This application is a U.S. National Phase of International Patent Application No. PCT/JP2019/018335 filed on May 8, 2019, which claims priority benefit of Japanese Patent Application No. JP 201 8-1 4401 7 filed in the Japan Patent Office on Jul. 31, 2018. Each of the above-referenced applications is hereby incorporated herein by reference in its entirety.
A technique disclosed in the present specification mainly relates to an information processing apparatus, an information processing method, and an acoustic system that process acoustic information.
In an acoustic field, there is a known technique for using a head related transfer function (HRTF) of a user to increase reproducibility of stereophonic sound. The HRTF largely depends on the shapes of the head and the auricles of the user, and it is desirable that the user be the subject in measuring the head related transfer function.
However, in a conventional measurement method of the head related transfer function, special equipment provided with a large number of speakers is necessary, and opportunities for the measurement of the user are limited. It is difficult to use the head related transfer functions directly measured from the user to reproduce the stereophonic sound. Therefore, a method is often used, in which a dummy head microphone simulating the head including the ears is used to measure and set head related transfer functions of an average user, and the stereophonic sound for individual users is reproduced.
For example, a head related transfer function selection apparatus is proposed that selects the head related transfer function suitable for the user from a database including a plurality of head related transfer functions (see PTL 1). However, the head related transfer function selection apparatus uses the head related transfer function considered to be close to the user among the head related transfer functions with average characteristics registered in the database. Therefore, compared to the case of using the head related transfer function obtained by directly measuring the user, it cannot be denied that the sense of realism is reduced in reproducing the stereophonic sound.
In addition, an apparatus is proposed that measures head related transfer functions simulating propagation characteristics of sound propagating from each direction to the ears (see PTL 2). However, the apparatus uses a large speaker traverse (movement apparatus) to measure head related transfer functions at equal intervals, and large-scale equipment is necessary. Therefore, it is considered that the measurement burden of the user as a subject is large.
On the other hand, a control apparatus is proposed that acquires a positional relation between the head or the ear of the user and the sound source from an image captured by a smartphone held by the user and that causes the smartphone to generate sound to simply measure the head related transfer function (see PTL 3). However, there is a demand for a head related transfer function measurement technique that increases the measurement accuracy without imposing a measurement burden on the user as much as possible.
Japanese Patent Laid-Open No. 2014-99797
Japanese Patent Laid-Open No. 2007-251248
Japanese Patent Laid-Open No. 2017-16062
An object of the technique disclosed in the present specification is to provide an information processing apparatus, an information processing method, and an acoustic system that carry out a process for deriving a head related transfer function.
The technique disclosed in the present specification has been made in view of the problem, and a first aspect of the technique provides an information processing apparatus including a detection unit that detects a position of a head of a user, a storage unit that stores a head related transfer function of the user, a determination unit that determines a position of a sound source for measuring the head related transfer function of the user based on the position of the head detected by the detection unit and information stored in the storage unit, and a control unit that controls the sound source to output measurement signal sound from the position determined by the determination unit. In addition, the information processing apparatus further includes a calculation unit that calculates the head related transfer function of the user based on collected sound data obtained by collecting, at the position of the head, the measurement signal sound output from the sound source.
The determination unit determines a position of a sound source for measuring the head related transfer function of the user next without overlapping the position where the head related transfer function is already measured, thereby allowing to efficiently measure the head related transfer function.
In addition, a second aspect of the technique disclosed in the present specification provides an information processing method including a detection step of detecting a position of a head of a user, a determination step of determining a position of a sound source for measuring a head related transfer function of the user based on the position of the head detected in the detection step and information stored in a storage unit that stores the head related transfer function of the user, and a control step of controlling the sound source to output measurement signal sound from the position determined in the determination step.
In addition, a third aspect of the technique disclosed in the present specification provides an acoustic system including a control apparatus and a terminal apparatus. The control apparatus includes a detection unit that detects a position of a head of a user, a storage unit that stores a head related transfer function of the user, a determination unit that determines a position of a sound source for measuring the head related transfer function of the user based on the position of the head detected by the detection unit and information stored in the storage unit, a control unit that controls the sound source to output measurement signal sound from the position determined by the determination unit, and a calculation unit that calculates the head related transfer function of the user based on collected sound data obtained by collecting, at the position of the head, the measurement signal sound output from the sound source. The terminal apparatus includes a sound collection unit that is mounted on the user and used and that collects, at the position of the head, the measurement signal sound output from the sound source, and a transmission unit that transmits data collected by the sound collection unit to the control apparatus.
The “system” mentioned here denotes a logical set of a plurality of apparatuses (or functional modules that realize specific functions), and whether or not the apparatuses or the functional modules are in a single housing does not particularly matter.
The technique disclosed in the present specification can provide the information processing apparatus, the information processing method, and the acoustic system that carry out the process for deriving the head related transfer function.
Note that the advantageous effects described in the present specification are illustrative only, and the advantageous effects of the present invention are not limited to these. In addition, the present invention may also attain additional advantageous effects other than the advantageous effects described above.
Other objects, features, and advantages of the technique disclosed in the present specification will become apparent from more detailed description based on the embodiment described later and the attached drawings.
Hereinafter, an embodiment of the technique disclosed in the present specification will be described in detail with reference to the drawings.
First, “position,” “angle,” and “distance” in an HRTF measurement system disclosed in the present specification will be described. In the embodiment described below, a position of a measurement point for measuring an HRTF of a user will be expressed by general spherical coordinates. As illustrated in
In the measurement using the speaker traverse (movement apparatus) disclosed in PTL 2, the distance r from the measurement sound source to the head is fixed, and ϕ and θ are changed for the measurement. When the distance r is fixed, the measurement is performed for a plurality of positions (ϕ, θ, r), and the HRTF at an optional position (ϕ′, θ′, r) that is not a measurement point can be calculated by interpolation using a technique such as spline interpolation.
The present embodiment allows measurement at a position different from the position where the position (ϕ, θ, r) is set as a measurement point. Such a system is useful in a case of measuring the HRTF in a situation in which the position is not fixed, such as when the subject moves or changes the posture in the HRTF measurement system. In a case where the position of the measurement of the head of the user is (ϕ′, θ′, r′), and the position of the set measurement point is (ϕ, θ, r), an approximate value of the HRTF is measured. In this case, interpolation calculation using a generally well-known technique can be performed to obtain the HRTF at the position (ϕ, θ, r) of the measurement point from the HRTF of the measurement point approximated by the position (ϕ′, θ′, r′) and from the values of a plurality of measurement points measured at more accurate positions in the surroundings. In addition, an absolute value of a position error d=(ϕ, θ, r)−(ϕ′, θ′, r′) allowed for the position that can be approximated may be set within a certain range (|d|<=D), and the HRTF measurement system may be set to allow the approximate measurement within the range. In such a way, the measurement in the allowed range can be performed.
Note that the “position” in the present specification has three meanings including “position” of the measurement point described above, “position” where the user as a subject exists, and “position” of a sound source or the like for measurement or for drawing attention of the subject. It should be noted that the meaning of “position” is properly used as necessary in the HRTF measurement system described below.
With reference to
In addition, a plurality of gates 5, 6, 7, 8, . . . , each including an arch-shaped frame, is installed in a traveling direction of the user indicated by reference number 4. A plurality of speakers included in an acoustic signal generation unit 106 (described later) is installed at different places on the gate 5 in the front. In addition, a user position posture detection unit 103 (described later) is installed on the second gate 6 from the front. The acoustic signal generation unit 106 and the user position posture detection unit 103 may be alternately installed on the third and following gates 7, 8, . . . .
The user may not go straight in the traveling direction 4 in a constant posture. The user may meander or may be crouched, and the relative position and the posture with respect to the acoustic signal generation unit 106 may vary.
With reference to
The storage unit 101, the user position posture detection unit 103, the sound source position determination unit 104, the sound source position changing unit 105, the acoustic signal generation unit 106, the calculation unit 107, and the communication unit 108 are housed in the control box 2. In addition, the user specification unit 102 is housed in the user specification apparatus 3, and the user specification apparatus 3 is externally connected to the control box 2. In addition, the sound collection unit 109, the communication unit 110, and the storage unit 111 are housed in the terminal apparatus 1 mounted on the head of the user as a measurement target of the HRTF. Furthermore, the communication unit 108 on the control box 2 side and the communication unit 110 on the terminal apparatus 1 side are connected to each other through, for example, wireless communication.
In a case where the terminal apparatus 1 and the control box 2 communicate through a radio wave, each of the communication unit 108 and the communication unit 110 is equipped with an antenna (not illustrated). However, optical communication, such as infrared rays, can be used between the terminal apparatus 1 and the control box 2 in an environment with a little influence of interference. In addition, although the terminal apparatus 1 is basically battery-driven, the terminal apparatus 1 may be driven by a commercial power supply.
The user specification unit 102 includes a device that uniquely determines the current measurement target of HRTF. The user specification unit 102 includes, for example, an apparatus that can read (or identify) an ID card with IC, a magnetic card, a piece of paper on which a one-dimensional or two-dimensional barcode is printed, a smartphone in which an application for specifying the user is executed, a watch-type device including a wireless tag, a bracelet-type device, and the like. In addition, the user specification unit 102 may be a device that specifies the user by reasoning biometric information, such as fingerprint impression and vein authentication. In addition, the user specification unit 102 may specify the user based on a recognition result of a two-dimensional image or three-dimensional data of the user acquired by a camera or a 3D scanner. The user is managed by a user identifier (ID) registered in advance. In this case, the user as a temporary user may perform the first measurement of the HRFT, and a specific user ID and the measured HRTF may be associated after the measurement.
A process for measuring the HRTF of each user specified by the user specification unit 102 is executed in the control box 2.
HRTF measurement data of each user specified by the user specification unit 102, data necessary for the HRTF measurement process, and the like are stored in the storage unit 101. A mapping table or the like of users and data management storage areas of the users can be prepared to manage the data. Note that in the following description, one piece of data will be described for each user. Ideally, it is desirable to collect the HRTF data of each of the left ear and the right ear for each user. To do so, the HRTF data of each user is sorted into data for left ear and data for right ear and managed in the storage unit 101.
For each user, the HRTFs need to be measured at a plurality of measurement points in the spherical coordinates. In other words, the measurement points of HRTFs exist all around the user, and a set of HRTFs measured at all the measurement points is the HRTF data of the user. In the HRTF measurement system 100 according to the present embodiment, the user position posture detection unit 103 uses a camera, a distance sensor, or the like to measure the position and the posture of the head of the user (direction of head (may be direction of face or part of face (such as nose, eyes, and mouth), which similarly applies hereinafter)), and the sound source position determination unit 104 uses the measurement result to determine whether or not there is a sound source at a position where the HRTF needs to be measured for the user next (that is, whether or not the measurement is possible at the position that requires the measurement of the HRTF). To efficiently measure the HRTF data of the user at a plurality of measurement points in a short time, the sound source position determination unit 104 needs to extract the position of the sound source for measuring the HRTF next from position information of unmeasured measurement points to sequentially determine the sound source at the position for measuring the HRTF at the extracted measurement point so that the HRTF at an already measured position is not repeatedly measured. However, the measurement points for which quality determination of measurement data described later or quality determination of calculated HRTF has failed may be recorded as “unmeasured” or “remeasurement,” and the measurement of the HRTF may be repeated later for the measurement points.
More specifically, there is also reflected sound and reverberation unique to each acoustic environment, and it is preferable to acquire the HRTF data of each user for each acoustic environment. In the present embodiment, it is assumed that the HRTF data of each user is managed in association with the acoustic environment information in the storage unit 101. In addition, the terminal apparatus 1 or the control box 2 may be equipped with a sensor for sensing the acoustic environment during the HRTF measurement, or the user may instruct and input the acoustic environment at the time of the measurement of the HRTF through a UI (User Interface). As an example of acquiring and managing the HRTF data of each user for each acoustic environment, it is sufficient if the HRTF data is stored and managed for each combination of the environment information identifier of the acoustic environment information and the user identifier (ID) of the user information in
The user position posture detection unit 103 measures at which coordinates in the HRTF measurement system 100 the position of the head of the user specified by the user specification unit 102 exists and in which direction the user is facing (that is, posture information of user) in the spherical coordinates around the coordinates in a case where the head of the user is placed at the coordinate position. The user position posture detection unit 103 includes, for example, one or more cameras, one of a TOF (Time Of Flight) sensor, a laser measurement device (such as LiDAR), an ultrasonic sensor, or the like, or a combination of a plurality of sensors. Therefore, the HRTF measurement system 100 can measure the distance r from each speaker included in the acoustic signal generation unit 106 to the head of the user. In the example illustrated in
The acoustic signal generation unit 106 includes one or more speakers and provides sound sources that generate signal sound for HRTF measurement. In addition, the sound sources can also be used as sound sources that generate signal sound as information viewed by the user (or information prompting the user to view) as described later. In the example illustrated in
The sound source position determination unit 104 selects a position (ϕ, θ, r) of the HRTF to be measured next for the user as the current measurement target of HRTF (or the user currently specified by the user specification unit 102) from the relative position of the position and posture information of the head of the user obtained by the user position posture detection unit 103 and the acoustic signal generation unit 106 and sequentially determines the sound source (speaker) at the position for measuring the TRTF of the selected position. It is preferable in terms of the efficiency of processing that each sound source hold an identifier (ID), and the sound source be controlled based on the ID after the sound source is determined based on the position. Furthermore, in the case where the stability information of the posture is provided as posture information as described above, the sound source position may be determined when the posture is stable. In such a way, the measurement of the HRTF can be more stable.
The sound source position changing unit 105 controls the acoustic signal generation unit 106 to generate signal sound for HRTF measurement from the position of the sound source determined by the sound source position determination unit 104. In the present embodiment, the acoustic signal generation unit 106 includes a plurality of speakers arranged at different positions as illustrated in
The sound collection unit 109 includes a microphone that converts a sound wave into an electrical signal. The sound collection unit 109 is housed in the terminal apparatus 1 mounted on the head of the user as a measurement target of the HRTF and collects signal sound for HRTF measurement emitted from the acoustic signal generation unit 106. Note that quality determination may be performed to determine whether or not there is an abnormality in the acoustic signal collected by the sound collection unit 109. Furthermore, the data measured by the sound collection unit 109 is temporarily stored in the storage unit 111 and transmitted from the terminal apparatus 1 to the control box 2 side through the communication unit 110.
The data measured by the sound collection unit 109 is time axis waveform information obtained by collecting the HRTF measurement signal emitted from the sound source at the position determined by the sound source position determination unit 104. On the control box 2 side, once the data measured by the sound collection unit 109 is received through the communication unit 108, the data is stored in the storage unit 101. Furthermore, the calculation unit 107 calculates the HRTF at the position of the sound source from the time axis waveform information measured for each position of the sound source and causes the storage unit 101 to store the HRTF. When the calculation unit 107 calculates the HRTF, quality determination is performed to determine whether or not the data measured by the sound collection unit 109 is correctly measured (or the quality determination may be performed in causing the storage unit 101 to store the HRTF). In addition, quality determination of the HRTF calculated by the calculation unit 107 is also performed. Note that the calculation unit 107 may calculate the HRTF in parallel with the sound collection of the HRTF measurement signal or may calculate the HRTF when some amount of unprocessed collected sound data is accumulated in the storage unit 101 or at optional timing. Although not illustrated, in a case where the terminal apparatus 1 further includes a position detection sensor such as a GPS (Global Positioning System), communication between the communication unit 110 of the terminal apparatus 1 and the communication unit 108 of the control box 2 can be used to transmit the information of the position detection sensor to the control box 2 to allow the control box 2 to use the information to measure the distance to the position of the head of the user. In such a way, there is an advantageous effect that the information of distance to the head of the user can be obtained even in a case where there is no distance measurement apparatus fixed to the HRTF measurement system 100.
Note that the process and the data management in at least part of the functional modules in the control box 2 illustrated in
The control box 2 side waits until the user specification unit 102 of the user specification apparatus 3 specifies the user (No in SEQ301). Here, it is assumed that the user is wearing the terminal apparatus 1 on the head.
Furthermore, once the user specification unit 102 specifies the user (Yes in SEQ301), the control box 2 transmits a connection request to the terminal apparatus 1 (SEQ302) and waits until a connection finish notification is received from the terminal apparatus 1 (No in SEQ303).
On other hand, the terminal apparatus 1 side waits until the connection request is received from the control box 2 (No in SEQ351). Furthermore, once the terminal apparatus 1 receives the connection request from the control box 2 (Yes in SEQ351), the terminal apparatus 1 executes a process of connecting to the control box 2 and then returns a connection finish notification to the control box 2 (SEQ352). Subsequently, the terminal apparatus 1 prepares the sound collection of the HRTF measurement signal to be executed by the sound collection unit 109 (SEQ353) and waits for a notification of output timing of the HRTF measurement signal from the control box 2 side (No in SEQ354).
Once the connection finish is received from the terminal apparatus 1 (Yes in SEQ303), the control box 2 notifies the terminal apparatus 1 of the output timing of the HRTF measurement signal (SEQ304). Furthermore, the control box 2 waits for a defined time (SEQ305) and outputs the HRTF measurement signal from the acoustic signal generation unit 106 (SEQ306). Specifically, the HRTF measurement signal is output from the sound source (speaker) corresponding to the sound source position changed by the sound source position changing unit 105 according to the determination by the sound source position determination unit 104. Subsequently, the control box 2 waits to receive the sound collection finish notification and the measurement data from the terminal apparatus 1 side (No in SEQ307).
In response to the notification of the output timing of the HRTF measurement signal from the control box 2 (Yes in SEQ354), the terminal apparatus 1 starts the sound collection process of the HRTF measurement signal (SEQ355). Furthermore, once the terminal apparatus 1 collects the HRTF measurement signal for a defined time (Yes in SEQ356), the terminal apparatus 1 transmits the sound collection finish notification and the measurement data to the control box 2 (SEQ357).
Once the control box 2 receives the sound collection finish notification and the measurement data from the terminal apparatus 1 side (Yes in SEQ307), the control box 2 checks whether or not the acquisition of the measurement data necessary and sufficient for calculating the HRTF of the user specified in SEQ351 is finished (SEQ308). Here, the control box 2 also performs the quality determination to determine whether or not there is an abnormality in the acoustic signal collected by the sound collection unit 109 on the terminal apparatus 1 side.
In a case where the acquisition of the measurement data necessary and sufficient for calculating the HRTF is not finished yet (No in SEQ308), the control box 2 transmits a measurement continuation notification to the terminal apparatus (SEQ309) and returns to SEQ304 to repeatedly carry out the notification of the output timing of the HRTF measurement signal and the transmission process of the HRTF measurement signal.
Furthermore, once the acquisition of the measurement data necessary and sufficient for calculating the HRTF is finished (Yes in SEQ308), the control box 2 transmits a measurement finish notification to the terminal apparatus 1 (SEQ310) and finishes the process for the HRTF measurement.
In a case where the measurement continuation notification is received from the control box 2 (No in SEQ358) after the transmission of the sound collection finish notification and the measurement data (SEQ357), the terminal apparatus 1 returns to SEQ354 to wait for the notification of the output timing of the HRTF measurement signal from the control box 2 side and repeatedly carries out the sound collection process of the HRTF measurement signal and the transmission of the sound collection finish notification and the measurement data to the control box 2.
Furthermore, in a case where the measurement finish notification is received from the control box 2 (Yes in SEQ358), the terminal apparatus 1 finishes the process for the HRTF measurement.
Basically, the sound source position can be set at the position of the measurement point of the HRTF, and the HRTF of the measurement point can be obtained based on the collected sound data of the HRTF measurement signal output from the sound source position. The required number and density (spatial distribution) of measurement points vary depending on the usage or the like of the HRTF. In addition, the number of sound source positions, that is, measurement points, varies according to the required accuracy of HRTF data.
The HRTFs of two or more measurement points cannot be measured at the same time exactly, and therefore, the HRTFs need to be sequentially measured for each measurement point. According to the configuration of the HRTF measurement system 100 illustrated in
In the terminal apparatus 1, the sound collection unit 109 collects the sound of the HRTF measurement signal and transmits the collected sound data to the control box 2 through the communication unit 110. The calculation unit 107 calculates the HRTF at the corresponding measurement point based on the received collected sound data and causes the storage unit 101 to store the HRTF.
For the current position and posture of the head of the user, the sound source position determination unit 104 determines, each time, the sound source position not overlapping the already measured measurement points. Furthermore, the sound source position changing unit 105 selects the speaker matching the sound source position sequentially determined according to the movement of the user and causes the speaker to output the HRTF measurement signal. In such a way, the sound collection at the measurement point and the measurement of the HRTF are carried out.
Therefore, the HRTF can be efficiently measured at the measurement points all around the user while the user goes through the gates 5, 6, 7, 8, . . . on foot. Note that in the case where the speaker is not arranged at the sound source position strictly matching the position of the measurement point, the HRTF at the desirable position may be interpolated based on the data obtained by collecting the signal sound output from two or more positions near the position of the measurement point. Furthermore, in the case where there is a measurement point not finished with the measurement of the HRTF due to stationary environmental noise of the surroundings or sudden noise, the interpolation may also be performed based on the HRTF data of the surrounding measurement point normally finished with the measurement.
It is preferable that the sound source position determination unit 104 uniformly select the measurement points from the entire circumference to, for example, completely measure the head related transfer functions all around the user. Alternatively, the priority of the HRTF measurement may be set in advance for each measurement point, and the sound source position determination unit 104 may determine the next measurement point with a higher priority among the measurement points not overlapping the already measured measurement points. Even in a case where, for example, the HRTFs of all of the measurement points cannot be acquired while the user passes through the gates 5, 6, 7, 8, . . . just once, the HRTFs of the measurement points with higher priorities can be acquired early with a small number of passes.
By the way, the resolution of the sound source position of a human is high in the direction of the median plane (median sagittal plane), followed by the downward direction. On the other hand, the resolution is relatively low in the left and right directions. The reason that the resolution is high in the median plane direction is also based on the fact that how the sound from the sound source in the median plane direction is heard varies between the left ear and the right ear due to the difference between the shapes of the left and right auricles of humans. Therefore, a high priority may be allocated to a measurement point close to the median plane direction.
The HRTF measurement system 100 with the functional configuration illustrated in
For example, as illustrated in
The user position posture detection unit 103 measures at which coordinates in the HRTF measurement system 100 the position of the head of the user specified by the user specification unit 102 exists and in which direction the user is facing (that is, posture information of user) in the spherical coordinates around the coordinates. The position measurement allows the HRTF measurement system 100 to measure the distance r from each speaker to the head of the user. The sound source position determination unit 104 determines the position (ϕ, θ, r) of the sound source for measuring the HRTF next, from the relative position between the position and posture information of the head of the user obtained by the user position posture detection unit 103 and each speaker. In this case, the sound source position determination unit 104 may determine the next measurement point without overlapping the already measured measurement points and may determine the next measurement point with a higher priority. Furthermore, the sound source position changing unit 105 causes one of the speakers to output the HRTF measurement signal to generate the signal sound for HRTF measurement from the position of the sound source determined by the sound source position determination unit 104. The subsequent sound collection process of the HRTF measurement signal and the calculation process of the HRTF based on the collected sound data are carried out according to the process sequence illustrated in
Although the user (one of the three people, the parents and the son) as a measurement target of the HRTF sits on the sofa in
Therefore, in the example illustrated in
In the HRTF measurement system 100 illustrated in
The user position posture detection unit 103 measures, every moment, the position and the posture of the head of the user as a measurement target. The sound source position determination unit 104 determines, each time, the sound source position not overlapping the already measured measurement points as the next measurement point, for the current position and posture of the head of the user. Furthermore, the sound source position movement apparatus 1001 causes the acoustic signal generation unit 106 to move to the measurement point determined by the sound source position determination unit 104.
The sound source position movement apparatus 1001 may be, for example, an autonomously moving pet robot or an unmanned aerial vehicle such as a drone. A speaker that can output the HRTF measurement signal is provided as the acoustic signal generation unit 106 on the pet robot or the drone. The sound source position movement apparatus 1001 moves to the measurement point determined by the sound source position determination unit 104 and causes the acoustic signal generation unit 106 to output the HRTF measurement signal. In addition, the sound source position movement apparatus 1001 may be further equipped with a sensor, such as a camera, that can measure the position and the posture of the head of the user, and in this case, the motion of the user as a measurement target can be followed so that the position of the speaker relative to the user matches the measurement point determined by the sound source position determination unit 104. The sound collection unit 109 collects, on the head of the user as a measurement target, the HRTF measurement signal output from the acoustic signal generation unit 106. Furthermore, the calculation unit 107 calculates the HRTF at the corresponding measurement point based on the collected sound data.
According to the HRTF measurement system 1000 illustrated in
Although the user sits on a chair 1103 and does not move, the pet robot 1101 walks around the user, and the relative position between the head of the user and the speaker provided on the pet robot 1101 changes every moment. The sound source position determination unit 104 determines, each time, the sound source position not overlapping the already measured measurement points, for the current position and posture of the head of the user. Furthermore, the pet robot 1101 walks around the user to output the HRTF measurement signals from the measurement points sequentially determined by the sound source position determination unit 104. In such a way, the sound collection and the measurement of HRTFs can be carried out at all of the measurement points to acquire the HRTF data of the user.
Note that although not illustrated, the pet robot 1101 can use a distance measurement sensor or the like to measure the distance to the head of the user to measure the distance r from the speaker of the pet robot 1101 to the head of the user and store the distance r in the database along with the HRTF measurement data. More specifically, the pet robot 1101 can move to the positions (ϕ, θ, r) of a plurality of measurement points around the head of the user to perform the HRTF measurement of the user. The pet robot 1101 as an autonomous measurement system can move to a predetermined position, and the HRTF can be measured without imposing a burden on the user. In the case of the pet robot 1101, the pet robot 1101 generally operates at a position lower than the head of the user. Therefore, the HRTF can be obviously measured at a measurement point position below the head of the user. On the other hand, because of the character of the pet robot 1101, the pet robot 1101 can make an action expressing fondness for the user to thereby prompt the user to make an action to change the direction of the face, and this can easily make the user perform an operation, such as lowering the posture to lower the position of the head and facing downward. Therefore, the HRTF measurement in which a position above the head of the user is the measurement point can also be naturally performed. That is, the HRTF of the user can be measured without losing the usefulness as a partner of the user that is the original purpose of the pet robot 1101, and there is an advantageous effect that sound information (such as music and voice service) can be localized and provided on a three-dimensional space according to the characteristics of the user.
In addition, although the user sits on the chair 1103 and does not move, the drone 1102 flies around the user, and the relative position between the head of the user and the speaker provided on the drone 1102 changes every moment. The sound source position determination unit 104 determines, each time, the sound source position not overlapping the already measured measurement points, for the current position and posture of the head of the user. Furthermore, the drone 1102 flies around the user to output the HRTF measurement signals from the measurement points sequentially determined by the sound source position determination unit 104. In such a way, the sound collection and the measurement of HRTFs can be carried out at all of the measurement points to acquire the HRTF data of the user.
Note that although not illustrated, the drone 1102 can use a distance measurement sensor or the like to measure the distance to the head of the user to measure the distance from the speaker of the drone 1102 to the head of the user and store the distance in the database along with the HRTF measurement data. More specifically, the drone 110 can move to the positions (ϕ, θ, r) of a plurality of measurement points around the head of the user to perform the HRTF measurement of the user. The drone 1102 as an autonomous measurement system can move to a predetermined position, and the HRTF can be measured without imposing a burden on the user. Furthermore, in the case of the drone 1102, it is assumed that the user uses the drone 1102 in a floating state to capture an image from the air, and an excellent advantageous effect is attained particularly in the HRTF measurement from a position higher than the head of the user.
To measure the HRTF of one user, only one mobile apparatus among the pet robot 1101, the drone 1102, and the like may be used, or two or more mobile apparatuses may be used at the same time.
The sound source movement apparatus, such as the pet robot 1101 and the drone 1102, may not only move or fly around the user to output the HRTF measurement signal from the position determined by the sound source position determination unit 104, but may also remain stationary or hover and use voice guidance, flickering of light, or the like to instruct the user to move or to change the posture.
Note that although the mobile apparatus, such as the pet robot 1101 and the drone 1102, is provided with the functions of the sound source position movement apparatus 1001 and the acoustic signal generation unit 106, the mobile apparatus may be further provided with part or all of the functions of the control box 1 and the user specification apparatus 3 in
In the HRTF measurement system 100 illustrated in
The user position posture detection unit 103 measures, every moment, the position and the posture of the head of the user as a measurement target. The sound source position determination unit 104 determines, each time, the sound source position not overlapping the already measured measurement points, for the current position and posture of the head of the user. In addition, the sound source position changing unit 105 selects the best sound source for performing the HRTF measurement at the position determined by the sound source position determination unit 104.
Here, the speaker selected by the sound source position changing unit 105 may be separated from the position determined by the sound source position determination unit 104. For example, the installation of the sound sources covering all of the measurement positions may not be possible even in the system configuration including a large number of sound sources as illustrated in
Therefore, the HRTF measurement system 1800 can guide the user to set a desirable positional relation between the speaker and the head of the user, and there is also an advantage that the position-based HRTFs can be measured all around the user by using fewer speakers.
The information presentation unit 1801 can be provided by using, for example, a display, an LED (Light Emitting Diode), a light bulb, or the like. Specifically, the information presentation unit 1801 presents information to be viewed by the user (or information prompting the user to view) at a predetermined position of the display. Furthermore, when the user faces the information, the speaker selected by the sound source position changing unit 105 is in a positional relation with the user that allows to measure the HRTF of the position determined by the sound source position determination unit 104. Alternatively, the sound source position changing unit 105 selects one speaker that is in a positional relation corresponding to the position determined by the sound source position determination unit 104 with respect to the head when the user faces the information presented on the display. In any case, the speaker selected by the sound source position changing unit 105 outputs the HRTF measurement signal from the position determined by the sound source position determination unit 104 to the head of the user.
In addition, the information presentation unit 1801 can be provided by using a movement apparatus, such as the pet robot and the drone described above. Specifically, the information presentation unit 1801 prompts the user to make an action by causing the pet robot or the drone to move to a place at which the information presentation unit 1801 wants the user to face. Furthermore, when the user faces the pet robot or the drone, the speaker selected by the sound source position changing unit 105 is in a positional relation with the user that allows to measure the HRTF of the position determined by the sound source position determination unit 104. Alternatively, the sound source position changing unit 105 selects one speaker that is in a positional relation corresponding to the position determined by the sound source position determination unit 104 with respect to the head when the user faces the pet robot or the drone. In any case, the speaker selected by the sound source position changing unit 105 outputs, to the user, the HRTF measurement signal from the position determined by the sound source position determination unit 104.
In addition, the information presentation unit 1801 can be provided by using one of the plurality of speakers included in the acoustic signal generation unit 106. Specifically, the information presentation unit 1801 presents, from the speaker at the place where the information presentation unit 1801 wants the user to face, acoustic information for the user to face. Furthermore, when the user faces the sound source information, the speaker selected by the sound source position changing unit 105 as a speaker for outputting the HRTF measurement signal is in a positional relation corresponding to the position determined by the sound source position determination unit 104 with respect to the head of the user. Alternatively, the sound source position changing unit 105 selects one speaker in a positional relation that allows to measure the HRTF of the position determined by the sound source position determination unit 104, with respect to the head when the user faces the speaker from which the information presentation unit 1801 presents the acoustic information. In any case, the speaker selected by the sound source position changing unit 105 outputs the HRTF measurement signal from the position determined by the sound source position determination unit 104 to the head of the user.
The user specification unit 102 specifies the users 1941, 1942, and 1943 in the room 1900. In addition, the user position posture detection unit 103 measures the position and the posture of the head of each of the users 1941, 1942, and 1943.
The sound source position determination unit 104 refers to the measured position information of the users 1941, 1942, and 1943 managed in the storage unit 101 and determines the position (ϕ, θ, r) of the sound source for measuring the HRTF next for each of the users 1941, 1942, and 1943 so that the HRTF of an already measured position is not repeatedly measured. Furthermore, the sound source position changing unit 105 respectively selects the best speakers 1901, 1902, and 1903 for performing the HRTF measurement at the positions determined by the sound source position determination unit 104 for the users 1941, 1942, and 1943. However, the speakers 1901, 1902, and 1903 are separated from the positions determined by the sound source position determination unit 104 for the users 1941, 1942, and 1943, respectively.
The information presentation unit 1801 presents information for prompting the user to view at predetermined positions of the displays 1911, 1921, and 1931. Specifically, the information presentation unit 1801 displays information 1951 for prompting the user 1941 to view on the display 1911 and displays information 1952 for prompting the user 1942 to view. In addition, the information presentation unit 1801 displays information 1953 for prompting the user 1943 to view on the display 1931. The information 1951, 1952, and 1953 includes image information for making opportunities for the users 1941, 1942, and 1943 to change the directions of the heads, and the information 1951, 1952, and 1953 may be, for example, avatars of the users 1941, 1942, and 1943.
When the user 1941 faces the information 1951, the speaker 1901 is in a positional relation corresponding to the position determined by the sound source position determination unit 104 with respect to the head of the user 1941. Similarly, when the user 1942 faces the information 1952, the speaker 1902 is in a positional relation with the head of the user 1942 that allows to measure the HRTF of the position determined by the sound source position determination unit 104, and when the user 1943 faces the information 1953, the speaker 1903 is in a positional relation with the head of the user 1943 that allows to measure the HRTF of the position determined by the sound source position determination unit 104. In such a way, the speakers 1901, 1902, and 1903 selected by the sound source position changing unit 105 output the HRTF measurement signals from the positions determined by the sound source position determination unit 104 to the heads of the users 1941, 1942, and 1943, respectively.
The sound collection unit 109 of the terminal apparatus mounted on the head of each of the users 1941, 1942, and 1943 collects the HRTF measurement signal. Furthermore, the data measured by the sound collection unit 109 is temporarily stored in the storage unit 111 and transmitted from the terminal apparatus 1 to the control box 2 side through the communication unit 110. On the control box 2 side, the data measured by the sound collection unit 109 is received through the communication unit 108, and the calculation unit 107 calculates the HRTF of the position determined by the sound source position determination unit 104 for each of the users 1941, 1942, and 1943.
An operation example of making the user change the posture in the HRTF measurement system 1800 (another implementation example of the HRTF measurement system 1800) will be more specifically described with reference to
Furthermore, as illustrated in
Next, a specific configuration of the terminal apparatus 1 will be described.
As already described with reference to
In the present embodiment, individual differences in the heads, the bodies, the shapes of the earlobes, and the like of the users are taken into account, and the terminal apparatus 1 including the sound collection unit 109 has an in-ear body structure in order to collect sound close to the state in which the sound reaches the eardrums of individual users.
As can be understood from
The holding unit 1201 has a hollow structure as illustrated, and almost all of the inside is an opening portion. The holding unit 1201 does not close the ear hole of the user even in the state in which the holding unit 1201 is inserted into the cavity of the concha. That is, the earhole of the user is open, and the terminal apparatus 1 is an open-ear type. It can be stated that the terminal apparatus 1 has acoustic permeability even during the sound collection of the HRTF measurement signal. Therefore, even in a case of, for example, measuring the HRTF while the user is relaxing in the living room as illustrated in
A change in the ambient sound may also occur due to the influence of the diffraction or the reflection on the surface of the human body, such as the head, the body, and the earlobe of the user. According to the terminal apparatus 1 of the present embodiment, the sound collection unit 109 is provided near the entrance of the ear canal, and therefore, the influence of the diffraction or the reflection on each part of the human body, such as the head, the body, and the earlobe of each user, can be taken into account to obtain a highly accurate head related transfer function expressing the change in the sound.
Next, signal processing in the HRTF measurement system 100 according to the present embodiment will be described.
On the control box 2 side, the sound source position determination unit 104 checks, in the storage unit 101, the measured position information of the user specified by the user specification unit 102 and further determines the position of the sound source for measuring the HRTF next from the relative position between the head position and posture information of the user obtained by the user position posture detection unit 103 and the acoustic signal generation unit 106 so that the HRTF of the measured position information is not repeatedly measured.
The acoustic signal generation unit 106 includes a plurality of speakers that can output HRTF measurement signals. The sound source position changing unit 105 causes the speaker at the position determined by the sound source position determination unit 104 to output the HRTF measurement signal. It is preferable that the HRTF measurement signal be a broadband signal with known phase and amplitude, such as TSP (Time Stretched Pulse). Detailed information regarding the HRTF measurement signal is stored in the storage unit 101, and the HRTF measurement signal based on the information is output from the speaker.
The HRTF measurement signal output from the acoustic signal generation unit 106 is propagated through the space. The acoustic transfer function unique to the user, such as the influence of the diffraction and the reflection on the surface of the human body including the head, the body, the earlobe, or the like of the user, is further applied to the HRTF measurement signal, and the sound of the HRTF measurement signal is collected by the sound collection unit 109 in the terminal apparatus 1 mounted on the user. Subsequently, the collected sound data is transmitted from the terminal apparatus 1 to the control box 2.
On the control box 2 side, once the communication unit 108 receives the collected sound data transmitted from the terminal apparatus 1, the collected sound data is stored as position-based time axis waveform information in the storage unit 101 in association with the position determined by the sound source position determination unit 104.
Subsequently, the calculation unit 107 reads the position-based time axis waveform information from the storage unit 101 to calculate the HRTF and stores the HRTF as a position-based HRTF in the storage unit 101. In addition, the information of the position where the HRTF is measured is stored as measured position information in the storage unit 101.
In calculating the HRTF, the calculation unit 107 performs quality determination to determine whether or not the data measured by the sound collection unit 109 is correctly measured. For example, the measurement data stored in the storage unit 101 is discarded in a case where large noise is mixed in the measurement data.
In addition, an unmeasured or remeasurement flag is set for the measurement point for which the quality determination has failed, and the measurement of the HRTF is repeated later. For example, the position for which the quality determination has failed can be deleted from the measured position information in the storage unit 101, and the sound source position determination unit 104 can later determine the position as the sound source position again.
For example, based on the relative positional relation between the head position of the user obtained by the user position posture detection unit 103 and the acoustic signal generation unit 106 (or speaker used to output HRTF measurement signal), there is a time domain in which the measurement signal is not measured due to the metric space delay of the sound wave until the sound is collected by the sound collection unit 109 after the HRTF measurement signal is output (see
In addition, the information of the acoustic environment at the measurement place of the HRTF (such as acoustic characteristics in the room) as illustrated in
Furthermore, quality determination of the HRTF data calculated by the calculation unit 107 is also performed. This can determine poor quality of the measurement that cannot be determined from the collected sound data. An unmeasured or remeasurement flag is set for the measurement point for which the quality determination of the HRTF has failed, and the measurement of the HRTF is repeated later. For example, the position for which the quality determination has failed can be deleted from the measured position information in the storage unit 101, and the sound source position determination unit 104 can later determine the position as the sound source position again.
An entry is defined for each measurement point (that is, each measurement point number) in the table. Each entry includes a field storing information of the position of the corresponding measurement point relative to the user, a field storing distance information between the head of the user at the measurement and the speaker used for the measurement, a position-based time axis waveform information field storing waveform data of the sound wave collected by the sound collection unit 109 from the HRTF measurement signal output at the measurement point, a position-based HRTF field calculated by the calculation unit 107 based on the waveform data stored in the position-based time axis waveform information field, a measured flag indicating whether or not the HRTF is measured at the measurement point or the like, and a priority field indicating the priority of measuring the measurement point. The measured flag is data of 2 or more bits indicating “measured,” “unmeasured,” “remeasurement,” “approximate measurement,” or the like. Although not illustrated in
On the control box 2 side, the sound source position determination unit 104 refers to the table of the user specified by the user specification unit 102 in the storage unit 101 to select a measurement point with a high priority from the measurement points in which the measurement flag is not “measured” (that is, HRTF is unmeasured) and determines the position of the sound source for measuring the HRTF next. Furthermore, the sound source position changing unit 105 causes the speaker at the position determined by the sound source position determination unit 104 to output the HRTF measurement signal.
The HRTF measurement signal output from the acoustic signal generation unit 106 is propagated through the space. The acoustic transfer function unique to the user, such as the influence of the diffraction and the reflection on the surface of the human body including the head, the body, the earlobe, or the like of the user, is further applied to the HRTF measurement signal, and the sound of the HRTF measurement signal is collected by the sound collection unit 109 in the terminal apparatus 1 mounted on the user. Subsequently, the collected sound data is transmitted from the terminal apparatus 1 to the control box 2.
On the control box 2 side, once the communication unit 108 receives the collected sound data transmitted from the terminal apparatus 1, the collected sound data is stored in the position-base time axis waveform information field of the entry corresponding to the position determined by the sound source position determination unit 104 in the table illustrated in FIG. 15. In this case, the “measured” flag of the same entry is set to prevent repeatedly measuring the HRTF at the same measurement point.
Quality determination is applied to the collected sound data stored in the position-based time axis waveform information field of each entry to determine whether or not the collected sound data is correctly measured. Here, in a case where the quality determination of the collected sound data has failed, the measured flag of the corresponding entry is set to “unmeasured.” The sound source position determination unit 104 can later determine the same measurement point as a sound source position again.
On the other hand, in a case where the quality determination is successful, the calculation unit 107 calculates the HRTF from the collected sound data and stores the HRTF in the position-based HRTF in the same entry. In addition, quality determination of the HRTF data calculated by the calculation unit 107 is also performed. In such a way, poor quality of the measurement that cannot be determined from the collected sound data can be determined. Here, in a case where the quality determination of the HRTF data has failed, the measured flag of the corresponding entry is set to “unmeasured.” The sound source position determination unit 104 can later determine the same measurement point as a sound source position again.
Note that in a case where the measurement of the HRTFs of all of the measurement points cannot be finished within a defined time, the HRTFs unique to the user that can be measured and the HRTFs and the feature amounts of the other users measured in the past may be used to complete the HRTF data unique to the user.
In addition, an average value of the HRTFs of a plurality of other users measured in the past may be stored as an initial value in the position-based HRTF field of each entry in the table in the initial state (see
The S/N of each frequency band in the HRTF measurement signal can be adjusted according to the stationary noise of the measurement environment to realize more robust measurement of HRTF. For example, in a case where there is a band that cannot secure the S/N in the normal HRTF measurement signal, the HRTF measurement signal can be processed to secure the S/N of the band to realize stable HRTF measurement. The HRTF measurement signal will be described with reference to
In general, the power is inversely proportional to the frequency in the stationary noise of the measurement environment, and the noise is often similar to so-called pink noise, in which the lower the frequency, the larger the noise. Therefore, when the normal TSP signal is used for the measurement, the lower the frequency, the worse the S/N ratio of the measured signal sound and the environmental noise tends to be (see
A pink TSP (see
However, the environmental stationary noise may not only be simple pink noise but may be environmental stationary noise including noise in a high level at a specific frequency as illustrated in
In addition, the HRTF largely depends on the shapes of the head and the auricles of the user. Therefore, the HRTF has a feature that individual differences in characteristics are large in a high frequency, but the differences in characteristics are relatively small in a low frequency. Therefore, in a case where the S/N ratio cannot be secured due to the influence of the environmental noise in the low frequency, the HRTF may not be measured in the low frequency. HRTF characteristics that are already measured and that are not influenced by the environmental noise in the low frequency may be combined to stabilize the HRTF measurement.
The HRTF corresponding to the position of the sound source, that is, the position from the head of the user, is accumulated in a position-based HRTF database 1701. Specifically, the HRTF measurement system 100 accumulates the HRTFs measured for the users on the basis of positions (that is, HRTF data of each user).
A sound source generation unit 1702 reproduces a sound signal for the user to listen. The sound source generation unit 1702 may be, for example, a content reproduction apparatus that reproduces a sound data file stored in a medium, such as a CD (Compact Disc) and a DVD (Digital Versatile Disc). Alternatively, the sound source generation unit 1702 may generate sound of music supplied (streaming delivery) from the outside through a wireless system, such as Bluetooth (registered trademark), Wi-Fi (registered trademark), and a mobile communication standard (LTE (Long Term Evolution), LTE-Advanced, 5G, or the like). Alternatively, the sound source generation unit 1702 may receive sound automatically generated or reproduced by a server on a network (or cloud), such as the Internet, using a function of artificial intelligence or the like, sound (including sound recorded in advance) obtained by collecting voice of a remote operator (or instructor, voice actor, coach, or the like), or the like through a network and generate the sound on the system 1700.
A sound image position control unit 1703 controls a sound image position of the sound signal reproduced from the sound source generation unit 1702. Specifically, the sound image position control unit 1703 reads, from the position-based HRTF database 1701, the position-based HRTFs at the time that the sound output from the sound source at a desirable position reaches the left and right ears of the user. The sound image position control unit 1703 sets the position-based HRTFs in filters 1704 and 1705. The filters 1704 and 1705 convolve the position-based HRTFs of the left and right ears of the user into the sound signal reproduced from the sound source generation unit 1702. Furthermore, the sound passing through the filters 1704 and 1705 is amplified by amplifiers 1708 and 1709 and acoustically output from speakers 1710 and 1711 toward the left and right ears of the user.
Although the sound output from the speakers 1710 and 1711 is heard in the head of the user in a case where the position-based HRTFs are not convolved, the sound can be localized outside of the head of the user by convolving the position-based HRTFs. Specifically, the user hears the sound as if the sound is generated from the sound source position at the position of the sound source in measuring the HRTF. That is, the filters 1704 and 1705 can convolve the position-based HRTFs, and the user can recognize the sense of direction and some distance of the sound source reproduced by the sound source generation unit 1702 to localize the sound. Note that the filters 1704 and 1705 that convolve the HRTFs can be realized by FIR (Finite Impulse Response) filters, and in addition, filters approximated by a combination of computation and IIR (Infinite Impulse Response) on the frequency axis can also similarly realize the sound localization.
In the acoustic output system 1700 illustrated in
The user may select the position of sound localization (position from the user to the virtual sound source) or the type of acoustic environment through a user interface (UI) 1714. The sound image position control unit 1703 and the acoustic environment control unit 1712 read corresponding filter coefficients from the position-based HRTF database 1701 and the ambient acoustic environment database 1713, respectively, according to the user operation through the user interface 1714 and set the filter coefficients in the filters 1704 and 1705 and the filters 1706 and 1707. For example, the position for localizing the sound source and the acoustic environment may vary according to the differences in hearing sensation of individual users or according to the use conditions. Therefore, if the user can designate the sound source position and the acoustic environment through the user interface 1714, the convenience of the acoustic output system 1700 increases. Note that an information terminal, such as a smartphone, possessed by the user can be utilized for the user interface 1714.
In the present embodiment, the HRTF measurement system 100 measures the position-based HRTFs for each user, and the position-based HRTFs of each user are accumulated in the position-based HRTF database 1701 on the acoustic output system 1700 side. For example, the acoustic system 1700 may be further provided with a user identification function (not illustrated) for identifying the user, and the sound image position control unit 1703 may read the position-based HRTFs corresponding to the identified user from the position-based HRTF database 1701 to automatically set the position-based HRTFs in the filters 1704 and 1705. Note that face authentication or biometric authentication using biometric information, such as fingerprint, voiceprint, iris, and vein, may be used as the user identification function.
Furthermore, the acoustic output system 1700 may execute a process of fixing the sound image position to the real space in conjunction with the motion of the head of the user. For example, a sensor unit 1715 including a GPS, an acceleration sensor, a gyro sensor, or the like detects the motion of the head of the user, and the sound image position control unit 1703 reads the position-based HRTFs from the position-based HRTF database 1701 according to the motion of the head and automatically updates the filter coefficients of the filters 1704 and 1705. In such a way, even in a case where the motion of the head of the user is changed, the HRTFs can be controlled so that the sound can be heard from the sound source at a certain place on the real space. It is preferable to control the HRTF automatic update after the user designates the position for localizing the sound of the sound source through the user interface 1714.
Note that the sound image position control unit 1703 and the acoustic environment control unit 1712 may be software modules realized by a program executed on a processor, such as a CPU (Central Processing Unit), or may be dedicated hardware modules. In addition, the position-based HRTF database 1701 and the ambient acoustic environment database 1713 may be stored in a local memory (not illustrated) of the acoustic output system 1700 or may be databases on an external storage apparatus that can be accessed through a network.
Physical characteristics of the user as a measurement target will be described. It is known that there are individual differences in the HRTF due to the influence of differences in the physical characteristics of the shapes of the auricles or the like. Therefore, in a case where the size of the head of the user as a subject can be measured during or before the use of the HRTF measurement system according to the present embodiment, the center of the spherical coordinates can be set at the center of the distance between the ears.
As for the measurement of the physical characteristics in the HRTF measurement system, an image capturing apparatus (not illustrated), such as a camera, can be incorporated in each example of the present specification to image the head of the user as a subject that operates in the HRTF measurement system, and a technique, such as image processing, can be used to analyze the captured image. In such a way, information, such as the vertical and horizontal size of the auricle of the ear of the user, the vertical and horizontal size of the cavity of the concha, the distance between the auricles as viewed from above the head, the distance between the ears (described above), the head posture (the front of the head (semicircle), the back of the head (semicircle)), and the head distance (distance from the tip of the nose to the edge of the back of the head as viewed from the side of the head), can be acquired, and the information can be used as parameters in the HRTF calculation. As a result, more accurate sound localization can be provided based on the HRTF data measured for individuals.
The technique disclosed in the present specification has been described in detail with reference to the specific embodiment. However, it is apparent that those skilled in the art can modify or substitute the embodiment without departing from the scope of the technique disclosed in the present specification.
The technique disclosed in the present specification can be applied to measure the position-based HRTFs all around the user without using large-scale equipment such as a large speaker traverse (movement apparatus). In addition, the HRTF measurement system according to the technique disclosed in the present specification sequentially determines the position of the sound source for measuring the HRTF next without repeatedly measuring the HRTFs of the already measured positions and measures the HRTFs at all of the measurement points. Therefore, there is no physical or mental burden of the user. In addition, according to the technique disclosed in the present specification, the measurement of the HRTF can be progressed in the living room or can be progressed by using the pet robot or the drone, without the user noticing the measurement in everyday life.
That is, the technique disclosed in the present specification has been described in a form of an example, and the description of the present invention should not be restrictively interpreted. The claims should be taken into account to determine the scope of the technique disclosed in the present specification.
Note that the technique disclosed in the present specification can also be configured as follows.
An information processing apparatus including:
a detection unit that detects a position of a head of a user;
a storage unit that stores a head related transfer function of the user;
a determination unit that determines a position of a sound source for measuring the head related transfer function of the user based on the position of the head detected by the detection unit and information stored in the storage unit; and
a control unit that controls the sound source to output measurement signal sound from the position determined by the determination unit.
The information processing apparatus according to (1), further including:
a specification unit that specifies the user.
The information processing apparatus according to (1) or (2), in which
the determination unit determines a position of a sound source for measuring the head related transfer function of the user next without overlapping the position where the head related transfer function is already measured.
The information processing apparatus according to any one of (1) to (3), in which
the control unit selects one of a plurality of sound sources arranged at different positions based on the position determined by the determination unit and causes the sound source to output measurement signal sound.
The information processing apparatus according to any one of (1) to (3), in which
the control unit causes the sound source moved based on the position determined by the determination unit to output measurement signal sound.
The information processing apparatus according to any one of (1) to (5), further including:
a calculation unit that calculates the head related transfer function of the user based on collected sound data obtained by collecting, at the position of the head, the measurement signal sound output from the sound source.
The information processing apparatus according to (6), further including:
a first determination unit that determines whether or not there is an abnormality in the collected sound data.
The information processing apparatus according to (7), in which
the first determination unit performs the determination by handling, as no-signal, the collected sound data in a time domain in which a measurement signal is not measured due to a metric space delay between the position of the head and the position of the sound source.
The information processing apparatus according to any one of (6) to (8), further including:
a second determination unit that determines whether or not there is an abnormality in the head related transfer function calculated by the calculation unit.
The information processing apparatus according to any one of (6) to (9), in which
the calculation unit uses collected sound data of measurement signal sound output from the sound source near the position determined by the determination unit to interpolate the head related transfer function at the position determined by the determination unit.
The information processing apparatus according to any one of (1) to (10), in which
the determination unit sequentially determines the position of the sound source for measuring the head related transfer function of the user next so as to evenly measure the head related transfer function throughout an area to be measured.
The information processing apparatus according to any one of (1) to (10), in which
the determination unit sequentially determines the position of the sound source for measuring the head related transfer function of the user next based on a priority set in the area to be measured.
The information processing apparatus according to any one of (1) to (12), further including:
an information presentation unit that presents information prompting the user to make an action of changing the position of the head when there is no sound source that generates the measurement signal sound at the position of the sound source determined by the determination unit.
The information processing apparatus according to (13), further including:
a display, in which
the information presentation unit presents the information to be viewed by the user at a predetermined position of the display, and
causes the sound source that is at a position determined by the determination unit for the head after the change in the position, to generate measurement signal sound when the user faces the information.
The information processing apparatus according to (13) or (14), including:
a plurality of sound sources, in which
the control unit controls a sound source that is arranged at a position determined by the determination unit for the head after the change in the position, to output measurement signal sound.
The information processing apparatus according to (13), including:
a plurality of sound sources, in which
the control unit determines a first sound source among the plurality of sound sources that outputs measurement signal sound,
the information presentation unit determines a second sound source among the plurality of sound sources that presents acoustic information prompting the user to make an action, and
when the user faces the acoustic information, the first sound source is in a positional relation corresponding to the position determined by the determination unit with respect to the position of the head.
The information processing apparatus according to any one of (1) to (16), in which
the measurement signal sound includes a time-stretch pulse in which power is inversely proportional to frequency.
The information processing apparatus according to any one of (1) to (16), in which
the measurement signal sound includes a time-stretch pulse in which amplitude at each frequency is adjusted according to a frequency spectrum of stationary noise of a measurement environment.
An information processing method including:
a detection step of detecting a position of a head of a user;
a determination step of determining a position of a sound source for measuring a head related transfer function of the user based on the position of the head detected in the detection step and information stored in a storage unit that stores the head related transfer function of the user; and
a control step of controlling the sound source to output measurement signal sound from the position determined in the determination step.
An acoustic system including:
a control apparatus including
a terminal apparatus including
1 . . . Terminal apparatus, 2 . . . Control box, 3 . . . User specification apparatus
5, 6, 8, 8, . . . Gate
100 . . . HRTF measurement system
101 . . . Storage unit, 102 . . . User specification unit
103 . . . User position posture detection unit, 104 . . . Sound source position determination unit
105 . . . Sound source position changing unit, 106 . . . Acoustic signal generation unit, 107 . . . Calculation unit
108 . . . Communication unit, 109 . . . Sound collection unit, 110 . . . Communication unit
1000 . . . HRTF measurement system, 1001 . . . Sound source position movement apparatus
1101 . . . Pet robot, 1102 . . . Drone, 1103 . . . Chair
1700 . . . Acoustic output system
1701 . . . Position-based HRTF database, 1702 . . . Sound source generation unit
1703 . . . Sound image position control unit, 1704, 1705 . . . Filter
1706, 1707 . . . Filter, 1708, 1709 . . . Amplifier
1710, 1711 . . . Speaker, 1712 . . . Acoustic environment control unit
1713 . . . Ambient acoustic environment database
1714 . . . User interface, 1715 . . . Sensor unit
1800 . . . HRTF measurement system, 1801 . . . Information presentation unit
1900 . . . Room, 1901, 1902, 1903 . . . Speaker
1910, 1920, 1930 . . . Wall surface
1911, 1921, 1931 . . . Display
Number | Date | Country | Kind |
---|---|---|---|
JP2018-144017 | Jul 2018 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2019/018335 | 5/8/2019 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2020/026548 | 2/6/2020 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
20130036452 | Yamashita | Feb 2013 | A1 |
20150156599 | Romigh | Jun 2015 | A1 |
20150312694 | Bilinski | Oct 2015 | A1 |
20160119731 | Lester, III | Apr 2016 | A1 |
20160373877 | Laitinen | Dec 2016 | A1 |
20170164085 | Romigh | Jun 2017 | A1 |
20180192226 | Woelfl | Jul 2018 | A1 |
20200068334 | Magariyachi et al. | Feb 2020 | A1 |
Number | Date | Country |
---|---|---|
102915734 | Feb 2013 | CN |
110036655 | Jul 2019 | CN |
3554098 | Oct 2019 | EP |
2007-251248 | Sep 2007 | JP |
2013-033368 | Feb 2013 | JP |
2014-99797 | May 2014 | JP |
2017-16062 | Jan 2017 | JP |
2016065137 | Apr 2016 | WO |
2017135063 | Aug 2017 | WO |
2018110269 | Jun 2018 | WO |
Entry |
---|
International Search Report and Written Opinion of PCT Application No. PCT/JP2019/018335, dated Jul. 16, 2019, 10 pages of ISRWO. |
Morise, et al., Warped-TSP: An Acoustic Measurement Signal Robust to Background Noise and Harmonic Distortion, Electronics and Communications in Japan, Part 3, vol. 90, No. 4, XP001504053, Jan. 1, 2007, pp. 18-26. |
Extended European Search Report of EP Application No. 19845437.3, dated Aug. 27, 2021, 10 pages. |
Number | Date | Country | |
---|---|---|---|
20210345057 A1 | Nov 2021 | US |