This application is a U.S. National Phase of International Patent Application No. PCT/JP2021/001406 filed on Jan. 18, 2021, which claims priority benefit of Japanese Patent Application No. JP 2020-020560 filed in the Japan Patent Office on Feb. 10, 2020. Each of the above-referenced applications is hereby incorporated herein by reference in its entirety.
The present disclosure relates to an information processing device or the like having a function capable of changing a sound source position of a sound element.
In recent years, an information processing device such as a virtual assistant has been known as a software agent that executes a task or service for an individual. The information processing device includes, for example, a function of reproducing a content, a function of performing an alarm notification, a dialog function in which artificial intelligence (AI) interacts with a user, and the like. The information processing device is connected to headphones or the like worn by the user, and outputs, for example, a reproduced content of music, a moving image, or the like, and a sound signal for an alarm, a dialog, or the like from the headphones. Consequently, the user can receive various services while listening to the sound from the headphones.
For example, in a case where the dialog and the reproduced content are simultaneously generated, the information processing device localizes the sound source position of the dialog and the sound source position of the reproduced content to the same inside head by the headphones. In this case, since the dialog and the reproduced content overlap and sound interference occurs between the dialog and the reproduced content, it becomes difficult for the user to listen to the dialog and the reproduced content. Therefore, in the information processing device, content reproduction is stopped while a dialog is being generated. Furthermore, in the information processing device, a technology for enabling the sound source position of a sound element to be changed using headphones is widely known.
However, in the information processing device, for example, if the content reproduction is stopped in a case where another sound element of a dialog or the like is generated during the content reproduction, sound of the reproduced content is interrupted, and thus there is a risk of giving discomfort to the user.
Therefore, in order to cope with such a situation, for example, there is a demand for a technology for suppressing, even in a case where another sound element is simultaneously generated during content reproduction, sound interference between the reproduced content and the other sound element by changing a sound source position where the other sound element is heard or a sound source position where the reproduced content is heard. Furthermore, while there is a technology of localizing a sound image by each input sound signal to the outside head in order to change the sound source position as a headphone device, there is no technology of controlling the sound image localization position of the input sound signal, which needs to be reported according to the user, and there is a demand for this technology.
Therefore, the present disclosure proposes an information processing device and the like capable of suppressing sound interference between a sound element of a reproduced content and another sound element even in a case where the sound element of the reproduced content and the another sound element are simultaneously generated.
In order to solve the above-described problem, an information processing device according to an aspect of the present disclosure includes an acquisition unit that acquires a sound element of a content being reproduced and one or a plurality of other sound elements, a determination unit that determines importance levels of the sound elements acquired by the acquisition unit, and a signal processing unit that changes a sound source position of either the sound element of the content being reproduced or the other sound element according to the importance levels of the sound elements determined by the determination unit.
Hereinafter, embodiments of the present disclosure will be described in detail on the basis of the drawings. Note that in each of the following embodiments, the same parts are denoted by the same reference numerals, and redundant description will be omitted.
Furthermore, the present disclosure will be described according to the following order of items.
<1-1. Outline of Information Processing System>
An information processing device includes, for example, a function of a reproducing content, a function of performing an alarm notification, a dialog function in which AI interacts with a user, and the like. The information processing device is connected to headphones or the like worn by the user, and outputs, for example, a reproduced content of music, a moving image, or the like, and a sound signal for an alarm, a dialog, or the like from the headphones. Consequently, the user can receive various services while listening to the sound from the headphones.
However, for example, in a case where a dialog occurs while a reproduced content is being output by the headphones, the information processing device stops reproduction of the content and outputs an AI voice of the dialog, and thus there is a case of giving discomfort to the user due to the interruption of the reproduced content. Therefore, for example, there is a demand for a technology for suppressing, even in a case where another sound element is simultaneously generated during content reproduction, sound interference between the reproduced content and the other sound element by changing a sound source position where the other sound element is heard or a sound source position where the reproduced content is heard.
Therefore, in the present embodiment, this problem is solved by the following means.
For example, an information processing device includes an acquisition unit that acquires a sound element of a content being reproduced and one or a plurality of other sound elements, a determination unit that determines importance levels of the sound elements acquired by the acquisition unit, and a signal processing unit that changes a sound source position of either the sound element of the content being reproduced or the other sound element according to the importance levels of the sound elements determined by the determination unit.
The information processing device changes the sound source position of either the sound element of the content being reproduced or the other sound element according to the importance level of the sound element of the content being reproduced or the one or the plurality of other sound elements. Consequently, sound interference between sound elements can be suppressed by changing the sound source position of each sound element according to the importance level of the sound element without interrupting the content being reproduced. Then, efficiency of information transmission can be improved.
The outline of the present embodiment has been described above, and an information processing system 1 of the first embodiment will be described in detail below.
<2-1. Configuration of Information Processing Device>
<2-2. Configuration of Determination Unit>
The determination unit 12 determines an importance level of the sound element acquired by the system sound acquisition unit 11. The importance level of the sound element is information that ranks the degree of importance of the sound element. The signal processing unit 13 changes a sound source position in the sound output device 3 of a sound image of either the sound element of the content being reproduced or the other sound element according to the importance level of the sound element. Furthermore, the signal processing unit 13 may control the direction of the sound source instead of the position of the sound source, or may perform only direction estimation. Moreover, with respect to a sound source arranged in a three-dimensional space, represented by an object audio, control for specifying and/or changing the position thereof may be performed, or control in combination of any of the above may be performed.
The determination unit 12 includes an A/D conversion unit 21, an A/D conversion unit 21A, a separation unit 22, a sound element specification unit 23, a sound position estimation unit 24, an importance level specification unit 25, an importance level table 26, and a comparison unit 27. The A/D conversion unit 21 performs digital conversion of a sound signal of the system sound acquired by the system sound acquisition unit 11. The A/D conversion unit 21A performs digital conversion of a sound signal of the external sound acquired by the external sound acquisition unit 11A. The separation unit 22 separates the sound signal after digital conversion by the A/D conversion unit 21 and the A/D conversion unit 21A for each sound element by frequency analysis or the like. The separation unit 22 separates the sound signal of the system sound or the external sound for each sound element. The sound element specification unit 23 refers to a table that manages types of sound elements, which are not illustrated, and specifies a sound element type of each sound element. For example, the sound element specification unit 23 specifies the sound element type such as a sound element of a reproduced content, a sound element of an alarm, a sound element of a telephone sound, and a sound element of a dialog.
The sound position estimation unit 24 estimates an actual sound source position of each sound element by analyzing the sound elements separated by the separation unit 22. The sound position estimation unit 24 estimates the actual sound source position of each sound element, that is, a sound source position where the sound of the sound element is actually heard. For example, in a case of the sound element of the system sound such as the sound element of the reproduced content, the sound position estimation unit 24 estimates the inside head as the sound source position of the sound image of the sound element of the system sound because the sound output device 3 is worn. The importance level specification unit 25 refers to the importance level table 26 and specifies an importance level of each sound element. The comparison unit 27 compares the importance levels of the sound elements with each other, particularly, the importance level of the sound element of the reproduced content with the importance level of another sound element.
<2-3. Configuration of Signal Processing Unit>
The signal processing unit 13 includes a volume setting unit 31, a sound position setting unit 32, a noise cancellation unit 33, a 3D signal processing unit 34, a D/A conversion unit 35, and an amplifier unit 36. The volume setting unit 31 sets the volume of each sound element after 3D signal processing according to the importance level of each sound element. The volume of each sound element after the 3D signal processing is the volume of each sound element generated by the 3D signal processing unit 34. Note that the volume setting unit 31 refers to a table, which is not illustrated, that manages the volume according to the importance level of the sound element, and sets the volume of each sound element after the 3D signal processing corresponding to the importance level of each sound element.
The sound position setting unit 32 sets the sound source position of each sound element after the 3D signal processing according to the importance level of each sound element. The sound source position of each sound element after the 3D signal processing is the sound source position of each sound element generated by the 3D signal processing unit 34. Note that the sound position setting unit 32 refers to a table that is not illustrated that manages the sound source position according to the importance level of the sound element, and sets a sound position of the sound image of each sound element after the 3D signal processing corresponding to the importance level of each sound element. The noise cancellation unit 33 outputs a noise cancellation signal obtained by performing a noise cancellation process on the sound signal of the external sound after the digital conversion by the A/D conversion unit 21A to the D/A conversion unit 35.
The 3D signal processing unit 34 performs 3D signal processing of convolving a head-related impulse response (HRIR) with a digital signal of a sound element on the basis of the sound source position and the volume of each sound element to generate a 3D sound signal for localizing the sound image of the sound element to a desired sound source position. The head-related impulse response is a representation of a head-related transfer function (HRTF) representing a characteristic until sound emitted from the sound source reaches the user's ear on a time axis. The 3D signal processing unit 34 synthesizes the 3D sound signal of each sound element and outputs the synthesized 3D sound signals of all the sound elements. The 3D signal processing unit 34 generates the 3D sound signal for localizing the sound image of the sound element to a sound source position, for example, to the inside or outside head. Note that the outside head includes, for example, a sound source position separated from the inside head by an arbitrary distance in addition to a sound source position separated from the inside head of the user wearing the sound output device 3 by a certain distance. Furthermore, the 3D signal processing unit 34 also includes a head tracking function of relatively moving the sound source position according to an angular change in the direction of the face of the user.
The D/A conversion unit 35 performs analog conversion of the 3D sound signal of the 3D signal processing unit 34 and the noise cancellation signal subjected to the noise cancellation process by the noise cancellation unit 33. The amplifier unit 36 amplifies the 3D sound signal and the noise cancellation signal after analog conversion and outputs them to the sound output device 3. The information processing device 2 can localize the sound image of each sound element to each sound source position according to the 3D sound signal in a state where the noise cancellation process is performed. Consequently, the user wearing the sound output device 3 can listen to each sound element in a state of not only inside head localization but also outside head localization that sounds as if sounds were emitted from the surroundings.
<2-4. Configuration of Importance Level Table>
The sound source positions of the sound elements include, for example, positions of the inside head or the outside head. Furthermore, the outside head includes, for example, a front surface, a back surface, left and right side surfaces, a head top, and the like. The outside head may be anywhere as long as it is a position around the user's head, and can be appropriately changed. For example, in a case of the sound elements of the reproduced content and the alarm, since the sound elements of these system sounds are output from the sound output device 3, for example, the inside head is set as the actual sound source position. Note that, for convenience of description, it is assumed that the sound source positions of the sound elements are such that, for example, the inside head is the most audible to the user, and are ranked in the order of inside head→outside head of the front surface→outside head of the left and right side surfaces→outside head of the back surface→outside head of the head top, and the degree of being easily audible decreases in this order. Therefore, it is assumed that the sound source position of the sound element with high importance level is set to the inside head→outside head of the front surface→ . . . , and the sound source position of the sound element with low importance level is set to the outside head of the head top, for example. The hierarchy of the sound source positions is managed by a table that is not illustrated that manages the sound source positions according to the importance level. Note that the order of these sound source positions is merely an example, and can be changed as appropriate.
<2-5. Example of Sound Source Position>
The sound position estimation unit 24 estimates the inside head as the actual sound source positions of the sound element of the alarm and the sound element of the reproduced content. On the other hand, the sound position setting unit 32 compares the importance level “1” of the sound element of the alarm with the importance level “3” of the sound element of the reproduced content. Then, since the importance level of the sound element of the alarm is higher, the sound position setting unit 32 sets the alarm to the inside head and the reproduced content to the outside head of the head top as the sound source positions after the 3D signal processing. Since the sound source position of the alarm is localized to the inside head and the sound source position of the reproduced content is localized to the outside head of the head top, the sound source positions of the alarm sound and the reproduced content do not overlap. Therefore, by suppressing sound interference between the reproduced content and the alarm, the user can easily hear both the reproduced content and the alarm sound.
<2-6. First 3D Signal Generating Process>
In a case where the sound signal has been acquired (step S11: Yes), the A/D conversion unit 21 in the information processing device 2 performs digital conversion of the sound signal (step S12). Note that the A/D conversion unit 21A performs digital conversion of the sound signal of the external sound acquired by the external sound acquisition unit 11A. Then, the noise cancellation unit 33 adds the noise cancellation signal obtained by performing the noise cancellation process on the sound signal of the external sound to the signal output from the 3D signal processing unit 34, and then outputs the signal to the D/A conversion unit 35. The separation unit 22 in the information processing device 2 separates the sound element of the system sound from the sound signal after digital conversion by frequency analysis, sound source separation technique, or the like (step S13). The sound element specification unit 23 in the information processing device 2 specifies the sound element type of each sound element on the basis of a sound element separation result (step S14). The importance level specification unit 25 in the information processing device 2 refers to the importance level table 26 and specifies the importance level of each sound element (step S15).
The sound position estimation unit 24 in the information processing device 2 estimates the actual sound source position of each sound element from an analysis result of each sound element (step S16). The comparison unit 27 in the information processing device 2 determines whether or not a sound element of the content being reproduced is present in the sound elements (step S17). In a case where a sound element of the content being reproduced is present in the sound elements (step S17: Yes), the comparison unit 27 determines whether or not another sound element other than the sound element of the content being reproduced is present in the sound elements (step S18). Note that the another sound element is a sound element other than the content being reproduced in the system sound.
In a case where another sound element is present (step S18: Yes), the comparison unit 27 compares the importance level of the sound element of the content being reproduced with the importance level of the another sound element (step S19). The sound position setting unit 32 in the information processing device 2 sets the sound source position of each sound element after the 3D signal processing according to the importance level of each sound element (step S20). For example, in a case where an alarm and content being reproduced occur as sound elements, the importance level of the alarm is “1”, and the importance level of the content being reproduced is “3”. Therefore, the sound position setting unit 32 sets the sound source position of the alarm to the inside head and the sound source position of the content being reproduced to the outside head of the head top as the sound source positions after the 3D signal processing.
Moreover, the volume setting unit 31 in the information processing device 2 sets the volume of each sound element after the 3D signal processing according to the importance level of each sound element type (step S21). The volume setting unit 31 sets, for example, the volume of the alarm to 1 time and the volume of the content being reproduced to 1 time as the volumes after the 3D signal processing. Note that, as the volumes after the 3D signal processing, for example, the volume setting unit 31 may set the volume of the alarm to 1 time, the volume of the content being reproduced to 0.75 times, the volume of the alarm to 2 times, and the volume of the content being reproduced to 1 time, and can appropriately change the volume. The 3D signal processing unit 34 in the information processing device 2 generates a 3D sound signal of each sound element on the basis of the set sound source position and volume of each sound element after the 3D signal processing, and synthesizes and outputs the 3D sound signals for all the sound elements (step S22). The 3D signal processing unit 34 convolves a head impact response with the sound signal of each sound element on the basis of the sound source position and the volume of each sound element to generate the 3D sound signal of each sound element. Then, the 3D signal processing unit 34 synthesizes and outputs the 3D sound signal of each of all the sound elements.
The D/A conversion unit 35 in the information processing device 2 performs analog conversion of the synthesized 3D sound signal and noise cancellation signal (step S23). The amplifier unit 36 in the information processing device 2 amplifies the 3D sound signal and the noise cancellation signal after analog conversion, outputs the amplified 3D sound signal and noise cancellation signal to the sound output device 3 (step S24), and ends the processing operation illustrated in
In a case where the sound signal has not been acquired (step S11: No), the system sound acquisition unit 11 ends the processing operation illustrated in
The information processing device 2 compares the importance level of the sound element of the content being reproduced with the importance level of another sound element, sets the sound source position and the volume of each sound element after the 3D signal processing according to the importance level, and generates the 3D sound signal on the basis of the sound source position and the volume of each sound element. Moreover, the information processing device 2 provides the sound output device 3 with the 3D sound signal and the noise cancellation signal at the sound source position and with the volume according to the importance level. The information processing device 2 can suppress sound interference between the sound element of the reproduced content and another sound element in the system sound by changing the sound source position and the volume of each sound element according to the importance level, for example, by changing the sound source position between the sound element of the reproduced content and the another sound element in the system. Moreover, sound interference between sound elements in the system sound can be suppressed. Then, efficiency of information transmission of the system sound can be improved.
In the first embodiment, the sound source position in the sound output device 3 of the sound image of either the sound element of the content being reproduced or another sound element is changed according to the importance level of the sound element of the content being reproduced and the one or the plurality of other sound elements. Consequently, the sound source position of each sound element is changed according to the importance level of the sound element without interrupting the content being reproduced, thereby suppressing sound interference between the sound elements in the system sound, thereby improving the efficiency of information transmission.
Furthermore, for example, sound source positions of the reproduced content and another sound element are changed to the outside head according to the importance level of the system sound, and the sound source position is laid out by separating another sound element desired to be heard and the reproduced content. Consequently, it is possible to easily listen to the sound desired to be heard by using the importance level of each sound element reflecting the intention of the user.
Note that, for convenience of description, as illustrated in
Furthermore, in a case where the alarm sound has urgency and the importance level is much higher as compared to the importance level of the sound element of the reproduced content, the volume setting unit 31 may increase the volume of the sound element of the alarm after the 3D signal processing and decrease the volume of the sound element of the reproduced content after the 3D signal processing. Note that the case where the importance level is much higher is, for example, a case where the importance levels are two or more levels apart. Further, in a case where the alarm has urgency and the importance level is much higher as compared to the importance level of the sound element of the reproduced content, the reproduction of the reproduced content may be stopped instead of adjusting the volume. Furthermore, the reproduced content may be canceled using the noise cancellation unit 33.
Note that, in the information processing device 2 of the first embodiment, the case where the sound source position and the volume of each sound element in the system sound are adjusted according to the importance level of the sound element in the system sound has been exemplified, but it is not limited to the sound element of the system sound, and can also be applied to an external sound such as a car sound, for example. Therefore, the embodiment will be described below.
<3-1. Configuration of Information Processing Device>
Moreover, the information processing device 2A includes a device specification unit 14 that specifies the type of the sound output device 3. Types of the sound output device 3 include a headphone of an ear hole opening type worn by the user in a state where the ear hole of the user is opened, a headphone of an ear hole non-opening type worn like an ear plug, and the like. The ear hole opening type headphone is, for example, an open type headphone, and the ear hole non-opening type headphone is, for example, a canal type headphone. The device specification unit 14 may specify, for example, the sound output device 3 by the device type input by the user through a setting operation when connecting the sound output device 3 to the information processing device 2A. Furthermore, the device specification unit 14 may specify, for example, the device type through negotiation with the information processing device 2A when the sound output device 3 is connected to the information processing device 2A, and can be appropriately changed.
The separation unit 22 in the determination unit 12 separates sound elements such as the external sound and the system sound from the sound signal after digital conversion by the A/D conversion unit 21 and the A/D conversion unit 21A. The sound element specification unit 23 specifies the sound element type of a sound element such as the external sound or the system sound. The sound position estimation unit 24 estimates the sound source position of an actual sound element from the sound element such as the external sound or the system sound. For example, the sound position estimation unit 24 estimates an actual sound source position at which a car sound is heard. The importance level specification unit 25 refers to an importance level table 26A as described later and specifies the importance level for each sound element of the external sound or the system sound.
In a case where the sound output device 3 worn by the user is the headphone of the ear hole opening type, the signal processing unit 13 adjusts the sound source position and the volume of a sound element in the system sound other than the sound element of the external sound. In a case where the sound output device 3 worn by the user is the headphones of the ear hole non-opening type, the signal processing unit 13 adjusts the sound source position and the volume of a sound element of the external sound in addition to the sound element of the system sound.
<3-2. Configuration of Importance Level Table>
<3-3. Example of Sound Source Position>
For example, it is assumed a case where the sound output device 3 worn by the user is the headphones of the ear hole non-opening type and an external sound and a reproduced content occur. In a case where the importance level of the external sound is higher than that of the reproduced content, the sound position setting unit 32 sets the external sound to the actual sound source position and the reproduced content to the outside head of the head top as the sound source positions after the 3D signal processing. Furthermore, in a case where the importance level of the reproduced content is higher than that of the external sound or the system sound, the sound position setting unit 32 sets the reproduced content to the inside head or the outside head of the front surface, and the external sound to a sound source position different from the sound source position of the reproduced content, as the sound source positions after the 3D signal processing.
It is assumed a case where the sound output device 3 worn by the user is the headphones of the ear hole opening type and an external sound and a reproduced content occur. In a case where the importance level of the external sound is higher than that of the reproduced content, the sound position setting unit 32 sets the reproduced content to the outside head of the head top as the sound source position after the 3D signal processing. Further, in a case where the importance level of the reproduced content is higher than that of the external sound, the sound position setting unit 32 sets the reproduced content to the inside head or the outside head of the front surface as the sound source position after the 3D signal processing. Furthermore, in a case where the importance level of the reproduced content is higher than those of other system sounds, the sound position setting unit 32 sets the reproduced content to the inside head or the outside head of the front surface, and the other system sounds to sound source positions different from that of the reproduced content, as the sound source positions after the 3D signal processing.
In a case where there is no conversation between the user and Mr. A, the sound position estimation unit 24 estimates the inside head as the actual sound source position of the reproduced content. Since there is no conversation between the user and Mr. A and the importance level of the sound element of the reproduced content is the highest, the sound position setting unit 32 sets the reproduced content to the outside head of the front surface as the sound source position after the 3D signal processing. On the other hand, in a case where there is a conversation between the user and Mr. A, the sound position estimation unit 24 estimates the actual sound source position of the voice of Mr. A. Since the sound element of the voice of Mr. A has the highest importance level, the sound position setting unit 32 sets the reproduced content to the outside head of the head top as the sound source position after the 3D signal processing. Consequently, since it is the headphones of the ear hole opening type, the user can directly listen to the voice of Mr. A from the actual sound source position while listening to the reproduced content from the head top like BGM. That is, since the information processing device 2A adjusts the sound source position where the reproduced content can be heard in order to prioritize the voice of Mr. A, it is possible to suppress sound interference between the voice of Mr. A and the reproduced content.
Furthermore, for example, it is assumed a case where the user wearing the headphones of the ear hole non-opening type is with Mr. A while reproducing a content. In a case where there is no conversation between the user and Mr. A, the sound position estimation unit 24 estimates the inside head as the actual sound source position of the reproduced content. Since there is no conversation between the user and Mr. A and the importance level of the sound element of the reproduced content is the highest, the sound position setting unit 32 sets the reproduced content to the outside head of the front surface as the sound source position after the 3D signal processing. On the other hand, in a case where there is a conversation between the user and Mr. A, the sound position estimation unit 24 estimates the actual sound source position as the sound source position of the voice of Mr. A. Since the sound element of the voice of Mr. A has the highest importance level, the sound position setting unit 32 sets the reproduced content to the outside head of the head top and the voice of Mr. A to the actual sound source position (outside head) as the sound source positions after the 3D signal processing. Consequently, since it is the headphones of the ear hole non-opening type, the user can listen to the voice of Mr. A from the actual sound source position while listening to the reproduced content from the head top like BGM. That is, since the information processing device 2A adjusts the sound source and the volume in order to prioritize the voice of Mr. A, it is possible to suppress sound interference between the voice of Mr. A and the reproduced content.
<3-4. Second 3D Signal Generating Process>
In a case where the device specification unit 14 in the information processing device 2A has acquired the sound signal (step S11A: Yes), it is determined whether or not the sound output device 3 connected to the information processing device 2 is the headphones of the ear hole non-opening type (step S31). The sound output device 3 connected to the information processing device 2 is the headphones of the ear hole non-opening type or the headphones of the ear hole opening type worn by the user. In a case where the sound output device 3 is the headphones of the ear hole non-opening type (step S31: Yes), the A/D conversion unit 21 performs digital conversion of the sound signal (step S12A). The separation unit 22 in the information processing device 2A separates a sound element including an external sound from the sound signal after digital conversion by the frequency analysis, the sound source separation technique, or the like (step S13A). The sound element specification unit 23 in the information processing device 2A specifies a sound element type of each sound element including the external sound on the basis of the sound element separation result (step S14A). Note that the sound element type includes the sound element type of the external sound in addition to the sound element type of the system sound.
The importance level specification unit 25 in the information processing device 2A refers to the importance level table 26A and specifies the importance level for each sound element including the external sound (step S15A). The sound position estimation unit 24 in the information processing device 2A estimates the actual sound source position of each sound image of the sound element including the external sound from the analysis result of each sound element (step S16A).
The comparison unit 27 in the information processing device 2A determines whether or not a sound element of the content being reproduced is present in the sound elements (step S17A). In a case where a sound element of the content being reproduced is present in the sound elements (step S17A: Yes), the comparison unit 27 determines whether or not another sound element other than the sound element of the content being reproduced is present in the sound elements (step S18A).
In a case where another sound element is present (step S18A: Yes), the comparison unit 27 compares the importance level of the sound element of the content being reproduced with the importance level of another sound element (step S19A). The sound position setting unit 32 in the information processing device 2A determines whether or not the sound output device 3 of the user is the headphones of the ear hole non-opening type (step S33). In a case where the sound output device 3 of the user is the headphones of the ear hole non-opening type (step S33: Yes), the sound position setting unit 32 sets the sound source position after the 3D signal processing of each sound element according to the importance level of each sound element including the external sound and the system sound (step S20A). Note that the sound position setting unit 32 determines that the importance level of the voice of Mr. A is the highest in a case where the sound output device 3 is the headphones of the ear hole non-opening type and the sound element is the voice of Mr. A and content being reproduced. The sound position setting unit 32 sets the voice of Mr. A to the actual sound source position (outside head) and the content being reproduced to the outside head of the head top as the sound source positions after the 3D signal processing.
Moreover, the volume setting unit 31 in the information processing device 2A sets the volume after the 3D signal processing of each sound element according to the importance level of each sound element including the system sound and the external sound (step S21A). Note that the volume setting unit 31 sets the volume of the voice of Mr. A to 1 time and the volume of the content being reproduced to 1 time as the volumes after the 3D signal processing. The 3D signal processing unit 34 in the information processing device 2A generates the 3D sound signal of each sound element on the basis of the sound source position and the volume after the 3D signal processing of each sound element, and synthesizes and outputs the 3D sound signals for all the sound elements (step S22A). The 3D signal processing unit 34 convolves the head impact response of each sound element on the basis of the sound source position and the volume of each sound element to generate the 3D sound signal of each sound element. Then, the 3D signal processing unit 34 synthesizes and outputs all the 3D sound signals of each sound element including the system sound and the external sound.
The D/A conversion unit 35 in the information processing device 2A performs analog conversion of the noise cancellation signal from the noise cancellation unit 33 and all the 3D sound signals (step S23A). The amplifier unit 36 in the information processing device 2A amplifies and outputs the 3D sound signal and the noise cancellation signal after analog conversion to the sound output device 3 (step S24A), and ends the processing operation illustrated in
In a case where the sound output device 3 connected to the information processing device 2 is not the headphones of the ear hole non-opening type (step S31: No), the device specification unit 14 determines that the device type is the headphones of the ear hole opening type (step S32), and proceeds to step S12A to convert the sound signal into a digital signal. Furthermore, after comparing the importance levels in step S19A, in a case where the sound output device 3 of the user is not the headphones of the ear hole non-opening type (step S33: No), the sound position setting unit 32 determines that the sound output device 3 of the user is the headphones of the ear hole opening type. The sound position setting unit 32 sets the sound source position after the 3D signal processing for each sound element according to the importance level of each sound element including the system sound other than the external sound (step S20B). Note that, for example, in a case where the sound element is the voice of Mr. A and the content being reproduced, the sound position setting unit 32 determines that the importance level of the voice of Mr. A is the highest. Consequently, the sound position setting unit 32 can suppress interference between the content being reproduced and the voice of Mr. A by setting the content being reproduced to the outside head of the head top as the sound source position after the 3D signal processing.
Moreover, the volume setting unit 31 sets the volume after the 3D signal processing of each sound element according to the importance level of each sound element including the system sound other than the external sound (step S21B), and proceeds to step S22A to generate the 3D sound signal on the basis of the volume and the sound source position of each sound element. Note that the volume setting unit 31 sets the volume of the sound element of the content being reproduced to 1 time without adjusting the volume of the voice of Mr. A.
In a case where the system sound acquisition unit 11 and the external sound acquisition unit 11A have not acquired the sound signal (step S11A: No), the processing operation illustrated in
In the information processing device 2A of the second embodiment, in a case where the sound output device 3 of the user is the headphones of the ear hole opening type, the sound source position of the sound element of the system sound other than the external sound is adjusted according to the importance level of the sound element of the system sound and the external sound. Consequently, the sound source position of the system sound can be adjusted so as not to disturb the external sound. That is, sound interference between a sound element in the system sound and another sound element in the external sound can be suppressed.
In the information processing device 2A, in a case where the sound output device 3 of the user is the headphones of the ear hole non-opening type, the sound source positions of the sound elements of the system sound and the external sound are adjusted according to the importance levels of the sound elements of the system sound and the external sound. Consequently, the sound source positions of the system sound and the external sound can be adjusted so as not to disturb the external sound. That is, sound interference between a sound element in the system sound and another sound element in the external sound can be suppressed.
In the above-described information processing system 1A of the second embodiment, the case has been exemplified where the sound source position of the control target of the sound element is changed according to the type of the sound output device 3 worn by the user. However, the importance level of each sound element may be changed according to the current position of the user wearing the sound output device 3, and an embodiment thereof will be described below.
<4-1. Configuration of Information Processing Device>
In a case where the current position of the user is the home, the determination unit 12 sets the importance level of each sound element according to the home. In a case where the current position of the user is outside the home, the determination unit 12 sets the importance level of each sound element according to outside the home. In a case where the current position is the home, the signal processing unit 13 adjusts the sound source position and the volume of the sound element on the basis of the importance level of each sound element according to the home. In a case where the current position is outside the home, the signal processing unit 13 adjusts the sound source position and the volume of the sound element on the basis of the importance level of each sound element according to outside the home. Furthermore, it is assumed that the signal processing unit 13 does not change the sound source position because, for example, a car sound among the external sounds of outside the home has no meaning if the actual sound source position is changed.
<4-2. Configuration of Importance Level Table>
The sound element of the importance level 1 at home includes, for example, sound elements of the external sounds of a baby's cry, a voice of Mr. A, and a telephone sound, and sound elements of the system sounds of an alarm and a fixed telephone sound. The sound element of the importance level 2 at home includes, for example, a sound element of the system sound of the content being reproduced. The sound element of the importance level 3 at home includes, for example, a sound element of the external sound of a voice of a person other than Mr. A or a car sound.
On the other hand, the sound element of the importance level 1 outside the home includes, for example, sound elements of the external sounds of a baby's cry, a voice of Mr. A, a telephone sound, and a car sound, and a sound element of the system sound of an alarm. The sound element of the importance level 2 outside the home includes, for example, a sound element of the system sound of the content being reproduced. The sound element of the importance level 3 outside the home includes, for example, sound elements of the external sounds of a voice of a person other than Mr. A and a fixed telephone sound.
In a case where the importance level at home is compared with the importance level outside the home, the importance level at home is set such that, for example, the importance level of the fixed telephone sound is high and the importance level of the car sound is low. On the other hand, as the importance level outside the home, for example, the importance level of the car sound is set to be high, and the importance level of the fixed telephone sound is set to be low. Furthermore, the importance level of each sound element may be appropriately set and changed by the user.
<4-3. Example of Sound Source Position>
With the importance level of the sound element of the car being highest because it is outside the home, the sound position setting unit 32 sets the reproduced content to the outside head of the head top and the car sound to the actual sound source position (outside head) as the sound source positions after the 3D signal processing. At this time, the volume setting unit 31 sets the volume of the car sound to 1.5 times and the volume of the reproduced content to 0.5 times as the volumes after the 3D signal processing. Consequently, since the user is outside the home, the risk with respect to the car is high, and the user can increase the volume of the car from the actual position of the car and strongly recognize the presence of the car. In recent years, since the sound of an automobile has been reduced, it is possible to make the user strongly recognize the presence of the automobile by adjusting the volume of the sound of the automobile to be increased.
On the other hand, for example, it is assumed a case where sound elements of the reproduced content and the car sound occur at the home of the user wearing the headphones of the ear hole non-opening type. The sound position estimation unit 24 estimates an actual sound source position (inside head) as the sound source position of the reproduced content and an actual sound source position as the sound source position of the car sound. With the importance level of the sound element of the reproduced content being highest because it is at home, the sound position setting unit 32 sets the reproduced content to the outside head of the front surface and the car sound to the actual sound source position (outside head) as the sound source positions of each 3D signal processing. At this time, the volume setting unit 31 sets the volume of the reproduced content to 1 time and the volume of the car sound to 0 times as the volumes after the 3D signal processing. The risk with respect to the car at home is low, the user can listen to the reproduced content from the outside head of the front surface by erasing the car sound. Furthermore, by erasing unnecessary external sounds, there is also an effect such as reducing the burden on the user's brain and being less tiring.
For example, it is assumed a case where sound elements of the reproduced content and the car sound occur at the home of the user wearing the headphones of the ear hole opening type. The sound position estimation unit 24 estimates an actual sound source position (inside head) as the sound source position of the reproduced content and an actual sound source position as the sound source position of the car sound. With the importance level of the sound element of the reproduced content being highest because it is at home, the sound position setting unit 32 sets the reproduced content to the outside head of the front surface as the sound source position after the 3D signal processing. At this time, the volume setting unit 31 sets the volume of the reproduced content to 1 time as the volume after the 3D signal processing. Since the risk with respect to the car at home is low, the user can listen to the reproduced content from the outside head of the front surface while listening to the actual car sound.
On the other hand, for example, it is assumed a case where the sound elements of the reproduced content and the car sound occur outside the home of the user wearing the headphones of the ear hole opening type. With the importance level of the sound element of the car being highest because it is outside the home, the sound position setting unit 32 sets the reproduced content to the outside head of the head top as the sound source position after the 3D signal processing. At this time, the volume setting unit 31 sets the volume of the reproduced content to 0.5 times as the volume after the 3D signal processing. Consequently, the risk with respect to the car outside the home is high, and the user can recognize presence of the car by directly listening to the car sound from the actual position of the car while listening to the reproduced content like BGM from the head top. Furthermore, the importance levels of sound elements other than the car may be appropriately set to be the highest.
Furthermore, for example, it is assumed a case where a sound element of the reproduced content and a sound element of the voice of Mr. A occur at the home of the user wearing the headphones of the ear hole non-opening type. The sound position estimation unit 24 estimates the sound source position (inside head) of the reproduced content and the actual sound source position (outside head) of the voice of Mr. A as the actual sound source positions. On the other hand, the sound position setting unit 32 sets the reproduced content to the outside head of the head top and the voice of Mr. A to the actual sound source position (outside head) as the sound source positions after the 3D signal processing. The volume setting unit 32 sets the volume of the reproduced content and the volume of the voice of Mr. A to 1 time as the volumes after the 3D signal processing. Consequently, the user can listen to the voice of Mr. A while listening to the reproduced content from the head top like BGM.
For example, it is assumed a case where a sound element of the reproduced content and a sound element of the voice of Mr. A occur at the home of the user wearing the headphones of the ear hole opening type. The sound position estimation unit 24 estimates the sound source position (inside head) of the reproduced content and the actual sound source position (outside head) of the voice of Mr. A as the actual sound source positions. On the other hand, the sound position setting unit 32 sets the reproduced content to the outside head of the head top as the sound source position after the 3D signal processing. The volume setting unit 32 sets the volume of the reproduced content to 1 time as the volume after the 3D signal processing. Consequently, the user can directly listen to the voice of Mr. A while listening to the reproduced content from the head top like BGM.
Furthermore, for example, it is assumed a case where a sound element of the reproduced content and a sound element of the car sound occur outside the home of the user wearing the headphones of the ear hole non-opening type. The sound position estimation unit 24 estimates the sound source position (inside head) of the reproduced content and the actual sound source position of the car sound as the actual sound source positions. On the other hand, the sound position setting unit 32 sets the reproduced content to the outside head of the head top and the car sound to the actual sound source position (outside head) as the sound source positions after the 3D signal processing. The volume setting unit 31 sets the volume of the reproduced content to 0.5 times and the volume of the car sound to 1 time as the volumes after the 3D signal processing. Consequently, the user can listen to the car sound while listening to the reproduced content from the outside head of the head top like BGM.
For example, it is assumed a case where a sound element of the reproduced content and a sound element of the car sound occur outside the home of the user wearing the headphones of the ear hole opening type. The sound position estimation unit 24 estimates the sound source position (inside head) of the reproduced content and the actual sound source position of the car sound as the actual sound source positions. On the other hand, the sound position setting unit 32 sets the reproduced content to the outside head of the head top as the sound source position after the 3D signal processing. The volume setting unit 31 sets the volume of the reproduced content to 0.5 times as the volume after the 3D signal processing. Consequently, the user can directly listen to the car sound while listening to the reproduced content from the head top like BGM.
Furthermore, for example, it is assumed a case where a sound element of the reproduced content and a sound element of an announcement sound of an external sound occur outside the home of the user wearing the headphones of the ear hole non-opening type. The sound position estimation unit 24 estimates the actual sound source position (inside head) of the reproduced content and the actual sound source position (outside head) of the announcement sound as the actual sound source positions. On the other hand, the sound position setting unit 32 sets the reproduced content to the outside head of the front surface and the sound source position of the announcement sound to the outside head of the head top as the sound source positions after the 3D signal processing. Consequently, the user can listen to the announcement sound while listening to the reproduced content from the head top like BGM.
For example, it is assumed a case where a sound element of the reproduced content, a sound element of the voice of Mr. A, and a sound element of the voice of Mr. B occur at the home of the user wearing the headphones of the ear hole opening type. The sound position estimation unit 24 estimates, as the actual sound source positions, the sound source position (inside head) of the reproduced content, the actual sound source position (outside head) of the voice of Mr. A, and the actual sound source position (outside head) of the voice of Mr. B. On the other hand, since the voice of Mr. A has the importance level 1, the reproduced content has the importance level 2, and the voice of Mr. B has the importance level 3, the sound position setting unit 32 sets the reproduced content to the outside head of the head top as the sound source position after the 3D signal processing. Consequently, the user can directly listen to the voice of Mr. A while listening to the reproduced content from the head top like BGM.
<4-4. Third 3D Signal Generating Process>
The importance level specification unit 25 in the information processing device 2B specifies the sound element type of each sound element including the external sound in step S14A, and then refers to the importance level table 26B to specify the importance level corresponding to the current position for each sound element including the external sound and the system sound (step S15C). After specifying the importance level corresponding to the current position, the sound position estimation unit 24 in the information processing device 2B proceeds to step S16A so as to estimate the sound source position of each sound element including the system sound and the external sound.
In the information processing device 2B, in a case where the sound output device 3 of the user at home is the headphones of the ear hole opening type, the sound source position of the sound element of the system sound other than the external sound is adjusted according to the importance levels of the system sound corresponding to the home and the sound element of the external sound. Consequently, even in a case where the user wears the ear hole opening type headphones and is at home, the sound source position of the system sound can be adjusted so as not to disturb the external sound. That is, sound interference between a sound element in the system sound and another sound element in the external sound can be suppressed.
In the information processing device 2B, in a case where the sound output device 3 of the user at home is the headphones of the ear hole non-opening type, the sound source positions of the sound elements of the system sound and the external sound are adjusted according to the importance levels of the sound elements of the system sound and the external sound corresponding to the home. Consequently, even in a case where the user wears the headphones of the ear hole non-opening type and is at home, the sound source positions of the system sound and the external sound can be adjusted so as not to disturb the external sound. That is, sound interference between a sound element in the system sound and another sound element in the external sound can be suppressed.
In the information processing device 2B, in a case where the sound output device 3 of the user outside the home is the headphones of the ear hole opening type, the sound source position of the sound element of the system sound other than the external sound is adjusted according to the importance levels of the sound elements of the system sound and the external sound corresponding to outside the home. Consequently, even in a case where the user wears the headphones of the ear hole opening type and is outside the home, the sound source position of the system sound can be adjusted so as not to disturb the external sound. That is, sound interference between a sound element in the system sound and another sound element in the external sound can be suppressed.
In the information processing device 2B, in a case where the sound output device 3 of the user outside the home is the headphones of the ear hole non-opening type, the sound source positions of the sound elements of the system sound and the external sound are adjusted according to the importance levels of the sound elements of the system sound and the external sound corresponding to outside the home. Consequently, even in a case where the user wears the headphones of the ear hole non-opening type and is outside the home, the sound source positions of the system sound and the external sound can be adjusted so as not to disturb the external sound. That is, sound interference between a sound element in the system sound and another sound element in the external sound can be suppressed.
<4-5. Example of Operation of Information Processing Device>
The A/D conversion unit 21 performs digital conversion of the sound signal acquired by the system sound acquisition unit 11 and outputs the sound signal after digital conversion to the separation unit 22. The A/D conversion unit 21A performs digital conversion of the sound signal of the external sound acquired by the external sound acquisition unit 11A, and outputs the sound signal after digital conversion to the separation unit 22. The separation unit 22 separates, for example, a sound element of the voice of Mr. A, a sound element of the car sound, a sound element of the voice of Mr. B, and a sound element of the reproduced content from the sound signal by the frequency analysis or the like or the sound source separation technique or the like. The sound element specification unit 23 specifies a sound element type of each separated sound element. Moreover, the detection unit 15 determines that the current position of the user wearing the sound output device 3 is outside the home. The importance level specification unit 25 refers to the importance level table 26B and specifies the importance level of each sound element outside the home. The importance level specification unit 25 determines the importance levels of the voice of Mr. A and the car sound as the importance level 1, the importance level of the voice of Mr. B as the importance level 3, and the importance level of the reproduced content as the importance level 2.
Furthermore, the sound position estimation unit 24 estimates the sound source position of each separated sound element. The sound position estimation unit 24 estimates, for example, the actual sound source position of the voice of Mr. A (for example, front surface) and the actual sound source position of the car sound (for example, right side surface) as the actual sound source positions. Moreover, the sound position estimation unit 24 estimates an actual sound source position (for example, left side surface) of the voice of Mr. B and an actual sound source position (inside head) of the reproduced content as the actual sound source positions.
The sound position setting unit 32 sets the sound source position after the 3D signal processing of each sound element according to the importance level of each sound element. As the sound source positions after the 3D signal processing, for example, the sound position setting unit 32 sets the voice of Mr. A to an actual sound source position (outside head of the front surface) and the car sound to an actual sound source position (for example, outside head of the right side surface). Moreover, as the sound source position after the 3D signal processing, the sound position setting unit 32 sets the actual sound source position (for example, outside head of the left side surface) of the voice of Mr. B and the reproduced content to the outside head of the head top.
The volume setting unit 31 sets the volume after the 3D signal processing of each sound element according to the importance level of each sound element. As the volumes after the 3D signal processing, the volume setting unit 31 sets, for example, the volume of the voice of Mr. A to 1 time, the volume of the car sound to 1 time, the volume of the voice of Mr. B to 0 times, and the volume of the reproduced content to 1 time. The 3D signal processing unit 34 generates a 3D sound signal of each sound element on the basis of the volume and the sound source position of each sound element after the 3D signal processing. The 3D signal processing unit 34 synthesizes the 3D sound signal of each sound element and outputs the synthesized 3D sound signal to the D/A conversion unit 35. Then, the D/A conversion unit 35 performs analog conversion of the 3D sound signal and the noise cancellation signal processed by the noise cancellation unit 33, and acoustically outputs the 3D sound signal and the noise cancellation signal after analog conversion to the sound output device 3. Consequently, the user wearing the sound output device 3 can listen to the voice of Mr. A and the car sound from the actual sound source position without listening to the voice of Mr. B while listening to the reproduced content flowing from the head top like BGM.
As the information processing device 2, the information processing device such as a virtual assistant connected to the sound output device 3 has been exemplified. However, the information processing device is not limited to the virtual assistant, and can be applied to, for example, a content reproduction device or a smartphone having a function of reproducing a reproduced content or a function of acquiring another sound element, and can be appropriately changed.
Furthermore, the information processing device 2 may be, for example, a server device on a cloud, and may execute processing of the determination unit 12 and the signal processing unit 13 on the cloud and transmit the 3D sound signal generated by the signal processing unit 13 to the sound output device 3. Furthermore, the determination unit 12 and the signal processing unit 13 may be incorporated in the sound output device 3.
The sound output device 3 may be, for example, headphones for a head mounted display and the like of virtual reality (VR), augmented reality (AR), or the like, and can be appropriately changed.
In the information processing device 2, the case has been exemplified where the sound source position of the sound element is changed according to the importance level of the sound element, but acoustic characteristics of the sound element may be adjusted in addition to the change of the sound source position of the sound element.
In the information processing device 2, the case has been exemplified where the sound source position of the sound element is changed according to the importance level and the volume at the changed sound source position is adjusted. However, instead of adjusting the volume, the frequency characteristic of the sound element may be adjusted such that the sound element having a high importance level is easy to hear.
In the information processing device 2, the cases have been exemplified where the current position of the user wearing the sound output device 3 is at home or outside the home, and where the importance level of each sound element corresponding to the current position is defined. However, it is not limited to the home or outside the home, and for example, the importance level of each sound element corresponding to a place such as a company or a train may be defined.
Furthermore, the information processing device 2 is not limited to the current position or the like of the user wearing the sound output device 3, and the importance level of each sound element may be defined in association with the time zone of the user or the state of the user, for example, a state in which the user is studying, sleeping, or the like. For example, in a case where the user is studying, the reproduced content may be changed from the inside head to the outside head of the head top. Furthermore, in a case where the user is undergoing dental treatment, the content being reproduced may be changed from the inside head to the outside head of the head top, a sound element of ringing sound of a treatment device may be acquired, and the ringing sound of the treatment device may be canceled by sound of an opposite phase.
Furthermore, in the information processing device 2, the case has been exemplified where the sound source position of the sound element of the reproduced content is changed from the inside head to the outside head of the head top according to the importance level of the sound element, and the volume of the sound image at the sound source position is adjusted. However, in the reproduced content, the voice of a lyrics portion may be removed, and only the sound of an accompaniment portion may be output.
Furthermore, in the information processing device 2, it is assumed a case where the sound element of the reproduced content is higher than the importance level of the sound element of a first notification sound, and the sound element of the reproduced content is lower than the importance level of the sound element of a second notification sound. In this case, the information processing device 2 may change the first notification sound to the outside head of the head top and the second notification sound to the outside head of the front surface or back surface.
Furthermore, in the information processing device 2, the case has been exemplified where the importance level of the sound element of the voice of Mr. A is defined in advance in the importance level table 26A. However, for example, a person having high relevance with the user may be specified according to the social graph of the SNS, and the definition in the importance level table 26A may be updated so that the importance level of the sound element of the specified person becomes high.
Furthermore, the information processing device 2 may have a function of detecting a reaction motion such as facing the direction of sound of the user in a case where a sound element occurs, and may change the importance level of each sound element in the importance level table 26 according to a combination of the sound element and the state of the user according to the reaction motion of the user.
Furthermore, in the information processing device 2, for example, the case has been exemplified where the importance level of the sound element of the car sound is set to the importance level 1 in the case of outside the home. However, even in the case of outside the home, for example, the risk with respect to the car is low in a place where the vehicle does not pass, such as a pedestrian bridge, and thus the importance level may be changed according to the place. Furthermore, even in the case of outside the home, in the case of an accident frequent area, the volume of the sound element of a car may be increased to alert the user to the presence of the car.
In the information processing device 2, the frequency of the car sound may be adjusted simultaneously with the volume of the car sound to make it easier for the user to hear the sound of the car, or a specific sound may be added to make it easier for the user to recognize the presence of the car.
In the information processing device 2, a notification sound for notifying a green light, a red light, or the like of a crosswalk or the like may vary depending on the region. However, in a case where the information processing device 2 detects the notification sound of the crosswalk, the information processing device 2 may replace the notification sound with another notification sound of a region known by the user.
The case has been exemplified where the information processing device 2 outputs the 3D sound signal in which the sound source position of each sound element is appropriately changed according to the importance level of each sound element to the sound output device 3. However, the sound output device is not limited to the sound output device 3, and the sound source position may be changed using a plurality of speakers.
The information processing device 2 of the present embodiment may be implemented by a dedicated computer system or a general-purpose computer system.
For example, a program for executing the above-described operation (for example, the first 3D signal generating process, the second 3D signal generating process, and the third 3D signal generating process) is stored in a computer-readable recording medium such as an optical disk, a semiconductor memory, a magnetic tape, or a flexible disk and distributed. Then, for example, the program is installed in a computer and the above-described processing is executed, to thereby configure the information processing device 2 (2A, 2B).
Furthermore, the program described above may be stored in a storage device included in another information processing device on a network such as the Internet so that download of the program to a computer or the like is possible. Furthermore, the above-described functions may be implemented by cooperation of an operating system (OS) and application software. In this case, a portion other than the OS may be stored in a medium and distributed, or a portion other than the OS may be stored in a server device so that download to a computer or the like is possible.
Furthermore, among the respective processes described in the above-described embodiments, all or a part of the processes described as being performed automatically can be performed manually, or all or a part of the processes described as being performed manually can be performed automatically by a known method. In addition, information including the processing procedures, the specific names, and the various data and parameters illustrated in the document and the drawings described above can be arbitrarily changed unless otherwise specified. For example, the various types of information illustrated in the drawings are not limited to the illustrated information. Furthermore, in the above-described embodiments, there is a portion where a specific value is illustrated and described, but the value is not limited to the example, and another value may be used.
Furthermore, each component of each device illustrated in the drawings is functionally conceptual, and is not necessarily physically configured as illustrated. That is, a specific form of distribution and integration of each device is not limited to the illustrated form, and all or a part thereof can be configured in a functionally or physically distributed and integrated manner in an arbitrary unit according to various loads, usage conditions, and the like.
Furthermore, the above-described embodiments can be appropriately combined in a region in which the processing contents do not contradict each other. Furthermore, the order of each step illustrated in the flowcharts and the sequence diagrams of the above-described embodiments can be changed as appropriate.
Furthermore, for example, the present embodiment can be implemented as any component that constitutes a device or a system, for example, a processor as system large scale integration (LSI) or the like, a module using a plurality of processors or the like, a unit using a plurality of modules or the like, a set obtained by further adding other functions to the unit, and the like (that is, a configuration of a part of the device).
Note that, in the present embodiment, the system means a set of a plurality of components (devices, modules (parts), and the like), and it does not matter whether or not all the components are in the same housing. Therefore, both of a plurality of devices housed in separate housings and connected via a network and a single device in which a plurality of modules is housed in one housing are systems.
Furthermore, the present embodiment can employ, for example, a configuration of cloud computing in which at least one function (for example, the determination unit 12 or the signal processing unit 13) is shared and processed in cooperation by a plurality of devices via a network.
As described above, an information processing device according to an embodiment of the present disclosure includes an acquisition unit that acquires a sound element of a content being reproduced and one or a plurality of other sound elements, a determination unit that determines importance levels of the sound elements acquired by the acquisition unit, and a signal processing unit that changes a sound source position of either the sound element of the content being reproduced or another sound element according to the importance levels of the sound elements. Consequently, sound interference between sound elements can be suppressed by changing the sound source position of each sound element according to the importance level of the sound element without interrupting the content being reproduced. Then, the user can easily hear the another sound element while listening to the reproduced content.
In a case where an importance level of the another sound element is higher than an importance level of the sound element of the content being reproduced, the information processing device changes the sound source position of the sound element of the content being reproduced. Consequently, sound interference between the sound element of the content being reproduced and the another sound element can be suppressed. Then, the user can easily hear the another sound element while listening to the reproduced content.
In a case where an importance level of the another sound element is higher than an importance level of the sound element of the content being reproduced, the information processing device changes the sound source position of the sound element of the content being reproduced to a sound source position different from the sound source position of the another sound element. Consequently, sound interference between the sound element of the content being reproduced and the another sound element can be suppressed. Then, the user can easily hear the another sound element while listening to the reproduced content.
In a case where an importance level of the another sound element is higher than an importance level of the sound element of the content being reproduced, the information processing device changes the sound source position of the sound source of the sound element of the content being reproduced to outside head localization in a sound output device. Consequently, sound interference between the sound element of the content being reproduced and the another sound element can be suppressed. Then, the user can easily hear the another sound element while listening to the reproduced content.
In a case where an importance level of the sound element of the content being reproduced is higher than the importance level of the other sound elements, the information processing device changes to outside head localization in the sound output device in which the sound source position of the sound element of the content being reproduced and the sound source position of the another sound element are different. Consequently, sound interference between the sound element of the content being reproduced and the another sound element can be suppressed. Then, the user can easily hear the another sound element while listening to the reproduced content.
In a case where a movement of the user who uses the sound output device to a predetermined space is detected, the information processing device refers to an importance level table that manages the importance level of each sound element in each space corresponding to the predetermined space, and determines the importance level of an acquired sound element. The predetermined space can be, for example, the user's home or various environments outside the home. Furthermore, the movement to the predetermined space can mean, for example, when the user moves from inside to outside the home, or when the user moves from outside to inside the home, the user moves from one space outside the home to another space. Consequently, it is possible to suppress sound interference between sound elements according to an importance level in each predetermined space by changing the importance level of the sound element in each predetermined space in which the user using the sound output device moves.
The information processing device refers to the importance level table and determines the importance level of each separated sound element. Consequently, sound interference between sound elements can be suppressed by changing the importance level of each sound element.
The information processing device refers to the importance level table that manages an importance level of each sound element including a sound element of a specific person and the importance level table, emphasizes another sound element of the specific person in a case where the importance level of the sound element of the specific person specified from the separated another sound element is higher than a sound element of the content being reproduced, and cancels the another sound element of the specific person in a case where an importance level of the another sound element of the specified specific person is not higher than the sound element of the content being reproduced. Consequently, in a case where there is a voice of a specific person higher than the importance level of the reproduced content, the sound element of the specific person is prioritized, and in a case where there is a voice of a person other than the specific person not higher than the importance level of the reproduced content, sound interference with the voice of the specific person can be suppressed by canceling the voice.
The information processing device updates the importance level of each sound element including the sound element of the specific person such that the importance level of the sound element of the specific person increases according to a relevance between the user of the sound output device and the specific person. Consequently, the importance level of each sound element of the specific person can be updated according to the relevance between the user and the specific person.
In the importance level table, the importance level may be appropriately updated in one importance level table, or a plurality of importance level tables may be stored in advance in a cloud, a database, or the like, and a predetermined importance level table may be appropriately referred to from a plurality of importance level tables at a timing when the importance level is changed, or both patterns may be appropriately combined.
An information processing device acquires a sound element emitted by an information processing device and another sound element outside the information processing device including a sound element of an external sound captured from the outside of the information processing device, and allows, in a case where a sound output device that outputs the sound element to a sound source position is a headphone of an ear hole opening type, changing the sound source position of the sound element according to the importance level of the sound element emitted by the information processing device. Furthermore, in a case where the sound output device is the headphones of the ear hole non-opening type, the information processing device allows changing the sound source positions of the sound element according to the importance level of the sound element emitted by the information processing device and the importance level of the sound element of the external sound. Consequently, in a case where the user wears the headphones of the ear hole opening type, the sound source position of the sound element emitted by the information processing device can be changed according to the importance level of the sound element, and in a case where the user wears the ear hole non-opening type headphones, the sound source positions of the sound element emitted by the information processing device and the sound element of the external sound can be changed according to the importance level of the sound element.
Although the embodiments of the present disclosure have been described above, the technical scope of the present disclosure is not limited to the above-described embodiments as it is, and various modifications can be made without departing from the gist of the present disclosure. Furthermore, components of different embodiments and modification examples may be appropriately combined.
Furthermore, the effects of the embodiments described herein are merely examples and are not limited, and other effects may be provided.
Note that the present technology can have configurations as follows.
(1)
An information processing device including:
(2)
The information processing device according to (1) above, in which
(3)
The information processing device according to (1) or (2) above, in which
(4)
The information processing device according to any one of (1) to (3) above, in which
(5)
The information processing device according to any one of (1) to (4) above, in which
(6)
The information processing device according to any one of (1) to (4) above, in which
(7)
The information processing device according to any one of (1) to (6) above, further including
(8)
The information processing device according to any one of (1) to (7) above, in which the importance level table updates the importance level of the sound element according to a predetermined space in which a user wearing the sound output device is present.
(9)
The information processing device according to any one of (1) to (8) above, in which in a case where movement of the user wearing the sound output device from a first predetermined space to a second predetermined space is detected, the importance level table is updated with an importance level definition of the sound elements.
(10)
The information processing device according to any one of (1) to (9) above, in which the importance level table updates the importance level of the sound element according to relevance between a user wearing the sound output device and a specific person.
(11)
The information processing device according to (1) above, further including
(12)
The information processing device according to any one of (1) to (11) above, in which
(13)
The information processing device according to any one of (1) to (12) above, further including
(14)
An information processing method including:
(15)
An information processing program causing a computer to execute processing including:
(16)
An information processing system including an information processing device that acquires one or a plurality of other sound elements and a sound output device that outputs the sound element acquired in the image processing device to a sound source position, in which
Number | Date | Country | Kind |
---|---|---|---|
2020-020560 | Feb 2020 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2021/001406 | 1/18/2021 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2021/161722 | 8/19/2021 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
11290837 | Brimijoin, II | Mar 2022 | B1 |
20150189457 | Donaldson | Jul 2015 | A1 |
20160104491 | Lee | Apr 2016 | A1 |
20180146289 | Namm | May 2018 | A1 |
20180359592 | Laaksonen | Dec 2018 | A1 |
20230093585 | Faundez Hoffmann | Mar 2023 | A1 |
Number | Date | Country |
---|---|---|
11-331992 | Nov 1999 | JP |
2002-044797 | Feb 2002 | JP |
2006-115364 | Apr 2006 | JP |
2007-036610 | Feb 2007 | JP |
2016009850 | Jan 2016 | WO |
Entry |
---|
International Search Report and Written Opinion of PCT Application No. PCT/JP2021/001406, issued on Mar. 30, 2021, 09 pages of ISRWO. |
Number | Date | Country | |
---|---|---|---|
20230118803 A1 | Apr 2023 | US |