This application is based upon and claims the benefit of priority from the Japanese Patent Application No. 2021-151457, filed Sep. 16, 2021, the entire contents of which are incorporated herein by reference.
Embodiments described herein relate generally to an online conversation management apparatus and a storage medium storing an online conversation management program.
A sound image localization technique is known in which a sound image is localized in a space around the head of a user by using various types of sound reproduction devices different in sound reproduction environment, such as two channels of loudspeakers arranged in front of the user, earphones attached to the ears of the user, and headphones attached to the head of the user. This sound image localization technique can provide the user with an illusion that the sound is heard from a direction different from the direction in which the reproduction device actually exists.
Recently, attempts are being made to use the sound image localization technique in online conversation. In the case of an online meeting, for example, it is sometimes difficult to distinguish between the voices of a plurality of speakers because the voices are concentrated. By contrast, when the sound images of individual speakers are localized in different directions of a space around the head of a user, the user can distinguish between the voices of the individual speakers.
To localize sound images in a space around the head of each user, information of the sound reproduction environment of a reproduction device of each user must be known. If the sound reproduction environments of voice reproduction devices of users are different, an inconvenience that sound images are appropriately localized for one user but are not appropriately localized for another user can occur.
An embodiment provides an online conversation management apparatus and a storage medium storing an online conversation management program, by which appropriately localized sound images are reproduced for each user even when the sound reproduction environments of voice reproduction devices of individual users are different in the case of online conversation.
In general, according to one embodiment, an online conversation management apparatus includes a processor. The processor acquires, across a network, reproduction environment information from at least one terminal that reproduces a sound image via a reproduction device. The reproduction environment information is information of a sound reproduction environment of the reproduction device. The processor acquires azimuth information. The azimuth information is information of a localization direction of the sound image with respect to a user of the terminal. The processor performs control for reproducing a sound image of each terminal based on the reproduction environment information and the azimuth information.
Embodiments will be explained below with reference to the accompanying drawings.
The processor 1 controls the overall operation of the terminal. For example, the processor 1 of the host terminal HT operates as a first acquisition unit 11, a second acquisition unit 12, and a control unit 13 by executing programs stored in the storage 3 or the like. In the first embodiment, the processor 1 of each of the guest terminals GT1, GT2, and GT3 is not necessarily be operable as the first acquisition unit 11, the second acquisition unit 12, and the control unit 13. The processor 1 is, e.g., a CPU. The processor 1 can also be an MPU, a GPU, an ASIC, an FPGA, or the like. Furthermore, the processor 1 can be a single CPU and can also be a plurality of CPUs.
The first acquisition unit 11 acquires reproduction environment information input on the terminals HT, GT1, GT2, and GT3 participating in the online conversation. The reproduction environment information is information on the sound reproduction environment of the voice reproduction device 4 used in each of the terminals HT, GT1, GT2, and GT3. This information on the sound reproduction environment contains information indicating a device to be used as the voice reproduction device 4. The information indicating a device to be used as the voice reproduction device 4 is information indicating which of, for example, stereo loudspeakers, headphones, and earphones are used as the voice reproduction device 4. When the stereo loudspeakers are used as the voice reproduction device 4, the information on the sound reproduction environment also contains information indicating, for example, the distance between the right and left loudspeakers.
The second acquisition unit 12 acquires azimuth information input on the terminal HT participating in the online conversation. The azimuth information is information of sound image localization directions with respect to each of the terminal users including the user HU of the terminal HT.
The control unit 13 performs control for reproducing sound images on the individual terminals including the terminal HT based on the reproduction environment information and the azimuth information. For example, based on the reproduction environment information and the azimuth information, the control unit 13 generates sound image filter coefficients suitable for the individual terminals, and transmits the generated sound image filter coefficients to these terminals. The sound image filter coefficient is a coefficient that is convoluted in right and left voice signals to be input to the voice reproduction device 4. For example, the sound image filter coefficient is generated based on a head transmission function C as the voice transmission characteristic between the voice reproduction device 4 and the head (the two ears) of a user, and a head transmission coefficient d as the voice transmission characteristic between a virtual sound source specified in accordance with the azimuth information and the head (the two ears) of the user. For example, the storage 3 stores a table of the head transmission function C of each reproduction environment information and a table of the head transmission function d of each azimuth information. The control unit 13 acquires the head transmission functions C and d in accordance with the reproduction environment information of each terminal acquired by the first acquisition unit 11 and the azimuth information of the terminal acquired by the second acquisition unit 12, thereby generating a sound image filter coefficient of each of the terminals.
The memory 2 includes a ROM and a RAM. The ROM is a nonvolatile memory. The ROM stores an activation program of the terminal and the like. The RAM is a volatile memory. The RAM is used as a work memory when, for example, the processor 1 performs processing.
The storage 3 is a storage such as a hard disk drive or a solid-state drive. The storage 3 stores various programs to be executed by the processor 1, such as an online conversation management program 31. The online conversation management program 31 is an application program that is downloaded from a predetermined download server or the like, and is a program for executing various kinds of processing pertaining to online conversation in the online conversation system. The storage 3 of each of the guest terminals GT1, GT2, and GT3 need not store the online conversation management program 31.
The voice reproduction device 4 is a device for reproducing voices. The voice reproduction device 4 according to this embodiment is a device capable of reproducing voices, and can include stereo loudspeakers, headphones, or earphones. When the voice reproduction device 4 reproduces a sound image signal that is a voice signal in which the above-described sound image filter coefficient is convoluted, a sound image is localized in a space around the head of the user. In this embodiment, the voice reproduction devices 4 of the individual terminals can be either identical or different. Also, the voice reproduction device 4 can be either a device incorporated into the terminal or an external device capable of communicating with the terminal.
The voice detection device 5 detects input of the voice of the user operating the terminal. For example, the voice detection device 5 is a microphone. The microphone of the voice detection device 5 can be either a stereo microphone or a monaural microphone. Also, the voice detection device 5 can be either a device incorporated into the terminal or an external device capable of communicating with the terminal.
The display device 6 is a display device such as a liquid crystal display or an organic EL display. The display device 6 displays various screens such as an input screen to be explained later. Also, the display device 6 can be either a display device incorporated into the terminal or an external display device capable of communicating with the terminal.
The input device 7 is an input device such as a touch panel, a keyboard, or a mouse. When the input device 7 is operated, a signal corresponding to the contents of the operation is input to the processor 1. The processor 1 performs various kinds of processing corresponding to the signal.
The communication device 8 is a communication device for allowing the terminal to mutually communicate with other terminals across the network NW. The communication device 8 can be either a communication device for wired communication or a communication device for wireless communication.
The operation of the online conversation system according to the first embodiment will be explained below.
First, the operation of the terminal HT will be explained. In step S1, the processor 1 of the terminal HT displays the screen for inputting the reproduction environment information and the azimuth information on the display device 6. Data for displaying the input screen of the reproduction environment information and the azimuth information can be stored in, e.g., the storage 3 of the terminal HT in advance.
As shown in
Also, as shown in
Referring to
In step S2, the processor 1 determines whether the user HU has input the reproduction environment information and the azimuth information or the reproduction environment information is received from the terminals GT1, GT2, and GT3. If it is determined in step S2 that the user HU has input the reproduction environment information and the azimuth information or the reproduction environment information is received from the terminals GT1, GT2, and GT3, the process advances to step S3. If it is determined in step S2 that the user HU has not input the reproduction environment information and the azimuth information or the reproduction environment information is not received from the terminals GT1, GT2, and GT3, the process advances to step S4.
In step S3, the processor 1 stores the input or received information in, e.g., the RAM of the memory 2.
In step S4, the processor 1 determines whether the information input is complete, i.e., whether the reproduction environment information and the azimuth information of each terminal are completely stored in the RAM or the like. If it is determined in step S4 that the information input is incomplete, the process returns to step S2. If it is determined in step S4 that the information input is complete, the process advances to step S5.
In step S5, the processor 1 generates a sound image filter coefficient for each terminal, i.e., for the user of each terminal, based on the reproduction environment information and the azimuth information of the terminal.
For example, a sound image filter coefficient for the user HU includes a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT1, which is input by the user GU1, and the azimuth information of the user HU, which is designated by the user HU, a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT2, which is input by the user GU2, and the azimuth information of the user HU, which is designated by the user HU, and a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT3, which is input by the user GU3, and the azimuth information of the user HU, which is designated by the user HU.
A sound image filter coefficient for the user GU1 includes a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal HT, which is input by the user HU, and the azimuth information of the user GU1, which is designated by the user HU, a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT2, which is input by the user GU2, and the azimuth information of the user GU1, which is designated by the user HU, and a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT3, which is input by the user GU3, and the azimuth information of the user GU1, which is designated by the user HU.
It is possible to similarly generate sound image filter coefficients for the users GU2 and GU3. That is, the sound image filter coefficient for the user GU2 can be generated based on the reproduction environment information of terminals except for the reproduction environment information of the voice reproduction device 4 of the terminal GT2, which is input by the user GU2, and the azimuth information of the user GU2, which is designated by the user HU. Likewise, the sound image filter coefficient for the user GU3 can be generated based on the reproduction environment information of terminals except for the reproduction environment information of the voice reproduction device 4 of the terminal GT3, which is input by the user GU3, and the azimuth information of the user GU3, which is designated by the user HU.
In step S6, the processor 1 stores the sound image filter coefficient generated for the user HU in, e.g., the storage 3. Also, the processor 1 transmits the sound image filter coefficients generated for the users GU1, GU2, and GU3 to the terminals of these users by using the communication device 8. Thus, initialization for the online conversion is complete.
In step S7, the processor 1 determines whether the voice of the user HU is input via the voice detection device 5. If it is determined in step S7 that the voice of the user HU is input, the process advances to step S8. If it is determined in step S7 that the voice of the user HU is not input, the process advances to step S10.
In step S8, the processor 1 convolutes the sound image filter coefficient for the user HU in a voice signal based of the voice of the user HU input via the voice detection device 5, thereby generating sound image signals for other users.
In step S9, the processor 1 transmits the sound image signals for the other users to the terminals GT1, GT2, and GT3 by using the communication device 8. After that, the process advances to step S13.
In step S10, the processor 1 determines whether a sound image signal is received from another terminal via the communication device 8. If it is determined in step S10 that a sound image signal is received from another terminal, the process advances to step S11. If it is determined in step S10 that no sound image signal is received from any other terminal, the process advances to step S13.
In step S11, the processor 1 separates a sound image signal for the user HU from the received sound image signal. For example, if the sound image signal is received from the terminal GT1, the processor 1 separates a sound image signal in which the sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal HT, which is input by the user HU, and the azimuth information of the user GU1, which is designated by the user HU, is convoluted.
In step S12, the processor 1 reproduces the sound image signal by the voice reproduction device 4. After that, the process advances to step S13.
In step S13, the processor 1 determines whether to terminate the online conversation. For example, if the user HU designates the termination of the online conversation by operating the input device 7, it is determined that the online conversation is to be terminated. If it is determined in step S13 that the online conversation is not to be terminated, the process returns to step S2. In this case, if the reproduction environment information or the azimuth information is changed during the online conversation, the processor 1 regenerates the sound image filter coefficient by reflecting the change, and continues the online conversation. If it is determined in step S13 that the online conversation is to be terminated, the processor 1 terminates the process shown in
Next, the operations of the terminals GT1, GT2, and GT3 will be explained. Since the operations of the terminals GT1, GT2, and GT3 are the same, the operation of the terminal GT1 will be explained below as a representative.
In step S101, the processor 1 of the terminal GT1 displays the reproduction environment information input screen on the display device 6. Data for displaying the reproduction environment information input screen can be stored in the storage 3 of the terminal GT1 in advance.
In step S102, the processor 1 determines whether the user GU1 has input the reproduction environment information. If it is determined in step S102 that the user GU1 has input the reproduction environment information, the process advances to step S103. If it is determined in step S102 that the user GU1 has not input the reproduction environment information, the process advances to step S104.
In step S103, the processor 1 transmits the input reproduction environment information to the terminal HT by using the communication device 8.
In step S104, the processor 1 determines whether the sound image filter coefficient for the user GU1 is received from the terminal HT. If it is determined in step S104 that the sound image filter coefficient for the user GU1 is not received, the process returns to step S102. If it is determined in step S104 that the sound image filter coefficient for the user GU1 is received, the process advances to step S105.
In step S105, the processor 1 stores the received sound image filter coefficient for the user GU1 in, e.g., the storage 3.
In step S106, the processor 1 determines whether the voice of the user GU1 is input via the voice detection device 5. If it is determined in step S106 that the voice of the user GU1 is input, the process advances to step S107. If it is determined in step S106 that the voice of the user GU1 is not input, the process advances to step S109.
In step S107, the processor 1 convolutes the sound image filter coefficient for the user GU1 in a voice signal based on the voice of the user GU1 input via the voice detection device 5, thereby generating sound image signals for other users.
In step S108, the processor 1 transmits the sound image signals for the other users to the terminals HT, GT2, and GT3 by using the communication device 8. After that, the process advances to step S112.
In step S109, the processor 1 determines whether a sound image signal is received from another terminal via the communication device 8. If it is determined in step S109 that a sound image signal is received from another terminal, the process advances to step S110. If it is determined in step S109 that no sound image signal is received from any other terminal, the process advances to step S112.
In step S110, the processor 1 separates a sound image signal for the user GU1 from the received sound image signal. For example, if the sound image signal is received from the terminal HT, the processor 1 separates a sound image signal in which the sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT1, which is input by the user GU1 and the azimuth information of the user HU, which is designated by the user HU, is convoluted.
In step S111, the processor 1 reproduces the sound image signal by using the voice reproduction device 4. After that, the process advances to step S112.
In step S112, the processor 1 determines whether to terminate the online conversation. For example, if the user GU1 designates the termination of the online conversation by operating the input device 7, it is determined that the online conversation is to be terminated. If it is determined in step S112 that the online conversation is not to be terminated, the process returns to step S102. In this case, if the reproduction environment information is changed during the online conversation, the processor 1 transmits this reproduction environment information to the terminal HT and continues the online conversation. If it is determined in step S112 that the online conversation is to be terminated, the processor 1 terminates the process shown in
In the first embodiment as described above, a sound image filter coefficient for the user of each terminal is generated in the host terminal HT based on the reproduction environment information and the azimuth information. Consequently, in accordance with the reproduction environment of the voice reproduction device 4 of each terminal, the sound images of other users can be localized. For example, if a plurality of users simultaneously speak, voices VA, VB, VC, and VD of the plurality of users are concentratedly heard as shown in
The generation of the sound image filter coefficient requires the reproduction environment information and the azimuth information. On the other hand, the host terminal cannot directly confirm the reproduction environment of the voice reproduction device of each guest terminal. In the first embodiment, however, each guest terminal transmits the reproduction environment information to the host terminal, and the host terminal generates a sound image filter coefficient of the terminal. As described above, the first embodiment is particularly suitable for an online conversation environment in which one terminal collectively manages the sound image filter coefficients.
In this embodiment, the host terminal generates a new sound image filter coefficient whenever acquiring the reproduction environment information and the azimuth information. However, if the host terminal and the guest terminals previously share a plurality of sound image filter coefficients that are assumed to be used, the host terminal can also determine a necessary sound image filter coefficient from the shared sound image filter coefficients whenever acquiring the reproduction environment information and the azimuth information. Instead of transmitting the sound image filter coefficient to each guest terminal, the host terminal can transmit only information of an index representing the determined sound image filter coefficient to each guest terminal. In this case, it is unnecessary to sequentially generate sound image filter coefficients during the online conversation.
Also, the first embodiment does not particularly refer to the transmission/reception of information other than voices during the online conversation. In the first embodiment, it is also possible to transmit/receive, e.g., video images other than voices.
Furthermore, the host terminal generate a sound image filter coefficient in the first embodiment. However, the host terminal does not necessarily generate a sound image filter coefficient. A sound image filter coefficient can be generated by a given guest terminal, and can also be generated by a device, such as a server, other than a terminal participating in the online conversation. In this case, the host terminal transmits, to the server or the like, the reproduction environment information and the azimuth information of each guest terminal participating in the online conversation, including the reproduction environment information acquired from each guest terminal.
The second embodiment will be explained below.
In the second embodiment, a server Sv is further connected so that the server Sv can communicate with the terminals HT, GT1, GT2, and GT3 across the network NW. In the second embodiment, the server Sv collectively performs control for localizing sound images in a space around the head of each of the users HU, GU1, GU2, and GU3 when performing the conversation using the terminals HT, GT1, GT2, and GT3. The server Sv shown in
The online conversation system of the second embodiment shown in
The processor 101 controls the overall operation of the server Sv. The processor 101 of the server Sv operates as a first acquisition unit 11, a second acquisition unit 12, a third acquisition unit 14, and a control unit 13 by executing programs stored in, e.g., the storage 103. In the second embodiment, the processor 1 of each of the host terminal HT and the guest terminals GT1, GT2, and GT3 is not necessarily operable as the first acquisition unit 11, the second acquisition unit 12, the third acquisition unit 14, and the control unit 13. The processor 101 is, e.g., a CPU. The processor 101 can also be an MPU, a GPU, an ASIC, an FPGA, or the like. The processor 101 can be a single CPU or the like, and can also be a plurality of CPUs or the like.
The first acquisition unit 11 and the second acquisition unit 12 are the same as the first embodiment, so an explanation thereof will be omitted. Also, the control unit 13 performs control for reproducing sound images at each of the terminals including the terminal HT based on reproduction environment information and azimuth information, in the same manner as explained in the first embodiment.
The third acquisition unit 14 acquires utilization information of the terminals HT, GT1, GT2, and GT3 participating in the online conversation. The utilization information is information on the utilization of sound images to be used on the terminals HT, GT1, GT2, and GT3. This utilization information contains, e.g., an attribute to be allocated to a user participating in the online conversation. In addition, the utilization information contains information of the group setting of a user participating in the online conversation. The utilization information can also contain other various kinds of information about the utilization of sound images.
The memory 102 includes a ROM and a RAM. The ROM is a nonvolatile memory. The ROM stores, e.g., an activation program of the server Sv. The RAM is a volatile memory. The RAM is used as, e.g., a work memory when the processor 101 performs processing.
The storage 103 is a storage such as a hard disk drive or a solid-state drive. The storage 103 stores various programs such as an online conversation management program 1031 to be executed by the processor 101. The online conversation management program 1031 is a program for executing various kinds of processing for the online conversation in the online conversation system.
The communication device 104 is a communication device to be used by the server Sv to communicate with each terminal across the network NW. The communication device 104 can be either a communication device for wired communication or a communication device for wireless communication.
Next, the operation of the online conversation system according to the second embodiment will be explained.
In step S201, the processor 101 transmits data of a screen for inputting the reproduction environment information and the azimuth information to the terminals HT, GT1, GT2, and GT3. That is, in the second embodiment, the input screen of the reproduction environment information and the azimuth information shown in
In step S202, the processor 101 determines whether the reproduction environment information and the azimuth information are received from the terminals HT, GT1, GT2, and GT3. If it is determined in step S202 that the reproduction environment information and the azimuth information are received from the terminals HT, GT1, GT2, and GT3, the process advances to step S203. If it is determined in step S202 that the reproduction environment information and the azimuth information are not received from the terminals HT, GT1, GT2, and GT3, the process advances to step S207.
In step S203, the processor 101 stores the received information in, e.g., the RAM of the memory 102.
In step S204, the processor 101 determines whether the input of the information is complete, i.e., whether the reproduction environment information and the azimuth information of each terminal are completely stored in the RAM or the like. If it is determined in step S204 that the input of the information is incomplete, the process returns to step S202. If it is determined in step S204 that the input of the information is complete, the process advances to step S205.
In step S205, the processor 101 generates a sound image filter coefficient for each terminal, i.e., the user of each terminal, based on the reproduction environment information and the azimuth information of the terminal.
For example, a sound image filter coefficient for the user HU includes a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT1, which is input by the user GU1, and the azimuth information of the user HU, which is designated by each of the users HU, GU1, GU2, and GU3, a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT2, which is input by the user GU2, and the azimuth information of the user HU, which is designated by each of the users HU, GU1, GU2, and GU3, and a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT3, which is input by the user GU3, and the azimuth information of the user HU, which is designated by each of the users HU, GU1, GU2, and GU3.
Also, a sound image filter coefficient for the user GU1 includes a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal HT, which is input by the user HU, and the azimuth information of the user GU1, which is designated by each of the users HU, GU2, and GU3, a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT2, which is input by the user GU2, and the azimuth information of the user GU1, which is designated by each of the users HU, GU1, GU2, and GU3, and a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT3, which is input by the user GU3, and the azimuth information of the user GU1, which is designated by each of the users HU, GU1, GU2, and GU3.
It is possible to similarly generate a sound image filter coefficient for the user GU2 and a sound image filter coefficient for the user GU3. That is, a sound image filter coefficient for the user GU2 includes a sound image filter coefficient generated based on the reproduction environment information of terminals except the reproduction environment information of the voice reproduction device 4 of the terminal GT2, which is input by the user GU2, and the azimuth information of the user GU2, which is designated by each of the users HU, GU1, GU2, and GU3. Also, a sound image filter coefficient for the user GU3 includes a sound image filter coefficient generated based on the reproduction environment information of terminals except the reproduction environment information of the voice reproduction device 4 of the terminal GT3, which is input by the user GU3, and the azimuth information of the user GU3, which is designated by each of the users HU, GU1, GU2, and GU3.
In step S206, the processor 101 transmits the sound image filter coefficients generated for the users HU, GU1, GU2, and GU3 to their terminals by using the communication device 104. Consequently, initialization for the online conversation is complete.
In step S207, the processor 101 determines whether a sound image signal is received from at least one of the terminals HT, GT1, GT2, and GT3 via the communication device 104. If it is determined in step S207 that a sound image signal is received from at least one terminal, the process advances to step S208. If it is determined in step S207 that no sound image signal is received from any terminal, the process advances to step S210.
In step S208, the processor 101 separates a sound image signal for each user from the received sound image signal. For example, if a sound image signal is received from the terminal HT, the processor 101 separates, as a sound image signal for the user GU1, a sound image signal in which a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT1, which is input by the user GU1, and the azimuth information of the user HU, which is designated by the user GU1, is convoluted. Similarly, the processor 101 separates, as a sound image signal for the user GU2, a sound image signal in which a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT2, which is input by the user GU2, and the azimuth information of the user HU, which is designated by the user GU2, is convoluted. Also, the processor 101 separates, as a sound image signal for the user GU3, a sound image signal in which a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT3, which is input by the user GU3, and the azimuth information of the user HU, which is designated by the user GU3, is convoluted.
In step S209, the processor 101 transmits each separated sound image signal to a corresponding terminal by using the communication device 104. After that, the process advances to step S210. Note that each terminal reproduces a sound image signal received in the same manner as the processing in step S12 of
In step S210, the processor 101 determines whether to terminate the online conversation. For example, if the termination of the online conversation is designated by the operations on the input devices 7 by all the users, it is determined that the online conversation is to be terminated. If it is determined in step S210 that the online conversation is not to be terminated, the process returns to step S202. In this case, if the reproduction environment information or the azimuth information is changed during the online conversation, the processor 101 regenerates a sound image filter coefficient by reflecting the change, and continues the online conversation. If it is determined in step S210 that the online conversation is to be terminated, the processor 101 terminates the process shown in
In step S301, the processor 101 transmits data of a screen for inputting the reproduction environment information and the azimuth information to the terminals HT, GT1, GT2, and GT3. Note that the processor 101 can also transmit data of a utilization information input screen to the terminals HT, GT1, GT2, and GT3.
In step S302, the processor 101 determines whether the reproduction environment information and the azimuth information are received from the terminals HT, GT1, GT2, and GT3. If it is determined in step S302 that the reproduction environment information and the azimuth information are received from the terminals HT, GT1, GT2, and GT3, the process advances to step S303. If it is determined in step S302 that the reproduction environment information and the azimuth information are not received from the terminals HT, GT1, GT2, and GT3, the process advances to step S307.
In step S303, the processor 101 stores the received information in, e.g., the RAM of the memory 102.
In step S304, the processor 101 determines whether the input of the information is complete, i.e., whether the reproduction environment information and the azimuth information of each terminal are completely stored in the RAM or the like. If it is determined in step S304 that the input of the information is incomplete, the process returns to step S302. If it is determined in step S304 that the input of the information is complete, the process advances to step S305.
In step S305, the processor 101 generates a sound image filter coefficient for each terminal, i.e., for each user based on the reproduction environment information and the azimuth information of the terminal. This sound image filter coefficient generated in step S305 can be the same as the sound image filter coefficient generated in step S205 of the first example.
In step S306, the processor 101 stores the sound image filter coefficient for each user in, e.g., the storage 103.
In step S307, the processor 101 determines whether a voice signal is received from at least one of the terminals HT, GT1, GT2, and GT3 via the communication device 104. If it is determined in step S307 that a voice signal is received from at least one terminal, the process advances to step S308. If it is determined in step S307 that no voice signal is received from any terminal, the process advances to step S310.
In step S308, the processor 101 generates a sound image signal for each user from the received voice signal. For example, if a voice is received from the terminal HT, the processor 101 generates a sound image signal for the user GU1 by convoluting, in the received voice signal, a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT1, which is input by the user GU1, and the azimuth information of the user HU, which is designated by the user GU1. Likewise, the processor 101 generates a sound image signal for the user GU2 by convoluting, in the received voice signal, a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT2, which is input by the user GU2, and the azimuth information of the user HU, which is designated by the user GU2. Also, the processor 101 generates a sound image signal for the user GU3 by convoluting, in the received voice signal, a sound image filter coefficient generated based on the reproduction environment information of the voice reproduction device 4 of the terminal GT3, which is input by the user GU3, and the azimuth information of the user HU, which is designated by the user GU3. Furthermore, if utilization information is available, the processor 101 can also adjust the generated sound image signal in accordance with the utilization information. This adjustment will be explained later.
In step S309, the processor 101 transmits each generated sound image signal to a corresponding terminal by using the communication device 104. After that, the process advances to step S310. Note that each terminal reproduces the received sound image signal in the same manner as the processing in step S12 of
In step S310, the processor 101 determines whether to terminate the online conversation. For example, if the termination of the online conversation is designated by the operations on the input devices 7 of all the users, it is determined that the online conversation is to be terminated. If it is determined in step S310 that the online conversation is not to be terminated, the process returns to step S302. In this case, if the reproduction environment information or the azimuth information is changed during the online conversation, the processor 101 regenerates a sound image filter coefficient by reflecting the change, and continues the online conversation. If it is determined in step S310 that the online conversation is to be terminated, the processor 101 terminates the process shown in
In the first example of the second embodiment, if the server, the host terminal, and the guest terminals previously share a plurality of sound image filter coefficients that are previously assumed to be used, the server can also determine a necessary sound image filter coefficient from the shared sound image filter coefficients whenever acquiring the reproduction environment information and the azimuth information. Instead of transmitting the sound image filter coefficient to the host terminal and each guest terminal, the server can transmit only information of an index representing the determined sound image filter coefficient to the host terminal and each guest terminal. In the second example of the second embodiment, the server can also determine a necessary sound image filter coefficient from a plurality of sound image filter coefficients that are previously supposed to be used whenever the reproduction environment information and the azimuth information are acquired. Then, the server can convolute the determined sound image filter coefficient in a voice signal.
In the second embodiment as explained above, the server Sv generates a sound image filter coefficient for the user of each terminal based on the reproduction environment information and the azimuth information. This can localize the sound images of other users in accordance with the reproduction environment of the voice reproduction device 4 of each terminal. Also, in the second embodiment, not the host terminal HT but the server Sv generates a sound image filter coefficient. Accordingly, the load on the host terminal HT can be reduced during the online conversation.
Furthermore, in the second embodiment, not only the host terminal HT but also the guest terminals GT1, GT2, and GT3 designate the reproduction environment information and the azimuth information, and sound image filter coefficients are generated based on these pieces of reproduction environment information and azimuth information. Therefore, each participant of the online conversation can determine sound image reproduction azimuths around the participant.
Modification 1 of the second embodiment will be explained below. In the first and second embodiments described above, the input screen including the azimuth input field 2602 shown in
This azimuth information input screen shown in
The azimuth information input screen shown in
Since the number of chairs is limited in
The azimuth information input screen shown in
Note that the arrangement of users when the number of participants in an online meeting is two or three is not limited to those shown in
Furthermore, the shape of the schematic view 2609 of a meeting table is not necessarily limited to a rectangle. For example, as shown in
It is not always necessary to use the schematic view of the meeting table shown in
Furthermore, the azimuth information can also be input on three-dimensional schematic views as shown in
Modification 2 of the second embodiment will be explained below. Modification 2 of the second embodiment is an example suitable for an online lecture, and is a practical example using utilization information.
As shown in
As shown in
For example, when the host user HU designates attributes, the processor 101 of the server Sv can adjust the reproduction of a sound image for each attribute. For example, when a voice signal of “presenter” and voice signals of other users are simultaneously input, the processor 101 can transmit only the voice of “presenter” to each terminal or localize a sound image so that the voice of “presenter” is clearly heard. The processor 101 can also transmit voices such as “mechanical sound” and “timekeeper” to only the terminal of “presenter” or localize sound images so that these voices cannot be heard on other terminals.
As shown in
The timekeeper set button 2623 is a button for performing various settings necessary for a timekeeper, such as the setting of the remaining time of the presentation, and the setting of the interval of the bell. The start button 2624 is a button that is selected when starting the presentation, and used to start timekeeping processes such as measuring the remaining time of the presentation and ringing the bell. The stop button 2625 is a button for stopping the timekeeping process. The pause/resume button 2626 is a button for switching pause/resume of the timekeeping process.
As shown in
The display screen to be displayed when the listener discussion button 2622 is selected further includes a make new group button 2630. The make new group button 2630 is selected when setting a new group not displayed in the group setting field 2628. When the make new group button 2630 is selected, the user sets, e.g., the name of the group. When making a new group, it is also possible to designate a user who is unwanted to participate in the group. For this user who is set to be unwanted to participate in the group, the processor 101 performs control so as not to display the participation button 2629 on the display screen. In
The display screen to be displayed when the listener discussion button 2622 is selected also includes a start button 2631 and a stop button 2632. The start button 2631 is a button for starting a listener discussion. The stop button 2632 is a button for stopping the listener discussion.
The display screen to be displayed when the listener discussion button 2622 is selected further includes a volume balance button 2633. The volume balance button 2633 is a button for designating the volume balance between the user as “presenter” and other users belonging to groups.
For example, when a group is set and the start button 2631 is selected, the processor 101 localizes sound images so that only users belonging to the group can hear voices. Also, the processor 101 adjusts the volume of the user as “presenter” and the volume of other users in accordance with the designation of the volume balance.
The group setting field 2628 can also be configured such that a user having initially set a group can switch active/inactive of the group. In this case, an active group and an inactive group can be displayed in different colors in the group setting field 2628.
The third embodiment will be explained below.
If the “meeting room” field 2641 is selected by the user, the processor 101 of the server Sv acquires echo data measured in advance in a small meeting room from the echo table 1032. If the “conference room” field 2642 is selected by the user, the processor 101 acquires echo data measured in advance in a large meeting room from the echo table 1032. If the “almost-echo-free room” 2643 is selected by the user, the processor 101 acquires echo data measured in advance in an anechoic room from the echo table 1032.
If the “internal member meeting” field 2645 is selected by the user, the processor 101 of the server Vs acquires echo data measured in advance in a small meeting room from the echo table 1032. If the “debrief meeting etc.” field 2646 is selected by the user, the processor 101 acquires echo data measured in advance in a large meeting room from the echo table 1032. If the “secret meeting” field 2647 is selected by the user, the processor 101 acquires echo data measured in advance in an anechoic room from the echo table 1032.
In the third embodiment as explained above, the server Sv holds echo information corresponding to the size of room, the purpose of use, and the atmosphere of meeting, in the form of a table. The server Sv adds an echo selected from the table to a voice signal for each user. This can reduce the feeling of fatigue when the voices of individual users are heard by the same volume level.
In the third embodiment, the echo table contains three types of echo data. However, the echo table can also contain one or two types of echo data or four or more types of echo data.
In the third embodiment, the storage 103 can further store a level attenuation table 1033. The level attenuation table 1033 has level attenuation data corresponding to the distance of a sound volume measured in advance in an anechoic room, as table data. In this case, the processor 101 of the server Sv acquires level attenuation data corresponding to a virtual distance between the user and a virtual sound source in which a sound image is supposed to be used, and adds level attenuation corresponding to the acquired level attenuation data to a sound image signal. This can also reduce the feeling of fatigue when the voices of individual users are heard by the same volume level.
While certain embodiments have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the inventions. Indeed, the novel embodiments described herein may be embodied in a variety of other forms; furthermore, various omissions, substitutions and changes in the form of the embodiments described herein may be made without departing from the spirit of the inventions. The accompanying claims and their equivalents are intended to cover such forms or modifications as would fall within the scope and spirit of the inventions.
Number | Date | Country | Kind |
---|---|---|---|
2021-151457 | Sep 2021 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
5594800 | Gerzon | Jan 1997 | A |
5757927 | Gerzon | May 1998 | A |
5812674 | Jot | Sep 1998 | A |
6021205 | Yamada | Feb 2000 | A |
9088854 | Enamito et al. | Jul 2015 | B2 |
20060045276 | Gamo | Mar 2006 | A1 |
20090002477 | Cutler | Jan 2009 | A1 |
20090238371 | Rumsey | Sep 2009 | A1 |
20090252356 | Goodwin | Oct 2009 | A1 |
20120250869 | Ohashi | Oct 2012 | A1 |
20130202116 | Par | Aug 2013 | A1 |
20130336490 | Someda | Dec 2013 | A1 |
20150086023 | Enamito et al. | Mar 2015 | A1 |
20150296086 | Eckert et al. | Oct 2015 | A1 |
20150350788 | Someda et al. | Dec 2015 | A1 |
20170092298 | Nakamura | Mar 2017 | A1 |
Number | Date | Country |
---|---|---|
2006-74386 | Mar 2006 | JP |
2008-160397 | Jul 2008 | JP |
2012-212982 | Nov 2012 | JP |
2013-51631 | Mar 2013 | JP |
2014-17813 | Jan 2014 | JP |
2015-65541 | Apr 2015 | JP |
5944567 | Jul 2016 | JP |
6255076 | Dec 2017 | JP |
6407568 | Oct 2018 | JP |
Entry |
---|
Kashiyama, K. “Railway noise evaluation system using visualization and audibility by VR technology” Noise Control vol. 44, No. 4, 2020 (15 pages) with Machine Translation. |
Okumura, H. “Sonification of sound using 3D sound technology ViReaITM” Noise Control vol. 44, No. 4, 2020 (15 pages) with Machine Translation. |
Kondo, K. “Examination of virtual audio conferencing client using enhanced acoustic reality” Engineering and Technical Research Survey on Telecommunications Technology 10-01027, Graduate School of Science and Engineering, Yamagata University, 2012 (31 pages) with Machine Translation. |
Number | Date | Country | |
---|---|---|---|
20230078804 A1 | Mar 2023 | US |