The present disclosure relates to the technical field of screen sound generation, and more specifically, to an audio control method, a control device, a driving circuit and a readable storage medium.
With the continuous development of display technology, the display screen is becoming larger and larger, and the mismatch between the supporting sound system and the large screen image is becoming more and more serious, which cannot achieve the audio-visual effect of sound and picture integration, and reduces user experience. In some application scenarios, the screen sound generation technology alleviates this problem by arranging multiple speakers under the display screen. However, the existing screen sound generation technology is only based on the traditional two-channel and three-channel audio playback technology, which is difficult to further improve the audio-visual effect of the sound and picture integration.
Some embodiments of the present disclosure provide an audio control method, a control device, a driving circuit and a readable storage medium for improving the sound and picture integration effect of the screen sound system.
According to an aspect of the present disclosure, an audio control method is provided, the method is applicable to a display screen configured with M speakers, wherein, M is an integer greater than or equal to 2, and the method comprises: obtaining a sound image coordinate of a sound object relative to the display screen; determining N speakers from the M speakers as loudspeakers according to the sound image coordinate and position coordinates of the M speakers relative to the display screen, wherein N is an integer less than or equal to M; determining output gains of the N loudspeakers according to distances between the N loudspeakers and a viewer of the display screen and sound attenuation coefficients; and calculating output audio data of the sound object in the display screen according to audio data of the sound object and the output gains of the N loudspeakers, and controlling the M speakers to play the output audio data.
According to some embodiments of the present disclosure, determining the N speakers from the M speakers as the loudspeakers comprises: calculating distances between the position coordinates of the M speakers and the sound image coordinate, and determining 3 speakers with nearest distances as the loudspeakers, wherein N=3.
According to some embodiments of the present disclosure, determining the output gains of the N loudspeakers according to the distances between the N loudspeakers and the viewer of the display screen and the sound attenuation coefficients comprises: obtaining N vectors pointed from the viewer to the N loudspeakers; updating vector modulus of the N vectors based on differences between the vector modulus of the N vectors, and using a vector-base amplitude panning algorithm to calculate N initial gains based on updated N vectors; and obtaining N sound attenuation coefficients respectively based on the vector modulus of the N vectors, and obtaining N output gains based on a product of the N sound attenuation coefficients and the N initial gains.
According to some embodiments of the present disclosure, updating the vector modulus of the N vectors based on the differences between the vector modulus of the N vectors, and using the vector-base amplitude panning algorithm to calculate the N initial gains based on the updated N vectors, comprises: determining a loudspeaker with a largest vector modulus among the N vectors of the N loudspeakers, wherein the loudspeaker with the largest vector modulus is represented as a first loudspeaker, a vector modulus of the first loudspeaker is represented as a first vector modulus, and loudspeakers other than the first loudspeaker among the N loudspeakers are represented as second loudspeakers; obtaining extended vectors based on vector directions of the second loudspeakers and the first vector modulus; and calculating N initial gains based on a vector of the first loudspeaker and the extended vectors of the second loudspeakers according to the vector-base amplitude panning algorithm.
According to some embodiments of the present disclosure, obtaining the N sound attenuation coefficients respectively based on the vector modulus of the N vectors comprises: for each of the second loudspeakers, calculating a difference d between vector modulus of the second loudspeakers and the first vector modulus, and calculating a sound attenuation coefficient k according to k=20 log(10, d) based on the difference d; and setting a sound attenuation coefficient of the first loudspeaker to be 0.
According to some embodiments of the present disclosure, the M speakers are equally spaced in the display screen in a form of matrix.
According to some embodiments of the present disclosure, calculating the output audio data of the sound object in the display screen according to the audio data of the sound object and the output gains of the N loudspeakers, and controlling the M speakers to play the output audio data, comprises: setting output gains of speakers other than the N loudspeakers among the M speakers to be 0; and multiplying the audio data with output gains of the M speakers respectively, to obtain output audio data comprising M audio components, and controlling the M speakers to output one of corresponding M audio components respectively.
According to some embodiments of the present disclosure, multiplying the audio data with the output gains of the M speakers respectively comprises: delaying the audio data for a predetermined time interval, and multiplying delayed audio data with the output gains of the M speakers.
According to some embodiments of the present disclosure, obtaining the sound image coordinate of the sound object relative to the display screen comprise: making video data comprising the sound object, wherein the sound object is controlled to move, wherein the display screen is used to output the video data; and recording moving track of the sound object to obtain the sound image coordinate.
According to another aspect of the present disclosure, an audio control device is provided, the device is applicable for a display screen equipped with M speakers, M is an integer greater than or equal to 2, and the device comprises: a sound image coordinate unit which is configured to obtain a sound image coordinate of a sound object relative to the display screen; a coordinate comparison unit which is configured to determine N speakers from the M speakers as loudspeakers according to the sound image coordinate and position coordinates of the M speakers relative to the display screen, wherein, N is an integer less than or equal to M; a gain calculation unit which is configured to determine output gains of the N loudspeakers according to distances between the N loudspeakers and a viewer of the display screen and sound attenuation coefficients; and an output unit which is configured to calculate output audio data of the sound object in the display screen according to audio data of the sound object and the output gains of the N loudspeakers, and controlling the M speakers to play the output audio data.
According to some embodiments of the present disclosure, determining the N speakers from the M speakers as the loudspeakers by the coordinate comparison unit comprises: calculating distances between the position coordinates of the M speakers and the sound image coordinate, and determining 3 speakers with nearest distances as the loudspeakers, wherein N=3.
According to some embodiments of the present disclosure, determining the output gains of the N loudspeakers according to the distances between the N loudspeakers and the viewer of the display screen and the sound attenuation coefficients by the gain calculation unit comprises: obtaining N vectors pointed from the viewer to the N loudspeakers; updating vector modulus of the N vectors based on differences between the vector modulus of the N vectors, and using a Vector-Base Amplitude Panning (VBAP) algorithm to calculate N initial gains based on updated N vectors; and obtaining N sound attenuation coefficients respectively based on the vector modulus of the N vectors, and obtaining N output gains based on a product of the N sound attenuation coefficients and the N initial gains.
According to some embodiments of the present disclosure, updating the vector modulus of the N vectors based on the differences between the vector modulus of the N vectors, and using the vector-base amplitude panning algorithm to calculate the N initial gains based on the updated N vectors by the gain calculation unit, comprises: determining a loudspeaker with a largest vector modulus among the N vectors of the N loudspeakers, wherein the loudspeaker with the largest vector modulus is represented as a first loudspeaker, a vector modulus of the first loudspeaker is represented as a first vector modulus, and loudspeakers other than the first loudspeaker among the N loudspeakers are represented as second loudspeakers; obtaining extended vectors based on vector directions of the second loudspeakers and the first vector modulus; and calculating N initial gains based on a vector of the first loudspeaker and the extended vectors of the second loudspeakers according to the vector amplitude translation algorithm.
According to some embodiments of the present disclosure, obtaining the N sound attenuation coefficients respectively based on the vector modulus of the N vectors by the gain calculation unit comprises: for each of the second loudspeakers, calculating a difference d between vector modulus of the second loudspeakers and the first vector modulus, and calculating a sound attenuation coefficient k according to k=20 log(10, d) based on the difference d; and setting a sound attenuation coefficient of the first loudspeaker to be 0.
According to some embodiments of the present disclosure, the M speakers are equally spaced in the display screen in a form of matrix.
According to some embodiments of the present disclosure, calculating the output audio data of the sound object in the display screen according to the audio data of the sound object and the output gains of the N loudspeakers, and controlling the M speakers to play the output audio data by the output unit comprises: setting output gains of speakers other than the N loudspeakers among the M speakers to be 0; and multiplying the audio data with output gains of the M speakers respectively, to obtain output audio data comprising M audio components, and controlling the M speakers to output one of corresponding M audio components respectively.
According to some embodiments of the present disclosure, multiplying the audio data with the output gains of the M speakers respectively by the output unit comprises: delaying the audio data for a predetermined time interval, and multiplying delayed audio data with the output gains of the M speakers.
According to some embodiments of the present disclosure, obtaining the sound image coordinate of the sound object relative to the display screen by the sound image coordinate unit comprises: making video data comprising the sound object, wherein the sound object is controlled to move, wherein the display screen is used to output the video data; and recording moving track of the sound object to obtain the sound image coordinate.
According to another aspect of the present disclosure, a driving circuit based on multi-channel splicing screen sound system is provided, the driving circuit comprises: a multi-channel sound card which is configured to receive sound data, wherein the sound data comprises sound channel data and sound image data, wherein, the sound image data comprises audio data and a coordinate of a sound object; an audio control circuit which is configured to obtain output audio data of the sound object in the display screen according to the audio control method described above; and a sound standard unit, wherein the sound standard unit comprises a power amplifier board and a screen sound components, and the sound standard unit is configured to output the channel data and the output audio data.
According to another aspect of the present disclosure, a non-volatile computer-readable storage medium is provided, on which instructions are stored, wherein the instructions causes the processor to execute the audio control method described above when executed by the processor.
Using the audio control method, control device, driving circuit and readable storage medium according to some embodiments of the present disclosure, the position of loudspeakers can be accurately determined according to the acoustic image coordinate of the sound object and the coordinates of a plurality of speakers, and further, gains of determined loudspeakers can be adjusted according to the position of the viewer and the sound attenuation coefficients, so as to improve the audio-visual effect of the sound and picture integration on the large screen, which can better realize the surround stereo experience for the sound object, and help to improve the viewing experience of large-screen users.
In order to more clearly illustrate the technical solutions in the embodiments of the present disclosure or those in the prior art, the following will briefly introduce the drawings used in the embodiments or the description of the prior art. It is obvious that the drawings in the following description are only some embodiments of the present disclosure. For those skilled in the art, other drawings can also be obtained from these drawings without creative work.
The technical solution in the embodiment of the present disclosure will be described clearly and completely below in combination with the drawings in the embodiment of the present disclosure. Obviously, the described embodiments are only a part of the embodiments of the present disclosure, not all of them. Based on the embodiments in the present disclosure, all other embodiments obtained by those skilled in the art without creative work fall within the protection scope the present disclosure.
The “first”, “second” and similar words used in the present disclosure do not indicate any order, quantity or importance, but are only used to distinguish different components. Similarly, similar terms such as “comprising” or “comprise” mean that the elements or objects appearing before the word cover the elements or objects listed after the word and their equivalents, without excluding other elements or objects. Similar terms such as “connecting” or “connection” are not limited to physical or mechanical connections, but can comprise electrical connections, whether direct or indirect.
A flowchart is used in the present disclosure to illustrate the steps of the method according to the embodiment of the present disclosure. It should be understood that the previous or subsequent steps are not necessarily carried out in order. Instead, various steps can be processed in reverse order or at the same time. At the same time, other operations can also be added to these processes. It can be understood that the professional terms and phrases involved in this article have the meanings known to those skilled in the art.
With the rapid development of display technology, the size of display screen is becoming larger and larger, which is used to meet the needs of application scenarios such as large-scale exhibitions for example. For the display screen with large display size, the mismatch between the supporting sound system and the large screen display is becoming more and more serious, and the playback effect of the sound and picture integration cannot be achieved. Specifically, the sound and picture integration can mean that the display pictures of the display screen are consistent with the played sound, or it can be called sound and picture synchronization. The display effect of the sound and picture integration can enhance the realism of pictures and improve the appeal of visual images.
Screen sound generation technology is used to solve the technical problem that large display screen is difficult to achieve the sound and picture integration. However, the existing screen sound generation technology still relies on the traditional two-channel or three-channel technology, but this technology does not completely solve the problem that the sound and picture cannot be integrated in applications of large screen size. Therefore, more accurate sound positioning system and more screen loudspeakers are needed to achieve the sound and picture integration. The existing screen sound system does not meet the scheme of multi-channel circuit driving. Although it can be spliced according to the scheme of two-channel circuit driving, this kind of splicing can only achieve an increase in number. There is no way to control the sound position and sound effect in real time according to film source contents to achieve better sound and picture integration effect.
Some embodiments of the present disclosure propose an audio control method, which is applicable to a display screen configured with multiple speakers. For example, speakers can be arranged below the display screen in an array structure to solve the problem that the sound and picture of the multi-channel screen cannot be integrated. As an example, the audio control method according to some embodiments of the present disclosure can be implemented in the multi-channel screen sound generation driving circuit to carry out audio driving control for the display screen configured with multiple under-screen speakers. Specifically, the audio control method can control the number and position of speakers in real time according to the position of the sound object, and control output gains of loudspeakers, to achieve better audio-visual experience. In addition, the audio control method according to the embodiment of the present disclosure can also be combined with an audio splicing unit to realize channel splicing, and can splice any number of channels in real time according to user needs.
Specifically, in step S101, the acoustic image coordinate of the sound object relative to the display screen can be understood as the coordinate of the sound object in the coordinate system relative to the display screen. For example, as shown in
Next, as shown in
According to some embodiments of the present disclosure, the steps to determine the loudspeakers can comprise: calculating distances between the position coordinates and the acoustic image coordinates of M speakers, and determining 3 speakers with nearest distances as loudspeakers, where N=3. In these embodiments, loudspeakers are selected based on distances, and the 3 speakers that are closest to the sound object are determined as the loudspeakers. It can be understood that the number of loudspeakers can also be other values.
Next, in step S103, output gains of N loudspeakers are determined respectively according to the distances between N loudspeakers and the viewer of the display screen and the sound attenuation coefficients. And, in step S104, the output audio data of the sound object in the display screen is calculated according to the audio data of the sound object and the output gains of N loudspeakers, and M speakers are controlled to play the output audio data.
According to the embodiment of the present disclosure, after determining N loudspeakers according to the distances, the gains of the loudspeakers are finely adjusted by further taking into account the position of the viewer relative to the display screen and the attenuation change of the sound. For example, the gains of N loudspeakers are set to different values, even if the sound intensities of the loudspeakers at different positions of the sound object are different, the audio-visual effect of the sound and picture integration is strengthened. The specific process of calculating output gains will be described in detail below.
In order to clearly understand the process of determining the output gains of the loudspeakers in the audio control method according to the embodiment of the present disclosure, the implementation process of the Vector-Base Amplitude Panning (VBAP) algorithm is first introduced. The VBAP algorithm is a method used to reproduce the 3D stereo effect by using multiple speakers and based on the position of the sound object in a 3D stereo scenario. According to the VBAP algorithm, 3 speakers can be used to reproduce the sound object, where the gain of each speaker corresponds to the position of the sound object.
As an example,
In addition, the position vector P indicating the position of the sound object is expressed as P=[P1, P2, P3], where P1, P2 and P3 represent the three-dimensional coordinates of the sound object respectively. Similarly, vectors L1, L2 and L3 can be expressed as L1=[L11, L12, L13], L2=[L21, L22, L23], L3=[L31, L32, L33], respectively.
Assuming that the gains of the 3 speakers corresponding to the position vectors L1, L2 and L3 are expressed as g1, g2 and g3 respectively, the following formula (1) should be met:
Therefore, according to the following formula (2), the gain of each speaker can be calculated from the position vector P of the sound object and the position vectors L1, L2 and L3 of the speaker.
After calculating the gain of each speaker, the audio signal of the sound object is multiplied with gains respectively and the results are played, so that the listener can obtain stereo surround effect. It can be understood that in the VBAP algorithm shown in
According to some embodiments of the present disclosure, determining the output gains of N loudspeakers respectively according to the distances between N loudspeakers and the viewer of the display screen and the sound attenuation coefficients (S103), comprises: S1031, obtaining N vectors pointing from the viewer to N loudspeakers; S1032, updating the vector modulus of N vectors based on the differences between the vector modulus of N vectors, and using the VBAP algorithm to calculate N initial gains based on the updated N vectors; S1033, obtaining N sound attenuation coefficients based on the vector modulus of N vectors respectively, and obtaining N output gains based on the products of N sound attenuation coefficients and N initial gains. Specifically, N=3 will be described as an example.
In order to facilitate understanding,
In the above step S1031, 3 vectors of the selected t3 loudspeakers will be obtained first, as shown in
In step S1032 above, the vector modulus of 3 vectors are updated based on the differences between the vector modulus of 3 vectors, and the VBAP algorithm shown in formula (2) above is used to calculate 3 initial gains based on the updated 3 vectors. Specifically, the process of obtaining initial gains can be described as steps: S10321, determining the loudspeaker with the largest vector modulus among N vectors of N loudspeakers, wherein, the loudspeaker with the largest vector modulus is represented as the first loudspeaker, and the vector modulus of the first loudspeaker is represented as the first vector modulus, the loudspeakers other than the first loudspeaker among N loudspeakers are represented as the second loudspeakers. For example, referring to
Next, S10322, obtaining the extended vector based on the vector direction of the second loudspeaker and the first vector modulus. That is to say, for speaker 1 and speaker 3 which are close to the listener, the modulus of their vectors are extended until the distances between them and the listener are equal to the distance between speaker 2 and the listener, and the vector directions are unchanged. Therefore, the distances between the extended speaker 1 and the listener is the same with the distance between the extended speaker 3 and the listener and the distance between speaker 2 and the listener, that is, all these distances are equal to the vector modulus R2, so that the position relationship between the updated speakers 1-3 and the listener meets the spherical relationship shown in
Next, S1033, calculating N initial gains based on the vector of the first loudspeaker and the extended vectors of the second loudspeakers according to the VBAP algorithm. The process of calculating the initial gains can be carried out with reference to the above formula (2).
According to the embodiment of the present disclosure, after obtaining the initial gains, the sound attenuation coefficients will also be calculated for the loudspeakers, and the calculated initial gains will be adjusted based on the sound attenuation coefficients. According to some embodiments of the present disclosure, obtaining N sound attenuation coefficients based on the vector modulus of N vectors comprises: for each of the second loudspeakers, calculating the difference d between the vector modulus of the second loudspeakers and the first vector modulus, and calculating the sound attenuation coefficient k according to k=20 log(10, d) based on the difference d. Specifically, in the example of
It can be understood that in the process of calculating the initial gains above, the vector modulus of speaker 1 and speaker 3 have been extended, so that the calculated initial gains do not conform to the real position relationship between the speakers and the screen. Therefore, the sound attenuation will be calculated for the extended speakers, and the initial gains will be adjusted based on the calculated sound attenuation information to obtain the final output gain, which can make the audio playback effect of the 3 loudspeakers more satisfying for the audio-visual experience of the sound and picture integration.
According to some embodiments of the present disclosure, calculating the output audio data of the sound object in the display screen according to the audio data of the sound object and the output gains of the N loudspeakers, and controlling M speakers to play the output audio data (S104) comprises: setting the output gains of the speakers other than the N loudspeakers in the M speakers to 0; and multiplying the audio data with the output gains of M speakers respectively to obtain the output audio data comprising M audio components, and controlling the M speakers to output one of the corresponding M audio components respectively.
As an example, for the 32 speakers in the display screen as shown in
In formula (3), Audio1 represents the audio data of the sound object, and gains Gain1_1 to Gain1_32 represent the output gains of 32 speakers in the display screen respectively, wherein only the output gains of the selected loudspeakers have specific values, while the output gains of other speakers are 0. After the multiplication process of formula (3) above, the audio components corresponding to 32 speakers respectively will be obtained for playback.
According to some embodiments of the present disclosure, before multiplying the audio data with the output gains of M speakers respectively, the audio data of the sound object can also be delayed for a predetermined time interval, and the delayed audio data can be multiplied with the output gains of M speakers. In actual applications, the acoustic image coordinate and audio data of the sound object are obtained synchronously, and a certain time delay is generated in the process of calculating the output gains according to the above steps S102-S103. Therefore, the synchronously received audio data can be delayed for a certain time interval to avoid the phenomenon of non-synchronization.
In the method according to the present disclosure, the information of the sound object is processed, and the information of the sound object is divided into audio data (Audio) and position information. For example, position information and audio data Audio can be obtained synchronously. For example, the audio control method according to the embodiment of the present disclosure can be implemented in the audio control circuit, and the audio control will simultaneously receive audio data and position information for a certain audio object. For the position information, it can be expressed as the acoustic image coordinate of the sound object relative to the display screen.
As shown in
According to some embodiments of the present disclosure, in order to adapt to the above changes in location information, the acoustic image coordinate module can first detect the sampling frequency (Fs) of the audio data Audio, and then judge whether the audio data Audio and acoustic image coordinate are synchronously input. If no new acoustic image coordinate is input, one or more speakers located in the center of the screen can be selected by default for sound generation. For example, if there is no background sound corresponding to the sound object, two speakers at the center of the screen can be directly selected to play audio data without having to carry out the audio control algorithm described above used to achieve the sound and picture integration. Until an acoustic image coordinate is detected, the acoustic image coordinate module can transmit the received acoustic image coordinate to the subsequent distance comparison process, and store the currently received acoustic image coordinate in the buffer. After receiving the next frame of audio data Audio, if a new acoustic image coordinate is received at the same time, the new acoustic image coordinate will be transferred to the distance comparison module, and the coordinate stored in the buffer will be refreshed at the same time. If no new acoustic image coordinate is received, the acoustic image coordinate stored in the buffer will be transferred to the distance comparison module at the back end.
After receiving the acoustic image coordinate, the distance comparison module can calculate the distances between the acoustic image coordinate and 32 pre-stored speaker coordinates respectively to obtain 32 distances, then compare them, and select 3 speakers with the smallest distances as the loudspeakers. In addition, if two distances are the same, choose either. Next, the output gains of the loudspeakers are determined respectively based on the speaker coordinates of the selected 3 loudspeakers and the acoustic image coordinate, and the output gains of remaining 29 speakers are set to zero. Then, a gain matrix can be obtained based on the output gains of 32 speakers. The gain matrix comprises the output gain of each speaker.
For received audio data, delay processing can be performed first to offset the time consumption of the above gain calculation. Then, the received audio data enters a mixing module Mixture for processing to obtain 32 audio components Audio1_1˜Audio1_32. The process of calculating the audio components can refer to the above formula (3).
The implementation process of the audio control method according to the embodiment of the present disclosure is described above for the case of one sound object. It can be understood that the audio control method according to the embodiment of the present disclosure can also be applied to a scenario of multiple sound objects, that is, according to the acoustic image coordinate and audio data of each sound object, the steps S101-S104 described above are carried out respectively, so as to play audio for different sound objects, which will not be repeated here.
According to some embodiments of the present disclosure, obtaining the acoustic image coordinate of the sound object relative to the display screen comprises: making video data comprising the sound object during, wherein the sound object is controlled to move, wherein the display screen is used to output video data; and recording the moving track of the sound object to obtain the acoustic image coordinate.
In some implementations, the video data comprising the sound object can be obtained based on programming software, and the audio data and acoustic image coordinate of the sound object can be recorded during the production process, so as to apply to the audio control method provided according to some embodiments of the present disclosure.
First of all, the layout of speakers can be drawn in a designed GUI interface and the coordinates of 32 speakers can be obtained, which will be used for the selection of loudspeakers. Next, a sound object, such as the helicopter shown in
By using the process shown in
The audio components that are obtained through audio control will enter a multi-channel sound card for playback. Specifically, the multi-channel sound card can be connected with 32 sound standard units, which correspond to the screen sound components in the display screen. For example, each sound standard unit can comprise an audio receiving format conversion unit, a digital to analog converter (DAC), a power amplifier board and other structures, which is not limited here.
Compared with the video data produced by programming software described above in combination with
As an implementation, the audio control method according to the embodiment of the present disclosure can also be configured to build a player software. As an example, a player software with 32 channels can be developed.
As shown in
As an implementation, the audio control method according to the embodiment of the present disclosure can also be applied to entertainment products, such as game scene playback. In the game scene, there are many kinds of sound objects, such as blasting sound, prompt sound, scene effect sound, etc. These sound objects have corresponding position coordinates in the game design process.
As an implementation method, the audio control method according to the embodiment of the present disclosure can also be applied to an integrated circuit (IC) to realize the real-time driving control of the acoustic image.
As shown in
Specifically, the standard unit box can comprise, for example, a power amplifier board and a speaker. For example, the number of standard unit boxes can be 32. Ethernet interface can be selected for audio interface, because other audio digital interfaces such as Inter-IC sound (IIS) protocol cannot realize long-distance transmission, and the transmission rate is low, which cannot realize real-time transmission of multi-channel data. Therefore, Ethernet interface and network cable are preferred for audio data transmission. The played sound data can be the audio data corresponding to the sound object or the channel data. The data format of sound data is shown as
In addition, considering that the frequency of pos data is generally 60 Hz or 120 Hz, while the frequency of audio data is generally 48 KHz, not every frame of audio data packet is configured with a pos data packet, and 800 or 400 frames of audio are required to be configured with a frame of pos data packet. As an example, configuring according to 400, which can make the transmission speed of audio data faster and save resources. As for channel selection, as shown in
Continue to refer to
If the acoustic image data is received, it can be transmitted to channels 33-64, and each sound object can occupy one channel. The channels 33-64 of these 32 sound objects can be output to the audio processing unit, and the audio control method provided according to some embodiments of the present disclosure is implemented in this audio processing unit.
Specifically, referring to
As described above, the frequency of pos data is generally 60 Hz or 120 Hz, while the frequency of audio data is generally 48 KHz, so not every frame of audio data packet is configured with one pos data packet, and 800 or 400 frames of audio are required to be configured with one frame of pos data packet. As an example, 400 frames of audio are configured with one frame of pos data packet. In this case, in the data separation unit, it will be controlled by a 9-bit counter. When the counter counts to 0-399, it will transmit the position data to a pos register, and at other times, it will output the audio data Audio to the back end. The pos register is set because the general amount of pos data is small, and the back end needs to get the same number of pos data packets as that of the audio data, which is stored in the pos register, so that each frame of audio data Audio gets a corresponding pos data in the pos register.
The audio data and acoustic image coordinate pos enter into the audio control unit in
The audio control method provided according to the embodiment of the present disclosure has been described in detail above in combination with various implementation methods. It can be understood that the audio control method can also be applied to other scenarios, which will not be repeated here.
By using the audio control method according to some embodiments of the present disclosure, the positions of loudspeakers can be accurately determined according to the acoustic image coordinate of the sound object and the coordinates of multiple speakers, and further, the gains of the determined loudspeakers can be adjusted according to the position of the viewer and the sound attenuation coefficients, so as to improve the audio-visual effect of the sound and picture integration on the large screen, which can better realize the surround stereo effect for the sound object, and help improve the viewing experience of large-screen users.
According to another aspect of the present disclosure, an audio control device is also provided.
As shown in
According to some embodiments of the present disclosure, the acoustic image coordinate unit 1010 can be configured to obtain the acoustic image coordinate of the sound object relative to the display screen; the coordinate comparison unit 1020 can be configured to determine N speakers from M speakers as loudspeakers according to the acoustic image coordinates and the position coordinates of M speakers relative to the display screen, where N is an integer less than or equal to M; the gain calculation unit 1030 can be configured to determine the output gains of N loudspeakers respectively according to the distances between the N loudspeakers and the viewer of the display screen and the sound attenuation coefficients; and the output unit 1040 can be configured to calculate the output audio data of the sound object in the display screen according to the audio data of the sound object and the output gains of N loudspeakers, and control M speakers to play the output audio data.
According to some embodiments of the present disclosure, determining N speakers from M speakers as loudspeakers by the coordinate comparison unit 1020 comprises: calculating the distances between the position coordinates and the acoustic image coordinates of M speakers respectively, and determining the 3 speakers with nearest distances as loudspeakers, where N=3.
According to some embodiments of the present disclosure, determining the output gains of the N loudspeakers respectively according to the distances between the N loudspeakers and the viewer of the display screen and the sound attenuation coefficients by the gain calculation unit 1030 comprises: obtaining N vectors pointing from the viewer to the N loudspeakers; updating the vector modulus of N vectors based on the differences between the vector modulus of N vectors, and using the VBAP algorithm to calculate N initial gains based on the updated N vectors; obtaining N sound attenuation coefficients based on the vector modulus of N vectors, and obtaining N output gains based on the products of N sound attenuation coefficients and N initial gains.
According to some embodiments of the present disclosure, updating the vector modulus of N vectors based on the differences between the vector modulus of N vectors, and using the VBAP algorithm to calculate the N initial gains based on the updated N vectors by the gain calculation unit 1030 comprises: determining the loudspeaker with the largest vector modulus among the N vectors of N loudspeakers, wherein, the loudspeaker with the largest vector modulus is expressed as the first loudspeaker, the vector modulus of the first loudspeaker is expressed as the first vector modulus, and the loudspeakers other than the first loudspeaker among the N loudspeakers are expressed as the second loudspeakers; obtaining an extended vector based on the vector direction of the second loudspeakers and the first vector modulus; and calculating N initial gains based on the vector of the first loudspeaker and the extended vector of the second loudspeakers according to the VBAP algorithm.
According to some embodiments of the present disclosure, obtaining N sound attenuation coefficients respectively based on the vector modulus of N vectors by the gain calculation unit 1030 comprises: for each of the second loudspeakers, calculating the difference r between the vector modulus of the second loudspeakers and the first vector modulus, and calculating the sound attenuation coefficient k based on the difference r according to k=20 log(10, r); and setting the sound attenuation coefficient of the first loudspeaker to 0. Specifically, the process of calculating the output gains of the loudspeakers by the gain calculation unit can refer to the above description in combination with
According to some embodiments of the present disclosure, M speakers are equally spaced in the display screen in a form of matrix.
According to some embodiments of the present disclosure, calculating the output audio data of the sound object in the display screen according to the audio data of the sound object and the output gains of N loudspeakers and controlling M speakers to play the output audio data by the output unit 1040, comprises: setting the output gains of the speakers other than the N loudspeakers in the M speakers to 0; and multiplying the audio data with the output gains of M speakers respectively to obtain the output audio data comprising M audio components, and controlling the M speakers to output one of the corresponding M audio components.
According to some embodiments of the present disclosure, multiplying the audio data with the output gains of M speakers respectively by the output unit 1040 comprises: delaying the audio data for a predetermined time interval, and multiplying the delayed audio data with the output gains of M speakers.
According to some embodiments of the present disclosure, obtaining the acoustic image coordinate of the sound object relative to the display screen by the acoustic image coordinate unit 1010, comprises: making video data comprising the sound object, wherein the sound object is controlled to move, wherein the display screen is used for outputting video data; and recording the moving track of the sound object to obtain the acoustic image coordinate. Specifically, the acoustic image coordinate unit 1010 can realize the steps described above in combination with
As an example, the above audio control device can be implemented as a circuit structure shown in
According to another aspect of the present disclosure, a driving circuit based on a multi-channel splicing screen sound system is also provided.
According to some embodiments of the present disclosure, the multi-channel sound card 2010 can be configured to receive sound data, wherein the sound data comprises channel data and acoustic image data, wherein the acoustic image data comprises audio data and the coordinate of the sound object. The audio control circuit 2020 can be configured to obtain the output audio data of the sound object in the display screen according to the audio control method described above. The sound standard unit 2030 can comprise a power amplifier board and a screen sound components. The sound standard unit can be configured to output channel data and output audio data. For the specific implementation structure of the driving circuit, please refer to the above description of
As an implementation method,
According to another aspect of the present disclosure, a non-volatile computer-readable storage medium is also provided, on which instructions are stored. The instructions cause the processor to execute the audio control method described above when executed by the processor.
As shown in
Those skilled in the art can understand that the contents disclosed in the present disclosure may have many variations and improvements. For example, the various devices or components described above can be implemented by hardware, software, firmware, or some or all of the three.
In addition, although the present disclosure makes various references to some units in the system according to the embodiments of the present disclosure, any number of different units can be used and run on clients and/or servers. The units are only illustrative, and different aspects of the system and method can use different units.
Those skilled in the art can understand that all or part of the steps in the above method can be completed by instructing the relevant hardware through a program, and the program can be stored in a computer-readable storage medium, such as read-only memory, magnetic disk or optical disk. Optionally, all or part of the steps of the above embodiments can also be implemented by using one or more integrated circuits. Accordingly, each module/unit in the above embodiments can be implemented in the form of hardware or software function modules. The present disclosure is not limited to any combination of specific forms of hardware and software.
Unless otherwise defined, all terms (comprising technical and scientific terms) used herein have the same meaning as those commonly understood by those skilled in the art to which this disclosure belongs. It should also be understood that terms such as those defined in the general dictionary should be interpreted as having the meaning consistent with their meaning in the context of the relevant technology, and should not be interpreted in the sense of idealization or extreme formalization, unless explicitly defined here.
The above is a description of the present disclosure and should not be considered as a limitation thereof. Although several exemplary embodiments of the present disclosure have been described, those skilled in the art will easily understand that many modifications can be made to the exemplary embodiments without departing from the novel teaching and advantages of the present disclosure. Therefore, all these modifications are intended to be comprised in the scope of the disclosure defined by the claims. It should be understood that the above is a description of the present disclosure, and should not be considered as limited to the specific embodiments disclosed, and the modification intention of the disclosed embodiments and other embodiments is comprised in the scope of the appended claims. The present disclosure is limited by the claims and their equivalents.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/CN2022/096380 | 5/31/2022 | WO |