This disclosure relates to the technical field of audio processing, and in particular, to an audio processing method, an electronic device, and a computer-readable storage medium.
A lens of an electronic device such as a gimbal camera or a surveillance camera may move under drive of a motor. To prevent noise interference and avoid a lens structure that is too complex, a microphone for acquiring audio is usually not provided on the lens, but on another component that does not rotate with the lens. In this way, when the lens rotates, an angle of view of images captured by the lens correspondingly changes, but a sound source orientation indicated by the audio acquired by the microphone cannot adapt to the change in the angle of view of the captured images, resulting in inconsistent visual and auditory senses of orientation of a captured video.
To resolve the foregoing problem that a sound source orientation indicated by recorded audio does not match images captured by a lens, embodiments of this disclosure provide an audio processing method, an electronic device, and a computer-readable storage medium.
A first aspect of some exemplary embodiments of this disclosure provides an audio processing method, including: obtaining relative attitude information between a lens and a plurality of microphones, where the lens is movable relative to at least one of the plurality of microphones; obtaining original audio signals by the plurality of microphones; determining weight information of the original audio signals based on the relative attitude information; and synthesizing the original audio signals based on the weight information to obtain a target audio signal to be played with images captured by the lens.
A second aspect of some exemplary embodiments of this disclosure provides an audio processing method, including: obtaining original audio signals by a plurality of microphones; synthesizing the original audio signals based on initial weight information of the original audio signals to obtain a target audio signal to be played with images captured by a lens; determining that the lens moves relative to at least one of the plurality of microphones; obtaining relative attitude information between the lens and the plurality of microphones; and adjusting the initial weight information based on the relative attitude information.
In the audio processing method provided in some exemplary embodiments of this disclosure, when the target audio signal is obtained by synthesizing the original audio signals acquired by a plurality of microphones, the weight information corresponding to the original audio signals is determined based on the relative attitude information between the lens and the microphones corresponding to the original audio signals. In this way, even if an angle of view of the images captured by the lens changes relative to the microphones, the target audio signal obtained after the synthesis based on the relative attitude information may still match the images captured by the lens, so as to provide users with consistent visual and auditory senses of orientation.
To describe the technical solutions in some exemplary embodiments of this disclosure, the accompanying drawings required to describe the embodiments will be briefly described below. Apparently, the accompanying drawings described below are only some exemplary embodiments of this disclosure. Those of ordinary skill in the art may further obtain other accompanying drawings based on these accompanying drawings without creative efforts.
The technical solutions in some exemplary embodiments of this disclosure will be described below with reference to the accompanying drawings. Apparently, the described embodiments are merely some rather than all of the embodiments of this disclosure. All other embodiments obtained by those of ordinary skill in the art based on the embodiments of this disclosure without creative efforts should fall within the scope of protection of this disclosure.
An electronic device with a video photographing function may be provided with a lens and a microphone. The lens, also referred to as a camera, may be used to capture images. The microphone may be used to acquire audio. After the captured images and the acquired audio are encapsulated in a specific format, a video (audio/video) may be obtained.
For convenience, the electronic device with the video photographing function is referred to as a photographing device in the exemplary embodiments of this disclosure. A lens of a traditional photographing device is fixed. When a user wants to photograph an object at different positions, the user needs to manually adjust a position of the photographing device such that the lens may aim at the object to be photographed. However, with development of science and technology, there are some new photographing devices whose lenses are no longer fixed, but may autonomously move or rotate as driven by of motors. There are many such photographing devices having movable lenses, such as an unmanned aerial vehicle (UAV) equipped with a gimbal, a gimbal camera, a surveillance camera, a robot, and a panoramic camera.
A gimbal camera may be used as an example for description. A lens of the gimbal camera may move. For example, when an intelligent tracking photographing function is enabled, the lens may lock onto a target and automatically rotate as the target moves. In another example, after a user inputs a rotation command, the lens may rotate as instructed by the rotation command.
To prevent noise interference of a gimbal and avoid an over-complex structure of the lens, a microphone for acquiring audio is usually not provided on the lens, but on another component that does not rotate with the lens, such as a base of the gimbal. In this way, when the lens of the gimbal camera rotates, an angle of view of images captured by the lens correspondingly changes, but a sound source orientation indicated by the audio acquired by the microphone cannot adapt to the change in the angle of view of the captured images, resulting in inconsistent visual and auditory senses of orientation of users for a captured video. This greatly affects user experience and even causes some users to have adverse reactions such as dizziness.
A person determined the sound orientation based on a difference between sound heard by the left and that heard by the right ear. In addition, at least two channels are required for a recorded audio to have a stereo effect. Multi-channel audio may be recorded by using a plurality of microphones. Specifically, during the recording, the plurality of microphones may simultaneously record (acquire) audio. Further, a plurality of pieces of recorded audio may be synthesized to obtain multi-channel audio. The three microphones in
D
L
=w
1L
D
1
+w
2L
D
2
+w
3L
D
3
D
R
=w
1R
D
1
+w
2R
D
2
+w
3R
D
3
where Di represents an original audio signal acquired by an ith microphone (i=1, 2, 3), and wi represents weights corresponding to the ith microphone. It should be noted that each microphone corresponds to two weights: one weight corresponding to the left channel and the other weight corresponding to the right channel. For example, the first microphone corresponds to two weights w1L and w1R, where w1L corresponds to the left channel, and w1R corresponds to the right channel. Correspondingly, the second microphone also corresponds to two weights w2L and w2R, and the third microphone also corresponds to two weights w3L and w3R.
These weights are fixed weights predetermined during previous work. Generally, the weights are determined in the following manner: firstly, designate an orientation as a default lens orientation (hereinafter referred to as the default orientation); next, determine the weights corresponding to audio signals acquired by the microphones based on the default orientation and the layout of the microphones.
For ease of understanding, an example regarding how the weights are determined is provided below with reference to
D
L
=w
1L
D
1
+w
3L
D
3
D
R
=w
2R
D
2
+w
3R
D
3
Since the foregoing weights corresponding to the microphones are determined under the assumption that an orientation of the lens is the default orientation, a sound source orientation indicated by the synthesized audio signals may match an angle of view of captured images only in the case where an actual orientation of the lens is the same as (or close to) the default orientation. In other words, if the actual orientation of the lens is different from the default orientation, the sound source orientation indicated by the synthesized audio signals does not match the angle of view of the captured images.
A specific example of video photographing is provided below. Reference may be made to
The sound source orientation indicated by the recorded audio always matches the default orientation. In practice, a sound source (the user B) is in the orientation of the third microphone relative to the lens, which is exactly the default orientation. Therefore, the sound source orientation indicated by the recorded audio is directly in front of the angle of view. When the user B is photographed, since the actual orientation of the lens is exactly the same as the default orientation, the sound source orientation indicated by the recorded audio matches the images captured by the lens. Specifically, in this example, the images show that the user B in front of the camera is speaking, and the audio heard also indicates that the sound source is in front of the camera. However, when the user C is photographed, since positions of the microphones do not change, and the audio is still synthesized in the same way, the sound source orientation indicated by the recorded audio is still directly in front of the angle of view. However, since the actual orientation of the lens is deviated from the default orientation, the sound source orientation indicated by the recorded audio does not match the images captured by the lens. Specifically, in this example, the images show that the user C directly in front of the camera is listening to the user B on the left side. However, the audio indicates that the sound source is directly in front of the camera, as if it were the user C who is speaking.
To resolve the foregoing problem, some exemplary embodiments of this disclosure provide an audio processing method. The audio processing method may be applied to the foregoing electronic device with the video photographing function, and the electronic device includes a lens and a plurality of microphones, where “the plurality of” can be understood as at least two. The lens of the electronic device is movable relative to at least one of the plurality of microphones. In other words, some of the plurality of microphones are disposed on the lens (and may move with the lens).
S301: Obtain relative attitude information between a lens and a plurality of microphones.
S302: Obtain original audio signals acquired by the plurality of microphones.
S303: Determine weight information corresponding to the original audio signals based on the relative attitude information.
S304: Synthesize the original audio signals based on the weight information to obtain a target audio signal.
In step S304, the synthesized target audio signal is played with images captured by the lens. Specifically, as mentioned above, the target audio signal may be encapsulated with the images captured by the lens in a specific video format to form a video file. When the video file is decapsulated and played, the target audio signal may be played with the images captured by the lens. In other words, the target audio signal may be an audio part of the recorded video, and forms the audio/video with the images captured by the lens.
In the audio processing method provided in some exemplary embodiments of this disclosure, the weight information corresponding to the original audio signal acquired by each microphone is no longer predetermined and fixed. The weight information corresponding to the original audio signals is determined based on the relative attitude information. The relative attitude information herein is the relative attitude information between the lens and the plurality of microphones, and may reflect relative orientation and position relationships between the lens and the microphones. In addition, the relative attitude information may be updated correspondingly after the lens moves relative to the microphones such that the relative attitude information obtained in step S301 may reflect real-time relative attitude between the lens and the microphones.
In some exemplary embodiments, there may be various ways for determining the relative attitude information. In some exemplary embodiments, the relative attitude information may be determined based on an orientation of each microphone and an attitude of the lens. The orientation of the microphone may be an orientation of the microphone relative to the lens. Specifically, the orientation of the microphone may be determined based on a position of the lens and a position of the microphone.
The attitude of the lens may include a position and/or an orientation of the lens. The position of the lens herein may be a position of the lens relative to the body, and the orientation of the lens corresponds to an angle of view of the captured images. In some exemplary embodiments, the lens may be mounted on a body (which may be a body of various devices or platforms) via a gimbal, and the microphones may be fixed on the body. Under control of the gimbal, the lens may move relative to the microphones. In this case, the relative attitude information may be determined based on orientation information of the gimbal. Specifically, the attitude of the lens may be determined based on the orientation information of the gimbal such that the relative attitude information may be determined based on the attitude of the lens and the orientations of the microphones.
The lens may rotate and move under the control of the gimbal. In many scenarios, the lens under the control of the gimbal may rotate. During the rotation, a main change is a change of the orientation of the lens. The position of the lens relative to the body may not change or slightly changes. However, in some scenarios, the lens may also move relative to the body under the control of the gimbal. For example, lenses equipped for some robots may extend, protrude, and slide under the control of the gimbal. During the movement of the lens, the position of the lens relative to the body may change. In other words, a position of the lens relative to the microphones also changes. In this case, the position of the lens may also be determined based on the orientation information of the gimbal.
As described above, to allow a recorded audio to have a stereoscopic effect, the recorded audio needs to have at least two channels of audio signals. In step S304, the synthesized target audio signal may be played on one of at least two channels. The channel corresponding to the target audio signal may be referred to as a target channel.
The channels herein are channels of sound recorded or played at different spatial positions, with corresponding orientations. For example, common dual channels consist of a left channel and a right channel, where “left” and “right” both describe the orientations corresponding to the respective channels. However, the orientations described as “left” and “right” are relative orientations, and actual orientations corresponding to the relative orientations need to be determined based on a reference direction. For example, the reference direction may be a facing direction. When north is faced, the actual orientation corresponding to the relative orientation “left” is west, and the actual orientation corresponding to the relative orientation “right” is east. When east is faced, the actual orientation corresponding to the relative orientation “left” is north, and the actual orientation corresponding to the relative orientation “right” is south.
The orientation of the target channel also has two types: relative orientation and actual orientation. However, considering that the relative orientation is not an absolute orientation, it is inconvenient to directly use the relative orientation in specific implementation. Therefore, an orientation corresponding to the target channel described in this disclosure refers to the actual orientation corresponding to the target channel. The orientation corresponding to the target channel may be determined based on a reference direction, and the reference direction may be the orientation of the lens.
For ease of understanding, reference may be made to
The target audio signal needs to be obtained through the synthesis based on the weight information corresponding to the original audio signals. Weight information of an original audio signal may essentially represent contribution of the original audio signal in the synthesis for obtaining the target audio signal (namely, a proportion of the original audio signal in the synthesis for obtaining the target audio signal). In some exemplary embodiments, contribution (weight information) of an original audio signal in the synthesis for obtaining the target audio signal may be determined based on the relative attitude information between the lens and a microphone corresponding to the original audio signal and the orientation corresponding to the target channel.
That the weight information corresponding to the original audio signal acquired by the microphone is determined based on the relative attitude information of the microphone and the orientation corresponding to the target channel may specifically include: determine deviation information between an orientation of the microphone and the orientation corresponding to the target channel based on the relative attitude information and the orientation corresponding to the target channel, and determine the corresponding weight information based on the deviation information. Reference may be made to
The deviation information may be represented in various forms. In some exemplary embodiments, the deviation information may be an angle between the orientation of the microphone and the orientation corresponding to the target channel (for convenience of reference, such an angle is referred to as a deviation angle hereinafter). Certainly, there are other implementations. For example, levels used to represent deviation degrees may be preset. As shown in
The weight information may be determined based on the deviation information. In some exemplary embodiments, when the deviation information is represented by the foregoing deviation angle, the weight information corresponding to the original audio signal may be determined based on a cosine of the deviation angle.
D
L
=w
1L
D
1
+w
2L
D
2
+w
3L
D
3
where Di represents an original audio signal acquired by an ith microphone (i=1, 2, 3), and wiL represents weight information corresponding to the original audio signal acquired by the ith microphone.
Considering that the orientation of the first microphone relative to the lens is on the right and the target channel is the left channel, if the foregoing deviation angle is used to represent the deviation, the deviation angle corresponding to the first microphone is θ1 in
The deviation angle θ1 of the first microphone in
D
L
=w
2L
D
2
+w
3L
D
3
The following formulas are used to calculate w2L and w3L:
where θ2 represents a deviation angle corresponding to a second microphone, and θ3 represents a deviation angle corresponding to a third microphone.
It may be understood that the cosine of the deviation angle reflects a projection of a unit vector in a same direction as the orientation of the microphone in the orientation corresponding to the target channel. Smaller deviation between the orientation of the microphone and the orientation corresponding to the target channel indicates a larger cosine of the deviation angle corresponding to the microphone, and larger weight information corresponding to the original audio signal acquired by the microphone.
In the foregoing formulas for calculating w2L and w3L, w2L and w3L are normalized. The normalization of the weight information may make the synthesized target audio signal more proper on an amplitude level.
It should be noted that in the audio processing method provided in some exemplary embodiments of this disclosure, when the weight information corresponding to the original audio signals is determined, the weight information corresponding to the original audio signal acquired by each microphone may also be determined. In the foregoing example corresponding to
It is readily understandable that although some exemplary embodiments of this disclosure is described in terms of the target audio signal corresponding to one of the at least two channels, in practical application, an audio signal corresponding to each channel may be obtained through synthesis by using the method provided in this disclosure. In the foregoing example corresponding to
For the foregoing formula, reference may be made to
After the audio signals DL and DR are obtained through synthesis, DL is played on the left channel and DR is played on the right channel. This may produce an auditory sense of orientation that matches the angle of view of the captured images.
In the audio processing method provided in some exemplary embodiments of this disclosure, when the target audio signal is obtained by synthesizing the original audio signals acquired by the plurality of microphones, the weight information corresponding to the original audio signals is determined based on the relative attitude information between the lens and the microphones corresponding to the original audio signals. In this way, even if the angle of view of the images captured by the lens changes relative to the microphones, the target audio signal obtained after the synthesis based on the relative attitude information may still match the images captured by the lens, so as to provide users with consistent visual and auditory senses of orientation.
In the above various implementations, the “orientation of the lens” refers to the actual orientation of the lens. After a series of processing based on the orientation of the lens, the target audio signal that provides a sense of orientation consistent with the captured images may be finally obtained after the synthesis. However, in a special scenario, a user does not want the recorded audio to have a sense of orientation consistent with the captured images, but hopes that a sound source orientation indicated by the recorded audio is a specified orientation.
To facilitate the understanding of the foregoing special scenario, the description may be provided in combination with the foregoing example corresponding to
For this special requirement of the user A, in some exemplary embodiments of this disclosure, the user may set the “orientation of the lens”. In this case, the “orientation of the lens” set by the user is actually a virtual orientation. The virtual orientation and the actual orientation of the lens are independent and unrelated to each other. The virtual orientation set by the user may be used to guide the synthesis for obtaining the target audio signal.
In the example corresponding to
The user may set the “orientation of the lens” such that the synthesized audio may provide a sense of orientation desired by the user, and may better adapt to requirements of different users.
The foregoing describes in detail an audio processing method provided in some exemplary embodiments of this disclosure.
S501: Obtain original audio signals acquired by a plurality of microphones.
S502: Synthesize the original audio signals based on initial weight information corresponding to the original audio signals to obtain a target audio signal.
The target audio signal is played with images captured by a lens.
S503: When the lens moves relative to at least one of the plurality of microphones, obtain relative attitude information between the lens and the plurality of microphones, and adjust the initial weight information based on the relative attitude information.
The lens is mounted on a body via a gimbal, and the microphones are fixed on the body.
The relative attitude information is determined based on orientation information of the gimbal.
In some exemplary embodiments, the relative attitude information is determined based on orientations of the microphones and an attitude of the lens.
In some exemplary embodiments, the attitude of the lens includes an orientation of the lens and/or a position of the lens.
In some exemplary embodiments, the target audio signal is played on a target channel of the at least two channels.
In some exemplary embodiments, that the initial weight information is adjusted based on the relative attitude information includes:
Adjust the initial weight information based on the relative attitude information and an orientation corresponding to the target channel, where the orientation corresponding to the target channel is determined based on an orientation of the lens.
In some exemplary embodiments, that the initial weight information is adjusted based on the relative attitude information and the orientation corresponding to the target channel includes:
Determine deviation information between an orientation of each microphone and the orientation corresponding to the target channel based on the relative attitude information and the orientation corresponding to the target channel, determine new weight information based on the deviation information, and adjust the initial weight information based on the new weight information.
In some exemplary embodiments, the deviation information includes an angle between the orientation of the microphone and the orientation corresponding to the target channel.
In some exemplary embodiments, the new weight information is determined based on a cosine of the angle.
In some exemplary embodiments, if the angle is greater than a preset angle, it is determined that the new weight information corresponding to the original audio signal acquired by the microphone is zero in the synthesis for obtaining the target audio signal.
In some exemplary embodiments, the new weight information is normalized.
In some exemplary embodiments, the orientation of the lens includes a virtual orientation set by a user, and the virtual orientation is independent of an actual orientation of the lens.
In some exemplary embodiments, the at least two channels include a left channel and a right channel.
In the audio processing method provided in some exemplary embodiments of this disclosure, when the target audio signal is obtained by synthesizing the original audio signals acquired by the plurality of microphones, the weight information corresponding to the original audio signals are obtained by adjusting the initial weight information corresponding to the original audio signals based on the relative attitude information, where the relative attitude information may reflect relative orientation and position relationships between the lens and the microphones corresponding to the original audio signals. In this way, even if an angle of view of the images captured by the lens changes relative to the microphones, the target audio signal obtained after the synthesis based on the relative attitude information may still match the images captured by the lens, so as to provide users with consistent visual and auditory senses of orientation.
For the audio processing method in the foregoing various implementations, reference may be made to the corresponding description of the foregoing first audio processing method. Details will not be described herein.
The processor implements the following steps when executing the computer program:
Obtain relative attitude information between the lens and the plurality of microphones.
Obtain original audio signals acquired by the plurality of microphones.
Determine weight information corresponding to the original audio signals based on the relative attitude information.
Synthesize the original audio signals based on the weight information to obtain a target audio signal, where the target audio signal is played with images captured by the lens.
In some exemplary embodiments, the electronic device may further include a gimbal. The lens is mounted on the body through the gimbal, and the microphones are fixed on the body.
The relative attitude information is determined based on orientation information of the gimbal.
In some exemplary embodiments, the relative attitude information is determined based on orientations of the microphones and an attitude of the lens.
In some exemplary embodiments, the attitude of the lens includes an orientation of the lens and/or a position of the lens.
In some exemplary embodiments, the target audio signal is played on a target channel of the at least two channels.
In some exemplary embodiments, when determining the weight information corresponding to the original audio signals based on the relative attitude information, the processor is specifically configured to determine the weight information based on the relative attitude information and an orientation corresponding to the target channel, where the orientation corresponding to the target channel is determined based on an orientation of the lens.
In some exemplary embodiments, when determining the weight information based on the relative attitude information and the orientation corresponding to the target channel, the processor is specifically configured to determine deviation information between an orientation of each microphone and the orientation corresponding to the target channel based on the relative attitude information and the orientation corresponding to the target channel, and determine the weight information based on the deviation information.
In some exemplary embodiments, the deviation information includes an angle between the orientation of the microphone and the orientation corresponding to the target channel.
In some exemplary embodiments, the weight information is determined based on a cosine of the angle.
In some exemplary embodiments, if the angle is greater than a preset angle, it is determined that the weight information corresponding to the original audio signal acquired by the microphone is zero during the synthesis for obtaining the target audio signal.
In some exemplary embodiments, the orientation of the lens includes a virtual orientation set by a user, and the virtual orientation is independent of an actual orientation of the lens.
In some exemplary embodiments, the weight information is normalized.
In some exemplary embodiments, the at least two channels include a left channel and a right channel.
In some exemplary embodiments, the electronic device may further include a plurality of speakers. The speakers have a one-to-one corresponding relationship with the channels.
In some exemplary embodiments, the electronic device may be any one of a UAV, a gimbal camera, a surveillance camera, a panoramic camera, and a robot.
In the electronic device provided in some exemplary embodiments of this disclosure, when the target audio signal is obtained by synthesizing the original audio signals acquired by the plurality of microphones, the weight information corresponding to the original audio signals is determined based on the relative attitude information between the lens and the microphones corresponding to the original audio signals. In this way, even if an angle of view of the images captured by the lens changes relative to the microphones, the target audio signal obtained after the synthesis based on the relative attitude information may still match the images captured by the lens, to provide users with consistent visual and auditory senses of orientation.
For the electronic device in the foregoing various implementations, reference may be made to the corresponding description of the foregoing first audio processing method. Details will not be described herein.
Some exemplary embodiments of this disclosure further provide an electronic device. Still refer to
The processor implements the following steps when executing the computer program:
obtaining original audio signals acquired by the plurality of microphones;
synthesizing the original audio signals based on initial weight information corresponding to the original audio signals to obtain a target audio signal, where the target audio signal is played with images captured by a lens; and
when the lens moves relative to at least one of the plurality of microphones, obtaining relative attitude information between the lens and the plurality of microphones, and adjusting the initial weight information based on the relative attitude information.
In some exemplary embodiments, the electronic device may further include a gimbal. The lens is mounted on the body through the gimbal, and the microphones are fixed on the body.
The relative attitude information is determined based on orientation information of the gimbal.
In some exemplary embodiments, the relative attitude information is determined based on orientations of the microphones and an attitude of the lens.
In some exemplary embodiments, the attitude of the lens includes an orientation of the lens and/or a position of the lens.
In some exemplary embodiments, the target audio signal is played on a target channel of the at least two channels.
In some exemplary embodiments, when adjusting the initial weight information based on the relative attitude information, the processor is specifically configured to adjust the initial weight information based on the relative attitude information and an orientation corresponding to the target channel, where the orientation corresponding to the target channel is determined based on an orientation of the lens.
In some exemplary embodiments, when adjusting the initial weight information based on the relative attitude information and an orientation corresponding to the target channel, the processor is specifically configured to determine deviation information between an orientation of each microphone and the orientation corresponding to the target channel based on the relative attitude information and the orientation corresponding to the target channel, determine new weight information based on the deviation information, and adjust the initial weight information based on the new weight information.
In some exemplary embodiments, the deviation information includes an angle between the orientation of the microphone and the orientation corresponding to the target channel.
In some exemplary embodiments, the new weight information is determined based on a cosine of the angle.
In some exemplary embodiments, if the angle is greater than a preset angle, it is determined that the new weight information corresponding to the original audio signal acquired by the microphone is zero during the synthesis for obtaining the target audio signal.
In some exemplary embodiments, the new weight information is normalized.
In some exemplary embodiments, the orientation of the lens includes a virtual orientation set by a user, and the virtual orientation is independent of an actual orientation of the lens.
In some exemplary embodiments, the at least two channels include a left channel and a right channel.
In some exemplary embodiments, the electronic device may further include a plurality of speakers. The speakers have a one-to-one corresponding relationship with the channels.
In some exemplary embodiments, the electronic device may be any one of a UAV, a gimbal camera, a surveillance camera, a panoramic camera, and a robot.
With the electronic device provided in some exemplary embodiments of this disclosure, when the target audio signal is obtained by synthesizing the original audio signals acquired by the plurality of microphones, the weight information corresponding to the original audio signals are obtained by adjusting the initial weight information corresponding to the original audio signals based on the relative attitude information, where the relative attitude information may reflect relative orientation and position relationships between the lens and the microphones corresponding to the original audio signals. In this way, even if an angle of view of the images captured by the lens changes relative to the microphones, the target audio signal obtained after the synthesis based on the relative attitude information may still match the images captured by the lens, to provide users with consistent visual and auditory senses of orientation.
For the electronic device in the foregoing various implementations, reference may be made to the corresponding description of the foregoing first audio processing method. Details will not be described herein.
Some exemplary embodiments of this disclosure further provide a computer-readable storage medium. The computer-readable storage medium stores a computer program. When the computer program is executed by a processor, the first audio processing method in the foregoing various implementations may be implemented.
Some exemplary embodiments of this disclosure further provide a computer-readable storage medium. The computer-readable storage medium stores a computer program. When the computer program is executed by a processor, the second audio processing method in the foregoing various implementations may be implemented.
Provided that there is no conflict or contradiction in the technical features provided in the foregoing exemplary embodiments, those skilled in the art may combine the technical features based on actual conditions to form various exemplary embodiments. A length of this disclosure is limited, and the various exemplary embodiments are not described, but it may be understood that the various exemplary embodiments also belong to the scope disclosed by the exemplary embodiments of this disclosure.
In addition, some exemplary embodiments of this disclosure may provide a form of a computer program product that is implemented on one or more computer-usable storage media (including, but not limited to, a disk memory, a compact disc read-only memory (CD-ROM), an optical memory, and the like) that include program code. The computer-usable storage media include non-volatile and volatile, and removable and non-removable media, and information storage may be implemented by any method or technology. The information may be computer-readable instructions, data structures, modules of programs, or other data. Examples of storage media of computers include but are not limited to a phase-change memory (PRAM), a static random access memory (SRAM), a dynamic random access memory (DRAM), another type of random access memory (RAM), a read-only memory (ROM), an electrically erasable programmable ROM (EEPROM), a flash memory or another memory technology, a CD-ROM, a digital versatile disc (DVD) or another optical memory, a magnetic tape cassette, magnetic tape and magnetic disk storage or another magnetic storage device, or any other non-transmission media. The storage media may be used to store information that may be accessed by a computing device.
It should be noted that relational terms herein such as first and second are merely used herein to distinguish one entity or operation from another entity or operation without necessarily requiring or implying any actual such relationship or order between the entities or operations. The term “comprise”, “include”, or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article, or device that includes a range of elements includes not only those elements but also other elements that are not explicitly listed or that are inherent to such process, method, article, or device. Without further restrictions, an element defined in the sentence “including a . . . ” do not exclude the existence of other identical elements in a process, method, article, or device including the element.
The method and apparatus provided in some exemplary embodiments of this disclosure have been described in detail above. The principles and implementations of this disclosure are described herein with specific examples. The description of these exemplary embodiments is merely provided to help understand the method and core idea of this disclosure. In addition, a person of ordinary skill in the art can make variations and modifications to this disclosure. Therefore, the contents provided herein shall not be construed as limitations on this disclosure.
This application is a continuation application of PCT application No. PCT/CN2020/092891, filed on May 28, 2020, and the content of which is incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2020/092891 | May 2020 | US |
Child | 17990870 | US |