Panning operations and surround sound decoding operations are mathematically distinct functions that affect the distribution of sound across a speaker system. Panning is the spread of a sound signal into a new multi-channel sound field. Panning is a common function in multi-channel audio systems. Panning functions distribute sound across multi-channel sound systems. In effect, panning “moves” the sound to a different speaker. If the audio is panned to the right, then the right speaker gets most of the audio stream and the left speaker output is reduced.
Surround sound decoding is the mathematical or matrix computations necessary to transform two-channel audio into the necessary multi-channel audio stream to support a surround sound system. Surround sound decoding is the process of transforming two-channel audio input into multi-channel audio output. Audio that is recorded in 5.1 is often encoded in a two-channel format to be broadcast in environments that only support the two-channel format, like broadcast television. Encoding can be of a mathematical form or a matrix form. Mathematical forms require a series of mathematical steps and algorithms to decode. DTS and Dolby Digital perform mathematical encoding. Matrix encoding relies on matrix transforms to encode 5.1 channel audio into a two-channel stream. Audio in matrix encoding can be played either encoded or decoded and be sound acceptable to the end user.
Some embodiments provide a panner that incorporates a surround sound decoder. The panner takes as input the desired panning effect that a user requests, separates sounds using surround sound decoding, and places the separated sounds in the desired places in an output sound field. Use of surround sound decoding by the panner provides several advantages for placing the sound in the field over the panners that do not use decoding.
Panners use collapsing and/or attenuating techniques to create a desired panning effect. Collapsing relocates the sound to a different location in the sound space. Attenuating increases the strength of one or more sounds and decreases the strength of one or more other sounds in order to create the panning effect. However, collapsing sounds folds down all input signal sources into a conglomerate of sounds and sends them to where the panning is directed to. As a result unwanted sounds that were not intended to be played at certain speakers cannot be separated from the desired sounds and are sent in the panning direction. Also, attenuating sounds without separating them often creates unwanted silence.
A collapsing panner that incorporates surround sound decoding increases the separation between the source signals prior to collapsing them and thereby provides the advantage that all signals are not folded into the same speaker. Another advantage of separating the sounds prior to collapsing them is preventing the same sound to be sent to multiple unwanted speakers thereby maintaining the uniqueness of the sounds at desired speakers. A panner that incorporates surround sound decoding also provides an enabling technology for attenuating panners in many situations where attenuating the sounds prior to separation creates silence.
The preceding Summary is intended to serve as a brief introduction to some embodiments of the invention. It is not meant to be an introduction or overview of all inventive subject matter disclosed in this document. The Detailed Description that follows and the Drawings that are referred to in the Detailed Description will further describe the embodiments described in the Summary as well as other embodiments. Accordingly, to understand all the embodiments described by this document, a full review of the Summary, Detailed Description and the Drawings is needed. Moreover, the claimed subject matters are not to be limited by the illustrative details in the Summary, Detailed Description and the Drawings, but rather are to be defined by the appended claims, because the claimed subject matters can be embodied in other specific forms without departing from the spirit of the subject matters.
The novel features of the invention are set forth in the appended claims. However, for purpose of explanation, several embodiments of the invention are set forth in the following figures.
In the following detailed description of the invention, numerous details, examples, and embodiments of the invention are set forth and described. However, it will be clear and apparent to one skilled in the art that the invention is not limited to the embodiments set forth and that the invention may be practiced without some of the specific details and examples discussed.
Some embodiments provide a panner that incorporates a surround sound decoder. The panner takes as input the desired panning effect that a user requests, separates sounds using surround sound decoding, and places the separated sounds in the desired places in an output sound field. Use of surround sound decoding by the panner provides several advantages for placing the sound in the field over the panners that do not use decoding.
Panners use collapsing and/or attenuating techniques to create a desired panning effect. Collapsing relocates the sound to a different location in the sound space. Attenuating increases the strength of one or more sounds and decreases the strength of one or more other sounds in order to create the panning effect. However, collapsing sounds folds down all input signal sources into a conglomerate of sounds and sends them to where the panning is directed to. As a result unwanted sounds that were not intended to be played at certain speakers cannot be separated from the desired sounds and are sent in the panning direction. Also, attenuating sounds without separating them often creates unwanted silence.
A collapsing panner that incorporates surround sound decoding increases the separation between the source signals prior to collapsing them and thereby provides the advantage that all signals are not folded into the same speaker. Another advantage of separating the sounds prior to collapsing them is preventing the same sound to be sent to multiple unwanted speakers thereby maintaining the uniqueness of the sounds at desired speakers. A panner that incorporates surround sound decoding also provides an enabling technology for attenuating panners in many situations where attenuating the sounds prior to separation creates silence.
Several more detailed embodiments of the invention are described in sections below. Section I provides an overview of panning and decoding operations. Next, Section II describes a panner that uses surround sound decoding in some embodiments. Section III describes rigging of master controls to subordinate controls in some embodiments. Section IV describes the graphical user interface of a media-editing application in some embodiments. Finally, a description of an electronic system with which some embodiments of the invention are implemented is provided in Section V.
A. Definitions
1. Audio Panning
Audio panning is the spreading of audio signal in a sound space. Panning can be done by moving a sound signal to certain audio speakers. Panning can also be done by changing the width, attenuating, and/or collapsing the audio signal. The width of an audio signal refers the width over which sound appears to originate to a listener at a reference point in the sound space (e.g., a width of 0.0 corresponds to a point source). Attenuation means that the strength of one or more sounds is increased and the strength of one or more other sounds is decreased. Collapsing means that sound is relocated (not re-proportioned) to a different location in the sound space.
Audio panners allow an operator to create an output signal from a source audio signal such that characteristics such as apparent origination and apparent amplitude of the sound are controlled. Some audio panners have a graphical user interface that depicts a sound space having a representation of one or more sound devices, such as audio speakers. As an example, the sound space may have five speakers placed in a configuration to represent a 5.1 surround sound environment. Typically, the sound space for 5.1 surround sound has three speakers to the front of the listener (front left (L) and front right (R), and center (C)), two surround speakers at the rear (left surround (Ls) and right surround (Rs)), and one channel for low frequency effects (LFE). A source signal for 5.1 surround sound has five audio channels and one LFE channel, such that each source channel is mapped to one audio speaker.
2. Surround Sound Decoding
Surround sound decoding is an audio technology where a finite number of discrete audio channels (e.g., two) are decoded into a larger number of channels on play back (e.g., five or seven). The channels may or may not be encoded before transmission or recording by an encoder. The terms “surround sound decoding” and “decoding” are used interchangeably throughout this specification.
As an example, a simple surround sound decoder uses the following formula to derive the surround sound signal from the encoded signals.
L=Lt
R=Rt
C=0.7*(Lt+Rt)
Ls=Rs=0.5*(Lt−Rt)
where L, R, C, Ls, Rs, Lt, and Rt are left, right, center, left surround, right surround, left total, and right total signals respectively.
B. Graphical User Interface
As shown in
The process finally sends (at 340) the decoded sound to the speakers. The process then ends. In some embodiments, after the panning input is used by the decoder to decode the signal, an actual panning is also performed (i.e., the sounds is physically moved towards the panned direction) when the output signal is sent to the speakers.
One of ordinary skill in the art will recognize that process 300 is a conceptual representation of the operations used to perform decoding by using panning inputs and to perform panning operations. The specific operations of process 300 may not be performed in the exact order shown and described. The specific operations may not be performed in one continuous series of operations, and different specific operations may be performed in different embodiments. Furthermore, the process could be implemented using several sub-processes, or as part of a larger macro process.
A. Examples of Panning Using Surround Sound Decoding
When only attenuating panning (and not decoding) is done (as shown by arrow 840), all speakers 810-830 are silent. Panning by attenuating does not relocate sound channels. Since the sound (as shown in
When only collapsing panning (and not decoding) is done (as shown by arrow 845), all speakers except the center speaker 820 are silent. Panning by collapsing relocates all sound channels to where the puck 835 is directed. As a result, the center speaker plays sounds from all channels including the judge, left player, right player, and crowd. Since the center speaker 820 is usually used for the sounds at the center of the stage (in this case the voice of the judge), having all sounds including the crowd and the left and right players to come out of the center speaker is not desirable.
In contrast, when decoding is used (as shown by arrow 850), some embodiments utilize the panning input (which is the movement of the puck 835 to the front center) to decode the channels in a way that the judge's sound is heard on the center speaker while all other speakers 810-815 and 825-830 are silent. Specifically, the voice of the judge is separated from the sounds of the players and the crowd by doing the surround sound decoding. The resulting sounds are then panned to the front center speaker. As a result, the judge's sound is heard on the center speaker and other speakers are left silent.
When only attenuating panning (and not decoding) is done (as shown by arrow 940), all speakers except the front left speaker 910 are silent. Since the sound (as shown in
When only collapsing panning (and not decoding) is done (as shown by arrow 945), the left front 910 and left surround 925 speakers receive sounds from all channels and other speakers 915-920 and 930 are silent. Therefore, panning using collapsing in this case has the undesired effect of playing the judge 625 and the left and right players (610 and 615, respectively) on the left surround speaker 925 and playing the judge 625, right player 615, and crowd 630 on the left front speaker 910.
In contrast, when decoding is used (as shown by arrow 950), some embodiments utilize the panning input (which is the movement of the puck 935 to the left most position) to decode the channels in a way that the left player's sound is played on the left front, the crowd is heard on the left surround speaker while all other speakers 915-920 and 930 are silent. Specifically, the voice of the left player and the crowd noise are separated from the other sounds by doing the surround sound decoding. The resulting sounds are then panned to the left. As a result, left player sound is sent to the left speaker 910, the crowd noise is sent to the left surround speaker 925, and other speakers are left silent.
When only attenuating panning (and not decoding) is done (as shown by arrow 1040), all speakers 1010-1030 are silent. Since the sound (as shown in
When only collapsing panning (and not decoding) is done (as shown by arrow 1045), the left surround 1025 and the right surround 1030 speakers receive sounds from all channels including the judge, left player, right player, and crowd which has the undesirable effect of hearing the sounds of the judge and left and right players on the surround speaker.
In contrast, when decoding is used (as shown by arrow 1050), some embodiments utilize the panning input (which is the movement of the puck 1035 to the center back position) to decode the channels in a way that left surround 1025 and the right surround 1030 speakers receive the crowd sound and all other speakers are silent. Specifically, the sounds are separated by doing surround sound decoding. The result is then panned to the center back which results in the crowd noise to be heard on the surround speakers 1025-1030.
As shown in the examples of
Specifically,
The right most column shows the position of the five speakers according to position of speakers in any of
For position A in the first row, the center speaker is shown to receive both Lt and Rt signals while all other four speakers are silent (as shown by number 0 above the lines that indicate the speaker positions). Similarly, other locations B-J in the input sound space are reproduced by proper settings of several panning and decoding inputs. Signals received at some speakers are weaker than the others. For instance, for position B, the Cl and Cr signals to surround speakers are weaker than Cl and Cr signals to the left and right front speakers due to the position of the puck between the center and front of the sound space.
As shown in
The values for the decode balance, Ls/Rs width, and F/R bias parameters shown in
FR Bias=−6y
LsRs Width=(x+1)2+2
Decoder Balance=(1−x2)−100(y4)−6x
where x and y are the x and y coordinates of the panner within the unit circle. The panner then does a mixture of collapsing and attenuating in equations.
B. Different Decoding Techniques Used
In some embodiments, the surround sound decoder takes the panning parameters and uses them to adjust the formulas that are used to do the surround sound decoding. Some formula coefficients also change in time both independent from the panning inputs as well as in response to changing of panning parameters. For instance, some decoders specify the center signal as follows:
C=0.7(G*Lt+(1−G)*Rt)
G=√{square root over (Σn=x−30x(Ltn2−Rtn2)*λn)}
where the Σ operator sums the difference between the squares of Lt and Rt signals over a certain number of previous samples (in this example over 30 previous samples), x identifies the current sample, n is the index identifying each sample, and λn denotes how fast the output signal level (i.e., the center signal, C) follows the changes of the input signals levels, i.e., Lt and Rt signals.
Using the above formulas allows compensating for the time varying signals. For instance, if overtime the left signal is louder, the above formula for C and G compensate for that. In some embodiments, the matrix formulas are dependent on the values of one or more of the panning and decoding parameters as well as the time. In these embodiments, changing the panning and/or decoding inputs adjusts the matrix and the quickness of the response to changes in the Lt and Rt signals.
Other embodiments use other formulas for surround sound decoding. For instance, the following program code is used in an embodiment that brings a louder channel down to the quieter channel. Specifically, the root-mean-square (RMS) of the right and left channels are compared and the channels are scaled based on the comparison. The output signals are then calculated using the scaled values.
Some embodiments perform additional enhancements during surround sound decoding. For instance, some embodiments delay the two surround outputs (e.g., the surround output would be ˜10 milliseconds after the left, center, and right outputs). Some embodiments apply lowpass or bandpass filters to the scaled input signals or the center and surround outputs. Furthermore, some embodiments additionally keep a running RMS of the center and surround signals to be used to drive attenuators on the output channels.
Furthermore, the decoding algorithm of different embodiments run any number of other decoding algorithms, including but not limited to Dolby Surround Dolby Pro Logic, DTS Neural Surround™ UpMix, DTS Neo:6, TC Electronic|Unwrap HD, SRS Circle Surround II, and Lexicon LOGIC 7™ Overview.
Also, some embodiments utilize different ways of generating surround sound in addition (or instead) of a typical decoding. For instance, some embodiments generate surround content with a surround reverb. Other embodiments perform some other techniques for source reconstruction. In all these embodiments, the decoding is used in conjunction with panning to achieve more convincing and realistic placement of sound in a virtual surround field.
The user interface module 1305 receives panning parameters 1325 and decoding parameters 1330 (e.g., through the GUI 200). The user interface module passes the panning parameters 1325 and decoding parameters 1330 to the decoding module 1320 and panning module 1335. The panning module 1335 and the decoding module 1320 use one or more of the techniques described in this specification to generate the output audio signal from the received input audio signal 1355. The “send output signal “module” sends the output audio signal to a set of speakers 1350 (five are shown).
The present application describes a graphical user interface that provides users with numerous ways to perform different sets of operations and functionalities. In some embodiments, these operations and functionalities are performed based on different commands that are received from users through different input devices (e.g., keyboard, trackpad, touchpad, mouse, etc.). For example, in some embodiments, the present application uses a cursor in the graphical user interface to control (e.g., select, move) objects in the graphical user interface. However, in some embodiments, objects in the graphical user interface can also be controlled or manipulated through other controls, such as touch control. In some embodiments, touch control is implemented through an input device that can detect the presence and location of touch on a display of the input device. An example of a device with such a functionality is a touch screen device (e.g., as incorporated into a smart phone, a tablet computer, etc.). In some embodiments with touch control, a user directly manipulates objects by interacting with the graphical user interface that is displayed on the display of the touch screen device. For instance, a user can select a particular object in the graphical user interface by simply touching that particular object on the display of the touch screen device. As such, when touch control is utilized, a cursor may not even be provided for enabling selection of an object of a graphical user interface in some embodiments. However, when a cursor is provided in a graphical user interface, touch control can be used to control the cursor in some embodiments.
In some embodiments, one or more parameters are used to control a larger set of decode and/or panning parameters.
The master control of some embodiments is a slider control. In stage 1 (1405) the master control 1400 has been set to a minimum value (at the far left of the slider) by the user selection input 1435. Stage 1 (1405) illustrates the values of the decode and pan controls when the master control 1400 is set to a minimum value. In the illustrated embodiment, the minimum value for the master control 1400 corresponds to a minimum value of the pan parameter. This minimum value of the pan parameter is shown by indicator 1430 with knob 1432, which is at the far left end of the indicator.
In this figure, the minimum value of the master control 1400 corresponds to the minimum possible value of the panning parameter. However, some embodiments provide master controls 1400 whose minimum values do not necessarily correspond to the minimum possible values of the subordinate parameters.
Stage 2 (1410) shows the values of the decode and pan parameters at an intermediate value of the master control 1400. Stage 2 (1410) demonstrates that some embodiments adjust different parameters by disproportionate amounts when the setting of the master control 1400 increases by a particular amount. The master control 1400 is set at an intermediate value (at about a third of the length of the master control slider). The decode parameter (as shown by knob 1427 of decode parameter indicator 1425) has increased considerably in response to the relatively small change in the master control's 1400 setting. However, the pan parameter (as shown by knob 1432 of decode parameter indicator 1430) has increased only slightly in response to that change in the master control's 1400 setting. That is, the small increase in the setting of the master control 1400 results in a large increase in one subordinate parameter and a small increase in another subordinate parameter.
Stage 3 (1415) shows the values of the decode and pan parameters at a large value of the master control. Stage 3 (1415) demonstrates that the master control can set the subordinate parameters in a non-linear manner. In this stage, the decode parameter has increased only slightly compared to its value in stage 2 (1410) even though the setting of the master control has gone up considerably. This contrasts with the large increase of the decode parameter from stage 1 (1405) to stage 2 (1410) when the master control setting went up only slightly. In stage 3 (1415) the pan parameter has increased proportional to the change in the master control's setting. Demonstrating that in some embodiments one parameter (here, the panning parameter) can have a linear relationship to the master control over part of the range of the master control even while another parameter (here the decode parameter) is non-linear over that range.
Stage 4 (1420) shows the values of the decode parameter and panning parameter when he master control's setting is at maximum. The master control's setting has gone up slightly compared to the setting in stage 3 (1415). The decode parameter has gone up very slightly, while the pan parameter has gone up significantly. The large increase in the panning parameter demonstrates that a parameter can have a linear relationship to the master control's setting for part of the range of the master control, but the same parameter can have a non-linear relationship to the master control's setting for another part of the range of the master control.
Although
The process defines (at 1530) GUI controls for the master and subordinate parameters. In some embodiments, defining GUI controls for the master includes assigning the master parameter to an existing control (e.g., an existing slider) in a particular display area of the GUI. The GUI controls for the subordinate parameters of some embodiments are designed to be indicators of the values of the subordinate parameters as set by the GUI control for the master parameter. As mentioned in the preceding paragraph, process 1500 of some embodiments defines a value for each of one or more subordinate parameters for a subset of the possible values of the master parameter. In some embodiments, when a program (not shown) implements the GUI controls for such embodiments, the program determines the values of the subordinate parameters based on the defined values. When the GUI control for the master parameter is set between two parameters for which subordinate parameter values are defined, some such programs determine the subordinate parameter values by interpolating the set values. Once the process 1500 defines (at 1530) the GUI controls, the process ends.
Although process 1500 utilizes a master control and a set of subordinate controls, some embodiments do not require a master control to control the set of subordinate parameters. In these embodiments, a set of parameters are rigged together and changing any of these parameters changes the other parameters. Similarly, all discussions for
One of ordinary skill in the art will recognize that process 1500 is a conceptual representation of the operations used to setting relationships between master parameters and subordinate parameters. The specific operations of process 1500 may not be performed in the exact order shown and described. The specific operations may not be performed in one continuous series of operations, and different specific operations may be performed in different embodiments. Furthermore, the process could be implemented using several sub-processes, or as part of a larger macro process.
Referring back to
Process 1600 then receives (at 1605) an interpolation function for interpolating values of parameters associated to each of a set of subordinate controls and a master control that are going to be rigged. In some embodiments, the GUI designer selects each control individually. For instance, a GUI designer selects the control for rotation parameter 1740. The designer then selects an interpolation function from a list of interpolation functions (e.g. a sine function 1720) by using control 1730. Process 1600 receives the interpolation function when the designer selects the associate button 1745 to associate the selected function to the selected control. The function is then used to determine values of each parameter based on the position of the associated controls as described below. The function is also used to interpolate values of parameters as described below. In some embodiments, process 1600 receives the interpolation function when the user enters a mathematical formula for the interpolation function through the text field 1720 and selects the associate button 1745.
Next, process 1600 receives (at 1610) positional settings for a set of subordinate controls that control a set of corresponding subordinate parameters. Referring to
Process 1600 then determines (at 1615) a value for each subordinate parameter based on the positional setting of the corresponding control. For instance, each value of the parameter “Balance” in display area 205 of
Next, process 1600 receives (at 1620) a positional setting for a control that controls the master parameter. For instance, the process receives a value after a user selection input positions master control 220 in
The process then receives (at 1630) a command to associate (or rig) the setting of the master control to the values of the set of subordinate parameters. For instance, in some embodiments when the save button 1705 is selected through a user selection input, process 1600 receives a command to associate (rig) the setting of the master control to the values of the selected subordinate parameters. The process stores (at 1635) the values of the master and subordinate parameters and the positional settings of their associated controls as one snapshot of the desired effect.
The process then determines (at 1640) whether another snapshot of the values is required. If so, the process proceeds to 1610 to receive another set of values for the master and subordinate parameters. Otherwise, the process optionally interpolates or extrapolates (at 1645) values of each parameter received to calculate intermediate values for the parameters. The process uses the interpolation function that is associated with each control. For instance when a master control parameter setting of 0 is associated with a subordinate control parameter setting of 6 and a master control parameter setting of 10 is associated with a subordinate control parameter setting of 12, then process 1600 (when a linear interpolation function is associated to the subordinate control) automatically associates a master control parameter setting of 5 (i.e., halfway between the received master control parameter settings) with a subordinate control parameter of 9 (i.e. halfway between the received subordinate control settings). Similarly, when interpolation function is non-linear (e.g., a sine function, a Bezier curve, etc.) the non-linear function is used to calculate the interpolated and extrapolated values. The process stores these values along with the received values of snapshots to create the desired effect.
The process then receives a name for the effect and associates (at 1650) the effect and the snapshot values to the master control. Referring to
One of ordinary skill in the art will recognize that process 1600 is a conceptual representation of the operations used for rigging a set of subordinate parameters to a master control to create a desired effect. The specific operations of process 1600 may not be performed in the exact order shown and described. The specific operations may not be performed in one continuous series of operations, and different specific operations may be performed in different embodiments. Furthermore, the process could be implemented using several sub-processes, or as part of a larger macro process.
The effect creation module 1820 receives inputs from user interface module and communicates with interpolation function determination module 1825, snapshot creation module 1830, range selection module 1835, and rigging module 1840. Interpolation function determination module 1825 receives the interpolation function associated with each control when the interpolation function is selected (either by entering a formula through the text field 1720 or by selection of an existing function through control 1720) and associate button 1745 is selected through a user selection input. Interpolation function determination module saves the interpolation function associated with each control into storage 1850. In some embodiments, a default linear interpolation function is assigned by the interpolation function determination module 1825 to each control prior to receiving an interpolation function for the control.
Snapshot creation module 1830 receives and saves values of the master and subordinate parameters for each snapshot. Range selection module 1835 receives the minimum and maximum range values for each control. Rigging module 1840 rigs the values of master and subordination controls. In some embodiments, rigging module 1840 communicates with rigging interpolation module 1845 to calculate additional snapshots by interpolating values of snapshots generated by snapshot creation module 1830. Storage 1850 is used to store and retrieve values of different ranges and parameters.
Next, process 1900 determines (at 1915) whether the new position of the master control was saved in a snapshot of the rigged values. As was described by reference to
When the new position of the master control is not saved in a snapshot, the process interpolates (or extrapolates) (at 1930) the values of the panning parameters rigged to the master control based on the new position of the master control, at least two saved adjacent positions of the master control, and the values of the rigged parameters corresponding to the saved adjacent master control positions. The process also changes the position of the associated controls for the rigged subordinate parameters.
Next, the process interpolates (or extrapolates) (at 1935) the values of the decoding parameters rigged to the master control based on the new position of the master control, at least two saved adjacent positions of the master control, and the values of the rigged parameters corresponding to the saved adjacent master control positions. The process also changes the position of the associated controls for the rigged subordinate parameters. The process then ends.
In some embodiments, in addition to receiving adjustments to master control (as shown in operation 1910), process 1900 receives adjustments to one or more rigged panning and/or decoding parameters. In some of these embodiments, such an adjustment takes the adjusted parameter out of the rig. In other embodiments, such an adjustment stops rigging all other parameters as well. Yet in other embodiments, such an adjustment does not take the adjusted parameter out of the rig but offsets the value of the adjusted parameter. These embodiments allow a user to modify the rig by offsetting the values of the rigged parameters.
Values of parameters shown on vertical lines 2025 are the saved snapshot values. The “in between” values 2010 are interpolated using a linear interpolation function that interpolates the values 2005 saved in snapshots. Similarly, the values 2015 of another rigged parameter are interpolated to derive interpolated values 2020.
Referring back to
One of ordinary skill in the art will recognize that process 1900 is a conceptual representation of the operations used to setting relationships between master parameters and subordinate parameters. The specific operations of process 1900 may not be performed in the exact order shown and described. The specific operations may not be performed in one continuous series of operations, and different specific operations may be performed in different embodiments. Furthermore, the process could be implemented using several sub-processes, or as part of a larger macro process.
Accordingly, in order to create the fly left surround to right front effect, the master control 220 is rigged to the puck 245 (which controls panning x and y values), panning collapse parameter (which shows how much sound is relocated to a different location in the sound space), and the decoding balance parameter (which indicates how much of the original sound versus the decoded sound is sent to speakers 235). In this example, other panning and decoding parameters are not rigged to the master control in order to create the fly from left surround to right front effect. Also as shown in
The user interface module 2405 receives (e.g., through the GUI 200) the position of a master control (e.g., position of slider control 220) that controls the value of a master parameter that is rigged to a set of subordinate parameters. The user interface module passes the master parameter value 2430 to set rigged parameters values module 2460. The set rigged parameters values module 2460 uses the master parameter value 2430 to determine the values of the rigged parameters. When the value of the master parameter corresponding to the received master control position is stored in a snapshot, snapshot retrieval module 2465 retrieves the values of the rigged parameters from the storage 2475 and sends them to set rigged parameters values module 2460. When the value of the master parameter is not stored in a snapshot, the rigging interpolation module 2470 calculates the values of the rigged parameters by interpolating or extrapolating the values of the parameters rigged to the master parameter based on the received value of the master parameter, at least two saved adjacent positions of the master parameter, and the values of the rigged parameters corresponding to the saved adjacent master parameters. Set rigged parameters values module 2460 sends the values of the rigged parameters to decoding module 2420 and panning module 2435. The panning module 2435 and the decoding module 2420 use one or more of the techniques described in this specification to generate the output audio signal from a received input audio signal 2455. The “send output signal “module” sends the output audio signal to a set of speakers 2450 (five are shown).
The clip library 2505 includes a set of folders through which a user accesses media clips (i.e. video clips, audio clips, etc.) that have been imported into the media-editing application. Some embodiments organize the media clips according to the device (e.g., physical storage device such as an internal or external hard drive, virtual storage device such as a hard drive partition, etc.) on which the media represented by the clips are stored. Some embodiments also enable the user to organize the media clips based on the date the media represented by the clips was created (e.g., recorded by a camera).
Within a storage device and/or date, users may group the media clips into “events”, or organized folders of media clips. For instance, a user might give the events descriptive names that indicate what media is stored in the event (e.g., the “New Event 2-8-09” event shown in clip library 2505 might be renamed “European Vacation” as a descriptor of the content). In some embodiments, the media files corresponding to these clips are stored in a file storage structure that mirrors the folders shown in the clip library.
Within the clip library, some embodiments enable a user to perform various clip management actions. These clip management actions may include moving clips between events, creating new events, merging two events together, duplicating events (which, in some embodiments, creates a duplicate copy of the media to which the clips in the event correspond), deleting events, etc. In addition, some embodiments allow a user to create sub-folders of an event. These sub-folders may include media clips filtered based on tags (e.g., keyword tags). For instance, in the “New Event 2-8-09” event, all media clips showing children might be tagged by the user with a “kids” keyword, and then these particular media clips could be displayed in a sub-folder of the event that filters clips in this event to only display media clips tagged with the “kids” keyword.
The clip browser 2510 allows the user to view clips from a selected folder (e.g., an event, a sub-folder, etc.) of the clip library 2505. As shown in this example, the folder “New Event 2-8-09” is selected in the clip library 2505, and the clips belonging to that folder are displayed in the clip browser 2510. Some embodiments display the clips as thumbnail filmstrips, as shown in this example. By moving a cursor (or a finger on a touchscreen) over one of the thumbnails (e.g., with a mouse, a touchpad, a touchscreen, etc.), the user can skim through the clip. That is, when the user places the cursor at a particular horizontal location within the thumbnail filmstrip, the media-editing application associates that horizontal location with a time in the associated media file, and displays the image from the media file for that time. In addition, the user can command the application to play back the media file in the thumbnail filmstrip.
In addition, the thumbnails for the clips in the browser display an audio waveform underneath the clip that represents the audio of the media file. In some embodiments, as a user skims through or plays back the thumbnail filmstrip, the audio plays as well.
Many of the features of the clip browser are user-modifiable. For instance, in some embodiments, the user can modify one or more of the thumbnail size, the percentage of the thumbnail occupied by the audio waveform, whether audio plays back when the user skims through the media files, etc. In addition, some embodiments enable the user to view the clips in the clip browser in a list view. In this view, the clips are presented as a list (e.g., with clip name, duration, etc.). Some embodiments also display a selected clip from the list in a filmstrip view at the top of the browser so that the user can skim through or playback the selected clip.
The timeline 2515 provides a visual representation of a composite presentation (or project) being created by the user of the media-editing application. Specifically, it displays one or more geometric shapes that represent one or more media clips that are part of the composite presentation. The timeline 2515 of some embodiments includes a primary lane (also called a “spine”, “primary compositing lane”, or “central compositing lane”) as well as one or more secondary lanes (also called “anchor lanes”). The spine represents a primary sequence of media which, in some embodiments, does not have any gaps. The clips in the anchor lanes are anchored to a particular position along the spine (or along a different anchor lane). Anchor lanes may be used for compositing (e.g., removing portions of one video and showing a different video in those portions), B-roll cuts (i.e., cutting away from the primary video to a different video whose clip is in the anchor lane), audio clips, or other composite presentation techniques.
The user can add media clips from the clip browser 2510 into the timeline 2515 in order to add the clip to a presentation represented in the timeline. Within the timeline, the user can perform further edits to the media clips (e.g., move the clips around, split the clips, trim the clips, apply effects to the clips, etc.). The length (i.e., horizontal expanse) of a clip in the timeline is a function of the length of media represented by the clip. As the timeline is broken into increments of time, a media clip occupies a particular length of time in the timeline. As shown, in some embodiments the clips within the timeline are shown as a series of images. The number of images displayed for a clip varies depending on the length of the clip in the timeline, as well as the size of the clips (as the aspect ratio of each image will stay constant).
As with the clips in the clip browser, the user can skim through the timeline or play back the timeline (either a portion of the timeline or the entire timeline). In some embodiments, the playback (or skimming) is not shown in the timeline clips, but rather in the preview display area 2520.
In some embodiments, the preview display area 2520 (also referred to as a “viewer”) displays images from video clips that the user is skimming through, playing back, or editing. These images may be from a composite presentation in the timeline 2515 or from a media clip in the clip browser 2510. In this example, the user has been skimming through the beginning of video clip 2540, and therefore an image from the start of this media file is displayed in the preview display area 2520. As shown, some embodiments will display the images as large as possible within the display area while maintaining the aspect ratio of the image.
The inspector display area 2525 displays detailed properties about a selected item and allows a user to modify some or all of these properties. In some embodiments, the inspector displays one of the GUIs shown in
The additional media display area 2530 displays various types of additional media, such as video effects, transitions, still images, titles, audio effects, standard audio clips, etc. In some embodiments, the set of effects is represented by a set of selectable UI items, each selectable UI item representing a particular effect. In some embodiments, each selectable UI item also includes a thumbnail image with the particular effect applied. The display area 2530 is currently displaying a set of effects for the user to apply to a clip. In this example, several video effects are shown in the display area 2530.
The toolbar 2535 includes various selectable items for editing, modifying what is displayed in one or more display areas, etc. The right side of the toolbar includes various selectable items for modifying what type of media is displayed in the additional media display area 2530. The illustrated toolbar 2535 includes items for video effects, visual transitions between media clips, photos, titles, generators and backgrounds, etc. In addition, the toolbar 2535 includes an inspector selectable item that causes the display of the inspector display area 2525 as well as the display of items for applying a retiming operation to a portion of the timeline, adjusting color, and other functions.
The left side of the toolbar 2535 includes selectable items for media management and editing. Selectable items are provided for adding clips from the clip browser 2510 to the timeline 2515. In some embodiments, different selectable items may be used to add a clip to the end of the spine, add a clip at a selected point in the spine (e.g., at the location of a playhead), add an anchored clip at the selected point, perform various trim operations on the media clips in the timeline, etc. The media management tools of some embodiments allow a user to mark selected clips as favorites, among other options.
One or ordinary skill will also recognize that the set of display areas shown in the GUI 2500 is one of many possible configurations for the GUI of some embodiments. For instance, in some embodiments, the presence or absence of many of the display areas can be toggled through the GUI (e.g., the inspector display area 2525, additional media display area 2530, and clip library 2505). In addition, some embodiments allow the user to modify the size of the various display areas within the UI. For instance, when the display area 2530 is removed, the timeline 2515 can increase in size to include that area. Similarly, the preview display area 2520 increases in size when the inspector display area 2525 is removed.
Many of the above-described features and applications are implemented as software processes that are specified as a set of instructions recorded on a computer readable storage medium (also referred to as computer readable medium). When these instructions are executed by one or more computational or processing unit(s) (e.g., one or more processors, cores of processors, or other processing units), they cause the processing unit(s) to perform the actions indicated in the instructions. Examples of computer readable media include, but are not limited to, CD-ROMs, flash drives, random access memory (RAM) chips, hard drives, erasable programmable read only memories (EPROMs), electrically erasable programmable read-only memories (EEPROMs), etc. The computer readable media does not include carrier waves and electronic signals passing wirelessly or over wired connections.
In this specification, the term “software” is meant to include firmware residing in read-only memory or applications stored in magnetic storage which can be read into memory for processing by a processor. Also, in some embodiments, multiple software inventions can be implemented as sub-parts of a larger program while remaining distinct software inventions. In some embodiments, multiple software inventions can also be implemented as separate programs. Finally, any combination of separate programs that together implement a software invention described here is within the scope of the invention. In some embodiments, the software programs, when installed to operate on one or more electronic systems, define one or more specific machine implementations that execute and perform the operations of the software programs.
The bus 2605 collectively represents all system, peripheral, and chipset buses that communicatively connect the numerous internal devices of the electronic system 2600. For instance, the bus 2605 communicatively connects the processing unit(s) 2610 with the read-only memory 2630, the GPU 2615, the system memory 2620, and the permanent storage device 2635.
From these various memory units, the processing unit(s) 2610 retrieves instructions to execute and data to process in order to execute the processes of the invention. The processing unit(s) may be a single processor or a multi-core processor in different embodiments. Some instructions are passed to and executed by the GPU 2615. The GPU 2615 can offload various computations or complement the image processing provided by the processing unit(s) 2610. In some embodiments, such functionality can be provided using CoreImage's kernel shading language.
The read-only-memory (ROM) 2630 stores static data and instructions that are needed by the processing unit(s) 2610 and other modules of the electronic system. The permanent storage device 2635, on the other hand, is a read-and-write memory device. This device is a non-volatile memory unit that stores instructions and data even when the electronic system 2600 is off. Some embodiments of the invention use a mass-storage device (such as a magnetic or optical disk and its corresponding disk drive) as the permanent storage device 2635.
Other embodiments use a removable storage device (such as a floppy disk, flash memory device, etc., and its corresponding disk drive) as the permanent storage device. Like the permanent storage device 2635, the system memory 2620 is a read-and-write memory device. However, unlike storage device 2635, the system memory 2620 is a volatile read-and-write memory, such a random access memory. The system memory 2620 stores some of the instructions and data that the processor needs at runtime. In some embodiments, the invention's processes are stored in the system memory 2620, the permanent storage device 2635, and/or the read-only memory 2630. For example, the various memory units include instructions for processing multimedia clips in accordance with some embodiments. From these various memory units, the processing unit(s) 2610 retrieves instructions to execute and data to process in order to execute the processes of some embodiments.
The bus 2605 also connects to the input and output devices 2640 and 2645. The input devices 2640 enable the user to communicate information and select commands to the electronic system. The input devices 2640 include alphanumeric keyboards and pointing devices (also called “cursor control devices”), cameras (e.g., webcams), microphones or similar devices for receiving voice commands, etc. The output devices 2645 display images generated by the electronic system or otherwise output data. The output devices 2645 include printers and display devices, such as cathode ray tubes (CRT) or liquid crystal displays (LCD), as well as speakers or similar audio output devices. Some embodiments include devices such as a touchscreen that function as both input and output devices.
Finally, as shown in
Some embodiments include electronic components, such as microprocessors, storage and memory that store computer program instructions in a machine-readable or computer-readable medium (alternatively referred to as computer-readable storage media, machine-readable media, or machine-readable storage media). Some examples of such computer-readable media include RAM, ROM, read-only compact discs (CD-ROM), recordable compact discs (CD-R), rewritable compact discs (CD-RW), read-only digital versatile discs (e.g., DVD-ROM, dual-layer DVD-ROM), a variety of recordable/rewritable DVDs (e.g., DVD-RAM, DVD-RW, DVD+RW, etc.), flash memory (e.g., SD cards, mini-SD cards, micro-SD cards, etc.), magnetic and/or solid state hard drives, read-only and recordable Blu-Ray® discs, ultra density optical discs, any other optical or magnetic media, and floppy disks. The computer-readable media may store a computer program that is executable by at least one processing unit and includes sets of instructions for performing various operations. Examples of computer programs or computer code include machine code, such as is produced by a compiler, and files including higher-level code that are executed by a computer, an electronic component, or a microprocessor using an interpreter.
While the above discussion primarily refers to microprocessor or multi-core processors that execute software, some embodiments are performed by one or more integrated circuits, such as application specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs). In some embodiments, such integrated circuits execute instructions that are stored on the circuit itself In addition, some embodiments execute software stored in programmable logic devices (PLDs), ROM, or RAM devices.
As used in this specification and any claims of this application, the terms “computer”, “server”, “processor”, and “memory” all refer to electronic or other technological devices. These terms exclude people or groups of people. For the purposes of the specification, the terms display or displaying means displaying on an electronic device. As used in this specification and any claims of this application, the terms “computer readable medium,” “computer readable media,” and “machine readable medium” are entirely restricted to tangible, physical objects that store information in a form that is readable by a computer. These terms exclude any wireless signals, wired download signals, and any other ephemeral signals.
While the invention has been described with reference to numerous specific details, one of ordinary skill in the art will recognize that the invention can be embodied in other specific forms without departing from the spirit of the invention. In addition, a number of the figures (including
While the invention has been described with reference to numerous specific details, one of ordinary skill in the art will recognize that the invention can be embodied in other specific forms without departing from the spirit of the invention. Thus, one of ordinary skill in the art would understand that the invention is not to be limited by the foregoing illustrative details, but rather is to be defined by the appended claims.
The present Application claims the benefit of U.S. Provisional Patent Application 61/443,670, entitled, “Audio Panning with Multi-Channel Surround Sound Decoding,” filed Feb. 16, 2011 and U.S. Provisional Patent Application 61/443,711, entitled, “Panning Presets,” filed Feb. 16, 2011. The contents of U.S. Provisional Patent Application 61/443,670 and U.S. Provisional Patent Application 61/443,711 are hereby incorporated by reference.
Number | Date | Country | |
---|---|---|---|
61443670 | Feb 2011 | US | |
61443711 | Feb 2011 | US |