The present patent application disclosure relates to analyzing music, and more particularly, to analyzing the tempo and beat of music.
Tempo and beat analysis is the basis of rhythm perception and music understanding. Although most humans can easily follow the beat of music by tapping their feet or clapping their hands, detecting a musical beat automatically remains a difficult task.
Various media editing and playback tools utilize automatic beat detection. For example, currently available movie editing tools permit a user to extract important video shots from a movie and to align transitions between these shots with the beat of a piece of music. Various photo viewing and presentation tools allow a user to put together a slideshow of photos set to music. Some of these photo presentation tools can align the transition between photos in the slideshow with the beat of the music. Other music playback media tools provide visualizations on a computer screen while playing back music. Music visualizations can be any sort of visual design such as circles, lines, flames, fountains, smoke, etc., that change in appearance while music is being played back. Transitions in the appearance of a music visualization that are linked to the beat of the music provide a more interesting experience for the user than if such transitions occur randomly.
The burgeoning use of computers to store, access, edit and playback various media through such media tools makes the task of music beat analysis and detection increasingly important. Accurate and efficient beat analysis and detection algorithms are therefore becoming basic components for various media editing and playback tools that perform tasks such as those mentioned above. However, prior methods and systems of beat analysis and detection have several disadvantages. One disadvantage is that most prior beat analysis and detection methods require that assumptions be made about the time signature and hierarchical meter of the music. For example, a typical assumption made in prior methods is that the time signature of the music is 4/4. Another disadvantage with prior methods/systems is that not all of the detected beats in such systems are in sync with the actual beat phase of the music. Often, there are detected beats that are out of sync or locked in a false beat phase. Furthermore, prior methods and systems do not offer a way to rectify the beats that are out of sync with the true beat phase of the music.
Accordingly, a need exists for improved beat analysis and detection that does not require assumptions regarding musical time signature and hierarchical meter, and that overcomes various disadvantages with prior methods such as those mentioned above.
A system and methods analyze music to detect musical beats and to rectify beats that are out of sync with the actual beat phase of the music. The music analysis includes onset detection, tempo/meter estimation, and beat analysis, which includes the rectification of out-of-sync beats.
The same reference numerals are used throughout the drawings to reference like components and features.
Overview
The following discussion is directed to a system that analyzes music to detect the beat of the music. Advantages of this system include an improved approach to beat detection that does not require an assumption of the musical time signature or hierarchical meter. Another advantage is a process for rectifying out-of-sync beats based on tempo consistency across the whole musical excerpt.
Exemplary Environment
The computing environment 100 includes a general-purpose computing system in the form of a computer 102. The components of computer 102 may include, but are not limited to, one or more processors or processing units 104, a system memory 106, and a system bus 108 that couples various system components including the processor 104 to the system memory 106.
The system bus 108 represents one or more of any of several types of bus structures, including a memory bus or memory controller, a peripheral bus, an accelerated graphics port, and a processor or local bus using any of a variety of bus architectures. An example of a system bus 108 would be a Peripheral Component Interconnects (PCI) bus, also known as a Mezzanine bus.
Computer 102 includes a variety of computer-readable media. Such media can be any available media that is accessible by computer 102 and includes both volatile and non-volatile media, removable and non-removable media. The system memory 106 includes computer readable media in the form of volatile memory, such as random access memory (RAM) 110, and/or non-volatile memory, such as read only memory (ROM) 112. A basic input/output system (BIOS) 114, containing the basic routines that help to transfer information between elements within computer 102, such as during start-up, is stored in ROM 112. RAM 110 contains data and/or program modules that are immediately accessible to and/or presently operated on by the processing unit 104.
Computer 102 may also include other removable/non-removable, volatile/non-volatile computer storage media. By way of example,
The disk drives and their associated computer-readable media provide non-volatile storage of computer readable instructions, data structures, program modules, and other data for computer 102. Although the example illustrates a hard disk 116, a removable magnetic disk 120, and a removable optical disk 124, it is to be appreciated that other types of computer readable media which can store data that is accessible by a computer, such as magnetic cassettes or other magnetic storage devices, flash memory cards, CD-ROM, digital versatile disks (DVD) or other optical storage, random access memories (RAM), read only memories (ROM), electrically erasable programmable read-only memory (EEPROM), and the like, can also be utilized to implement the exemplary computing system and environment.
Any number of program modules can be stored on the hard disk 116, magnetic disk 120, optical disk 124, ROM 112, and/or RAM 110, including by way of example, an operating system 126, one or more application programs 128, other program modules 130, and program data 132. Each of such operating system 126, one or more application programs 128, other program modules 130, and program data 132 (or some combination thereof) may include an embodiment of a caching scheme for user network access information.
Computer 102 can include a variety of computer/processor readable media identified as communication media. Communication media embodies computer readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared, and other wireless media. Combinations of any of the above are also included within the scope of computer readable media.
A user can enter commands and information into computer system 102 via input devices such as a keyboard 134 and a pointing device 136 (e.g., a “mouse”). Other input devices 138 (not shown specifically) may include a microphone, joystick, game pad, satellite dish, serial port, scanner, and/or the like. These and other input devices are connected to the processing unit 104 via input/output interfaces 140 that are coupled to the system bus 108, but may be connected by other interface and bus structures, such as a parallel port, game port, or a universal serial bus (USB).
A monitor 142 or other type of display device may also be connected to the system bus 108 via an interface, such as a video adapter 144. In addition to the monitor 142, other output peripheral devices may include components such as speakers (not shown) and a printer 146 which can be connected to computer 102 via the input/output interfaces 140.
Computer 102 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computing device 148. By way of example, the remote computing device 148 can be a personal computer, portable computer, a server, a router, a network computer, a peer device or other common network node, and the like. The remote computing device 148 is illustrated as a portable computer that may include many or all of the elements and features described herein relative to computer system 102.
Logical connections between computer 102 and the remote computer 148 are depicted as a local area network (LAN) 150 and a general wide area network (WAN) 152. Such networking environments are commonplace in offices, enterprise-wide computer networks, intranets, and the Internet. When implemented in a LAN networking environment, the computer 102 is connected to a local network 150 via a network interface or adapter 154. When implemented in a WAN networking environment, the computer 102 includes a modem 156 or other means for establishing communications over the wide network 152. The modem 156, which can be internal or external to computer 102, can be connected to the system bus 108 via the input/output interfaces 140 or other appropriate mechanisms. It is to be appreciated that the illustrated network connections are exemplary and that other means of establishing communication link(s) between the computers 102 and 148 can be employed.
In a networked environment, such as that illustrated with computing environment 100, program modules depicted relative to the computer 102, or portions thereof, may be stored in a remote memory storage device. By way of example, remote application programs 158 reside on a memory device of remote computer 148. For purposes of illustration, application programs and other executable program components, such as the operating system, are illustrated herein as discrete blocks, although it is recognized that such programs and components reside at various times in different storage components of the computer system 102, and are executed by the data processor(s) of the computer.
Exemplary Embodiments
The music analyzer 202, its components, and their respective functions can be briefly described as follows. In general, the music analyzer 202 detects onsets in a music clip using onset detection algorithm 204. An onset is the beginning of a musical sound where the energy usually has a big variance. For example, an onset may be the time when a piano key is pressed down. As discussed below, onsets are usually detected as local peaks from an onset curve. After detecting onsets in a music clip with onset detection algorithm 204, the music analyzer 202 estimates tempo (or meter) using tempo estimation algorithm 206. Tempo is the period of beats, representing basic recurrent rhythmical pattern in the music. Tempo is estimated based on an auto-correlation of the onset curve of the music clip as discussed below. After tempo estimation algorithm 206 estimates the tempo of the music, the beat detection algorithm 208 detects beat sequences based on the onset curve and estimated tempo of the music. After beat sequences are determined, segments containing continuous beat sequences are used to build a phase tree based on which segments share the same beat phase. The rectification algorithm 210 determines which group of segments contains the largest number of beats and assumes those segments to be in sync with the actual beat phase of the music. Segments that are not part of the group of segments which contains the largest number of beats are segments that are assumed to be out-of-sync with the actual beat phase of the music. These out-of-sync segments are then rectified by following the actual beat phase.
Onset detection and tempo estimation will now be discussed in greater detail with primary reference to
After conversion into a uniform format, the data from the music clip is divided into non-overlapping temporal frames, such as 16 microsecond-long frames.
Use of a 16 microsecond frame length is also only an example, and various other non-overlapping frame lengths may also be suitable. The spectrum of each frame is then calculated by FFT (Fast Fourier Transform). Each frame is divided into a number of octave-based sub-bands (Sub-Band 1-Sub-Band N). In this example, each frame is divided into six octave-based sub-bands. The amplitude envelope of each sub-band is then calculated by convolving with a half raise cosine Hanning window. From the amplitude envelope, an onset curve is detected by calculating the variance of the envelope of each sub-band using a Canny operator, that is,
Oi(n)=Ai(n){circle around (×)}C(n) (1)
where Oi(n) is the onset curve in the i-th sub-band, Ai(n) is the amplitude envelope of the i-th sub-band and C(n) is the Canny operator with a Gaussian kernel,
where Lc is the length of Canny operator and the σ is used to control the operator's shape. In a preferred implementation, Lc and σ are set as 12 and 4, respectively. Use of the Canny operator, rather than a one-order difference, has the potential of finding more onsets that have slopes with gradual transitions in the energy envelope. A one-order difference can only catch the abrupt changes in the energy envelope. Use of a half Hanning window and a Canny estimator are both well-known processes to those skilled in the art, and they will therefore not be further described.
An onset curve is a sequence of potential onsets along the time line. The onset curve represents the energy variance at each time slot. Onsets are detected as the local peaks from the onset curve. The onsets, or local peaks, represent the local maximum variance of the energy envelope. From the onsets detected from each sub-band, the lowest and the highest sub-bands contain the most obvious, regular and representative beat patterns. This is reasonable since most beats are indicated by low-frequency and high-frequency instrumentals, especially those using bass drum and snare drum in popular music. Considering this fact, only these two sub-bands (i.e., the lowest sub-band and the highest sub-band) are used for tempo estimation and final beat detection. Thus, in the current example implementation where each frame is divided into six octave-based sub-bands, only the first and sixth sub-bands are used for tempo estimation and final beat detection.
Referring still to
O(n)=ΣOi(n) (3)
where O(n) represents the onset curve of the music.
Auto-correlation is then used to estimate the tempo. Auto-correlation uses memory efficiently and can find subtle meter structure, as demonstrated in the following discussion. Based on all the prominent local peaks of the auto-correlation curve, tempo is estimated as their maximum common divisor, which is also a prominent peak according to equation (4) as follows:
where Pk are the prominent local peaks. In a preferred implementation, the prominent local peaks are detected with a threshold 0.1.
The bar length, or measure, represents a higher structure than beat. A bar, or measure, in music, is one of the small equal parts into which a piece of music is divided. It contains a fixed number of beats. In the present embodiment, the bar length is estimated using certain rules based on the first three maximum peaks of the auto-correlation curve as shown, for example, in
The first rule for estimating the bar length is that if the three peaks of the auto-correlation curve are regularly placed along the period, then the maximum common divisor of the three peaks is used as the estimation of the bar length. Otherwise, the position of the maximum peak along the period is used as the estimation for the bar length. The length is finally normalized to an approximate range, by iterative halving or doubling if the corresponding position also has a local peak in the auto-correlation function.
It should be noted that the bar length detected by this method is prone to be a half or double of the truth value. However, it can still indicate a more subtle structure of the meter. For example, if the bar length is three multiples of the tempo, the meter can be classified into “ternary” meter as shown in
Beat analysis and the rectification of out-of-sync beats will now be discussed in greater detail with primary reference to
As tempo information is obtained, a beat pattern template is established to calculate the confidence that each onset is a beat candidate in the onset sequence (i.e., onset curve). Recall that onsets are detected as the local peaks from the onset curve and they represent the local maximum variance of the energy envelope of the onset curve. The beat template is designed to represent the rhythm pattern of the music.
The beat confidence of each onset is calculated by matching the beat pattern template along the onset sequence, as
where Conf(n) is beat confidence at n-th frame, and PT(k) is the beat pattern template. Thus, for a given onset, if there also appear onsets at estimated positions having regular intervals of tempo, the confidence is high and the onset is more likely to be a beat. Otherwise, the confidence is low and the onset is less likely to be a beat. A potential beat, or beat candidate, is then detected or determined based on confidence level. When the confidence of an onset is above a certain threshold, the onset is detected as a beat candidate. The threshold is adaptively set based on the following:
The beat sequence search process is illustrated in
Based on the above tracking process, the beat search alternates between being in a state of sync and out-of-sync. Thus, final results may contain several independent segments of beats where each segment contains a continuous beat sequence with the interval of the tempo period, but where two contiguous beat segments are not at the interval of multiple tempos. This means that some of segments may be out of sync with the actual beat position, i.e., falsely locked on the wrong beat phase. An example of such a beat search result is demonstrated in
The rectification algorithm 210 determines which segments are synced with the actual beat phase and which segments are out-of-sync with the actual beat phase by first looking for those segments which share the same beat phase. The rectification algorithm 210 assumes that most of the detected beats are correctly phase-locked. Therefore the group of segments having the largest number of beats can be considered to be properly synced with the actual beat phase. Conversely, those segments not falling in with this group, are segments which are considered to be out-of-sync with the actual beat phase.
In order to find the largest sequence of beats from each segment that share the same beat phase (and thereby finding the highest number of detected beats), rectification algorithm 210 builds a phase tree from each segment.
After finding the segment sequence with the largest number of beats, which is assumed to be in sync with the actual beat phase, those segments that are out-of-sync can be easily rectified, just by following the actual beat phases.
As an example,
Exemplary Methods
Example methods for beat analysis and detection in music will now be described with primary reference to the flow diagrams of
A “processor-readable medium,” as used herein, can be any means that can contain, store, communicate, propagate, or transport instructions for use or execution by a processor. A processor-readable medium can be, without limitation, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. More specific examples of a processor-readable medium include, among others, an electrical connection (electronic) having one or more wires, a portable computer diskette (magnetic), a random access memory (RAM) (magnetic), a read-only memory (ROM) (magnetic), an erasable programmable-read-only memory (EPROM or Flash memory), an optical fiber (optical), a rewritable compact disc (CD-RW) (optical), and a portable compact disc read-only memory (CDROM) (optical).
At block 1002 of method 1000, onsets from a music clip are determined. The general process for determining or detecting musical onsets includes various steps. The music clip is first down-sampled to a uniform format such as a 16 kilohertz, 16 bit, mono-channel sample. The music clip is then divided into plurality of frames that are, for example, 16 microseconds in length. The frequency spectrum of each frame is then calculated using FFT (Fast Fourier Transform), and each frame is divided into a number of octave-based frequency sub-bands. In a preferred implementation, frames are divided into 6 octave-based frequency sub-bands. The amplitude envelope of the lowest and the highest sub-bands are calculated by convolving these sub-bands with a half raised, Hanning window. The onset curve is then determined from the amplitude envelope by calculating the variance of the amplitude of the lowest and highest sub-bands. The music onsets can then be determined as the local maximum variances in the amplitude envelope.
At block 1004 of method 1000, the tempo of the music clip is estimated from the onset curve. Estimating the tempo includes summing the onset curves of the lowest and highest sub-bands to first determine the onset curve of the music clip. An auto-correlation curve is then generated from the onset curve of the music clip, and the maximum common divisor of prominent local peaks of the auto-correlation curve is calculated.
At block 1006, the length of a bar (i.e., the length of a measure) of music is estimated. The bar length estimation includes calculating the length as a maximum common divisor of three peaks in the auto-correlation curve if the three peaks are evenly spaced within the tempo of the music clip. However, if the three peaks are not evenly spaced within the tempo of the music clip, the length is selected as the position of the maximum peak within the tempo. The length is finally normalized to an approximate range.
The method 1000 continues with block 1008 of
At block 1010 of method 1000, segments of beat sequence are detected in order to determine parts of the beat sequence that are synced with the actual beat and parts that may not be synced with the actual beat. Locking beat phases includes finding at least 3 continuous beat candidates that have intervals of one or more tempos. The 3 continuous beat candidates are then confirmed as beats.
At block 1012 of method 1000, the segments of beat sequences that are found to be out-of-sync with actual beat phase are rectified. Rectification of out-of-sync segments includes building phase trees from all the beat segments and searching through the phase tree for the largest sequence of segments that share the same beat phase. Then, it is assumed that the segments making up this largest sequence of segments are segments that are synced with the actual beat phase. Conversely, it is assumed that all segments that are not synced segments are out-of-sync segments. The out-of-sync segments are then rectified by following the actual beat phase.
Building the phase tree out of beat segments includes determining if a subsequent segment shares the same beat phase as a current segment. If the subsequent segment shares the same beat phase as the current segment, the subsequent segment is inserted into the phase tree as a child segment of the current segment. This process is repeated until all of the beat segments are processed.
Although the invention has been described in language specific to structural features and/or methodological acts, it is to be understood that the invention defined in the appended claims is not necessarily limited to the specific features or acts described. Rather, the specific features and acts are disclosed as exemplary forms of implementing the claimed invention.
This continuation application claims priority to U.S. patent application Ser. No. 10/811,287 to Lie Lu et al., filed Mar. 25, 2004, entitled, “Beat Analysis of Music Signals.”
Number | Date | Country | |
---|---|---|---|
Parent | 10811287 | Mar 2004 | US |
Child | 11264326 | Nov 2005 | US |