Methods, systems and media for performing visualized quantitative vibrato analysis are provided.
Vibrato is an indispensable technique for obtaining profound music expression in vocal and instrumental music. For example, a violinist can play a musical note with vibrato through to-and-fro motion of his fingers. The conflicting theories about the different types of vibrato frequency variations and their production techniques can confuse music students who are learning how to play vibrato. In addition, the conventional methods of analyzing and teaching vibrato are highly subjective and lack objective standards that can be easily followed by music students.
Accordingly, it is desirable to teach how to produce as vibrato effect using a system that can perform visualized quantitative vibrato analysis.
Methods, systems, and media for performing visualized quantitative vibrato analysis are provided. In some embodiments, methods for analyzing musical vibrato in an audio file are provided, the methods comprising: receiving, using a hardware processor, a target note from a user; receiving, using the hardware processor, a time-domain signal representing a piece of music comprising a plurality of notes, wherein the plurality of notes include the target note and the target note is played with a vibrato effect; converting, using the hardware processor, the time-domain signal to a frequency-domain signal; determining, using the hardware processor, a plurality of changes in frequency and intensity of the vibrato effect over time based on the frequency-domain signal; determining, using the hardware processor, a target frequency corresponding to the target note; and displaying, on a display, data about the changes in frequency and intensity of the vibrato effect over time and data about the target frequency.
In some embodiments, systems for analyzing musical vibrato in an audio file are provided, the systems comprising: at least one hardware processor that is configured to: receive a target note from a user; receive a time-domain signal representing a piece of music comprising a plurality of notes, wherein the plurality of notes include the target note and the target note is played with a vibrato effect; convert the time-domain signal to a frequency-domain signal; determine a plurality of changes in frequency and intensity of the vibrato effect over time based on the frequency-domain signal; and determine a target frequency corresponding to the target note; and at least one display that is configured to: display data about the changes in frequency and intensity of the vibrato effect over time and data about the target frequency.
In some embodiments, computer readable media containing computer-executable instructions that, when executed by a processor, cause the processor to perform a method for analyzing musical vibrato in an audio file are provided, the method comprising: receiving a target note from a user; receiving a time-domain signal representing a piece of music comprising a plurality of notes, wherein the plurality of notes include the target note and the target note is played with a vibrato effect; converting the time-domain signal to a frequency-domain signal; determining a plurality of changes in frequency and intensity of the vibrato effect over time based on the frequency-domain signal; determining a target frequency corresponding to the target note; and displaying data about the changes in frequency and intensity of the vibrato effect over time and data about the target frequency.
Various objects, features, and advantages of the disclosed subject matter can be more fully appreciated with reference to the following detailed description of the disclosed subject matter when considered in connection with the following drawings, in which like reference numerals identify like elements.
In accordance with various embodiments, as described in more detail below, mechanisms for performing visualized quantitative vibrato analysis are provided. These mechanisms can be used to teach and learn techniques for singing a vocal vibrato or playing vibrato with a musical instrument such as a violin, a cello, a guitar, etc.
In some embodiments, after receiving a target note to which a vibrato effect can be applied, an audio file containing a piece of music in which the target note is played with vibrato can be received. For example, an audio file can be imported from a storage device. Alternatively or additionally, a user can play the piece of music by applying vibrato to the target note. An audio file can then be produced by recording the user's real-time performance. Spectrum analysis can be performed on the audio file using any suitable methods. For example, a fast Fourier transform (FFT) can be performed on a time-domain signal representing the piece of music contained in the audio file and a frequency-domain signal can be generated accordingly. Spectrum data about the vibrato applied to the target note can be generated based on the frequency-domain signal. Quantitative analysis can be performed on the spectrum data based on various parameters of the vibrato, such as the frequency variation, the intensity variation, the frequencies with the highest intensities during each cycle of vibrato, the correspondent relationship between the frequency and intensity variations during each cycle of vibrato, etc. Multiple characteristics of the vibrato can be determined based on the spectrum analysis and the quantitative analysis. In some embodiments, the characteristics of the vibrato can be displayed on a display device. The multiple characteristics of the vibrato can be used to evaluate the quality of the performance recorded in the audio file, such as intonation, stability, etc.
In some embodiments, a user can study the preferable way to play a vibrato using the disclosed invention. For example, the user can play a piece of music by applying vibrato effect to a target note. The user's performance can be recorded by a suitable recording device. A sample performance of the vibrato can be generated to illustrate a desirable way to perform the vibrato. The user can playback his performance and the sample performance using a suitable playback device. Spectrum and quantitative analysis can be performed on the user's performance and the sample performance to generate two sets of statistics. The two sets of statistics can be displayed on a screen. The user can study the techniques for playing a vibrato using the two sets of statistics.
Turning to
Computing device 110 can be a general purpose device such as a computer or a special purpose device such as a client, a server, etc. Any of these general or special purpose devices can include any suitable components such as a processor (which can be microprocessor, digital signal processor, a controller, etc.), memory, communication interfaces, display controllers, input devices, etc. For example, computing device 110 can be implemented as a personal computer, a tablet computing device, a personal digital assistant (PDA), a portable email device, a multimedia terminal, a mobile telephone, a gaming device, a set-top box, a television, etc.
As illustrated in
Display device 120 can be provided as a stand-alone device or integrated with other elements of system 100. Display device 120 can be one or more of a monitor, a television, a liquid crystal display (LCD) for a mobile device, or any other suitable equipment for displaying visual images. In some embodiments, display device 120 can be three-dimensional capable. In some embodiments, display device 120 can be a touchscreen.
Recording device 130 can be a digital audio recorder, an analog audio recorder, a personal video recorder (PVR), digital video recorder (DVR), video cassette recorder (VCR), a digital video disk (DVD) recorder, compact disc recorder, a smartphone, or any other suitable recording device or storage device. In some embodiments, recording device 130 can include a storage device for storing or recording content or data recorded or provided by other components of system 100.
Playback device 140 can be as gaming system (e.g., X-BOX, PLAYSTATION, or GAMECUBE) or a portable electronic device, such as a portable DVD player, a portable gaming device, a cellular telephone, a personal digital assistant (PDA), a music player (e.g., a MP3 player), or any other suitable fixed or portable device.
Input device 150 can be a computer keyboard, a mouse, a keypad, a cursor-controller, a remote control, or any other suitable input device as would be used by a designer of input systems or process control systems. Alternatively, input device 150 can be a finger-sensitive or stylus-sensitive touch screen input of display device 120. Input device can also be a microphone or other voice recognition device which can receive acoustic input from a user.
Turning to
Process 200 can begin by receiving a target note from a user at 202. For example, as shown in
Referring back to
In another example, a user can click panel 508 of user interface 500 to record an audio file containing a piece of music in which the target note is played with vibrato. More particularly, for example, the user can play a piece of music by applying vibrato to the target note. Recording device 130 can record the user's performance in real time and produce an audio file accordingly. In some embodiments, the audio file can be stored in storage module 118 or any suitable storage device.
Referring back to
At 208, computing device 110 can perform quantitative analysis on the spectrum data and determine multiple characteristics of the vibrato applied to the target note. For example, analysis module 114 can analyze the spectrum data within the duration of the vibrato. It can then calculate a target pitch ratio which can represent the ratio of a time period during which the fundamental frequency matches the target frequency to the duration of the vibrato. Analysis module 114 can also calculate an above target pitch ratio which can represent the ratio of a time period during which the fundamental frequency is greater than the target frequency to the duration of the vibrato. Similarly, analysis module 114 can calculate a below target pitch ratio which can represent the ratio of a time period during which the fundamental frequency is less than the target frequency to the duration of the vibrato. Additionally or alternatively, analysis module 114 can calculate a vibrato ratio which can represent the ratio of the above target pitch ratio to the below target pitch ratio.
As another example, analysis module 114 can determine the correspondent relationship between the frequency and intensity variations. More particularly, for example, analysis module 114 can determine the positions of the intensity peaks in a cycle of vibrato. Analysis module 114 can then determine the correspondent relationship between the intensity peaks and the fundamental frequencies. For example, analysis module 114 can determine that the intensity peaks are aligned with the fundamental frequency peaks, the fundamental frequency troughs, the target pitch, etc. In some embodiments, a close and consistent alignment between the intensity peaks and the target pitch can indicate the intonation of the vibrato recorded in the audio file. For example, analysis module 114 can determine that the intensity peaks are substantially and consistently aligned with the target pitch in a cycle of vibrato. Analysis module 114 can then determine that the vibrato recorded in the audio file was played with good intonation. As another example, analysis module 114 can determine that the intensity peaks are substantially and consistently aligned with the fundamental frequency peaks or troughs in a cycle of vibrato. Analysis module 114 can then determine that the vibrato recorded in the audio file was played with good intonation. Alternatively, analysis module 114 can determine that some of the intensity peaks in a cycle of vibrato are substantially aligned with the fundamental frequency peaks while the others are substantially aligned with the fundamental frequency troughs. Analysis module 114 can then determine that the vibrato recorded in the audio file was not played with good intonation.
As yet another example, analysis module 114 can determine the type of the vibrato recorded in the audio file. As illustrated in
In some embodiments, analysis module 114 can determine the type of the vibrato recorded in the audio file based on an initial motion after the fundamental frequency reaches the target pitch for the first time and the correspondent relationship between the intensity peaks and the fundamental frequencies in a vibrato cycle. For example, analysis module 114 can determine that the vibrato recorded in the audio file was played by starting at the target pitch and producing an oscillation using an initial motion after the target pitch was reached for the first time. In a more particular example, analysis module 114 can determine that the vibrato was produced by a backward motion (e.g., moving one's fingers toward the violin scroll) or a forward motion (e.g., moving one's fingers toward the violin bridge) after the target pitch was reached for the first time. In another more particular example, analysis module 114 can determine that the vibrato recorded in the audio file was produced by a backward motion after the target pitch was reached for the first time and the intensity peaks were substantially aligned with the fundamental frequency peaks. Analysis module 114 can then determine that the vibrato recorded in the audio file is a TB vibrato. In another more particular example, analysis module 114 can determine that the vibrato recorded in the audio file was produced by a forward motion after the target pitch was reached for the first time and the intensity peaks were substantially aligned with the fundamental frequency troughs. Analysis module 114 can then determine that the vibrato recorded in the audio file is a TA vibrato. In another more particular example, analysis module 114 can determine that the vibrato recorded in the audio file was produced by a forward motion after the target pitch was reached for the first time and the intensity peaks are aligned with the target pitch. Analysis module 114 can then determine that the vibrato recorded in the audio file is a TBA vibrato.
Alternatively or additionally, analysis module 114 can determine the type of vibrato recorded in the audio file based on the vibrato ratio. For example, analysis module 114 can determine that the vibrato recorded in the audio file is a TB vibrato if the vibrato ratio is less than or equal to 0.5. As another example, analysis module 114 can determine that the vibrato recorded in the audio file is a TA vibrato lithe vibrato ratio is greater than or equal to 2. Alternatively, analysis module 114 can determine that the vibrato recorded in the audio file is a TBA/TAB vibrato if the vibrato ratio falls within the interval (0.5, 2).
At 210, display device 120 can display to the user the spectrum data and the characteristics of the vibrato applied to the target note. For example, display device 120 can display the spectrum data and the characteristics using a suitable user interface. More particularly, for example, as shown in
It should be understood that some of the above steps of the flow diagram of
Turning to
As illustrated, process 300 can begin by receiving a time-domain signal representing the piece of music contained in the audio file by computing device 110 at 302. For example, processing module 112 can load an audio file in any suitable format and read a time-domain signal which can represent a sound waveform recorded in the audio file.
Next, at 304, computing device 110 can convert the time-domain signal into a frequency-domain signal. For example, processing module 112 can perform a fast Fourier transform (FFT) on the time-domain signal. More particular, for example, processing module 112 can compute the frequency spectrum corresponding to a time series segment of the time-domain signal using the Hann window function. Processing module 112 can identify the frequency with the highest intensity near a note's standard frequency as the instantaneous frequency of the time series segment. Processing, module 117 can then applying FFT analysis to another segment of the time-domain signal. For example, processing module 112 can analyze a second time-series segment having the same length as that of the first time-series segment while shifting the width of the second time-series segment with a small percentage along the time series (e.g., 5%). Processing module 112 can perform the spectrum analysis described above iteratively and compute the spectrum and the instantaneous frequencies for the entire time-domain signal.
At 306, computing device 110 can collect data about variations of the fundamental frequency and the intensity of the fundamental frequency over time. For example, graphic module 116 can generate a two-dimensional spectrum graphic representing the changes in the fundamental frequency and the intensity of the fundamental frequency over time. More particular, for example, graphic module 116 can generate graphic 700 as illustrated in
Alternatively or additionally, graphic module 116 can generate a three-dimensional spectrum graphic (e.g., a waterfall graphic) which can represent the changes in the fundamental frequency and the intensity of the fundamental frequency over time. More particular, for example, graphic module 116 can generate a three-dimensional graphic 800 as illustrated in
Referring back to
Referring back to
It should be understood that some of the above steps of the flow diagram of
Turning to
As illustrated, music sheet 400 includes notes 420, 410, and 430 which can be played sequentially. In some embodiments, note 410 can be played with vibrato. As described above, a user can identify note 410 as a target note. The user can also identify notes 420 and 430 as a preceding note and a subsequent note, respectively.
In some embodiments, a user can play music sheet 400 and apply vibrato to note 410 with a musical instrument. As descried above, recording device 130 can record the user's performance in real time and produce an audio file accordingly. In some embodiments, a user can import an audio file on which music sheet 400 played by a musician has been recorded.
Turning to
As described above in connection with
As described above in connection with
Alternatively or additionally, a user can click perform button 508 of user interface 500 to record an audio file containing a piece of music in which the target note is played with vibrato. For example, the user can play a piece of music by applying vibrato to the target note with a musical instrument. Recording device 130 can record the user's performance and produce an audio file accordingly. The user can click play button 516 of user interface 500 to instruct playback device 140 to play the recorded audio file. Alternatively or additionally, the user can click perform button 508 to record a video file. The user can then click play button 516 to display the video in video display area 518.
As illustrated in
Turning to
In some embodiments, a user can identify as target note by entering the name of the vibrato note in area 902. For example, the user can enter Bb4 as a target note in area 902. Alternatively or additionally, the user can identify a target note by selecting a portion of staff 904 corresponding to the vibrato note. For example, a user can click the portion of staff 904 corresponding to B4 and the flat button “b” to identify Bb4 as a target note.
The user can also select a vibrato type that will be applied to the target note in area 906. As described above in connection with
The user can then click play sample button 908 of user interface 900 to instruct playback device 140 to play a sample audio file in which the target note is played with the type of vibrato selected by the user. For example, computing device 110 can impart the type of vibrato effect selected by the user to the target tone and generate an audio file accordingly. As another example, computing device 110 can retrieve an audio file including the target note played with vibrato by a musician. In some embodiments, the user can use the sample audio file to study the preferable way to play the type of vibrato selected by the user.
The user can then click vibrato exercises panel 910 to practice playing vibrato with a musical instrument. In some embodiments, the user can play the target note without vibrato. Computing device 110 can then determine the target pitch corresponding to the target note played by the user. In some embodiments, the user can play the target note with vibrato. Computing device 110 can analyze the user's performance. For example, computing device 110 can perform spectrum and quantitative analysis on the user's performance based on the methods described in connection with
Computing device 110 can also generate multiple statistics about the sample audio file and the user's performance. Display device 120 can then display the statistics in display are 912 of user interface 900. Computing device 110 can also generate a graphic illustrating the statistics and the difference between the sample performance and the user's performance. Display device 120 can then display the graphic in display area 914 of user interface 900. For example, as described above in connection with
In some implementations, any suitable computer readable media can be used for storing instructions for performing the processes described herein. For example, in some implementations, computer readable media can be transitory or non-transitory. For example, non-transitory computer readable media can include media such as magnetic media (such as hard disks, floppy disks, etc.), optical media (such as compact discs, digital video discs, Blu-ray discs, etc.), semiconductor media (such as flash memory, electrically programmable read only memory (EPROM), electrically erasable programmable read only memory (EEPROM), etc.), any suitable media that is not fleeting or devoid of any semblance of permanence during transmission, and/or any suitable tangible media. As another example, transitory computer readable media can include signals on networks, in wires, conductors, optical fibers, circuits, any suitable media that is fleeting and devoid of any semblance of permanence during transmission, and/or any suitable intangible media.
Although the application is described herein as being implemented on a user computer and/or server, this is only illustrative. The application may be implemented on any suitable platform (e.g., a personal computer (“PC”), a data display, a portable computer, a palmtop computer, a handheld PC, a laptop computer, a cellular phone, a personal digital assistant (“PDA”), a combined cellular phone and PDA, etc.) to provide such features.
Accordingly, methods, systems, and media for performing visualized quantitative vibrato analysis are provided.
Although the disclosed subject matter has been described and illustrated in the foregoing illustrative implementations, it is understood that the present disclosure has been made only by way of example, and that numerous changes in the details of implementation of the disclosed subject matter can be made without departing from the spirit and scope of the disclosed subject matter. For example, although the present invention has been described as performing visualized quantitative vibrato analysis in connection with the playing of a violin, visualized quantitative vibrato analysis can be performed in connection with vocal music or any suitable musical instruments, such as a cello, a viola, a guitar, any other musical instrument in which vibrato can be effected, etc. Features of the disclosed implementations can be combined and rearranged in various ways.
Number | Date | Country | Kind |
---|---|---|---|
101122407 | Jun 2012 | TW | national |
This application claims the benefit of U.S. Provisional Patent Application No. 61/758,583, filed Jan. 30, 2013, and Taiwan Patent Application No. 101122407, filed Jun. 22, 2012, each of which is hereby incorporated by reference herein in its entirety.
Number | Date | Country | |
---|---|---|---|
61758583 | Jan 2013 | US |