This application is a continuation-in-part of, and claims the benefit of, U.S. patent application Ser. No. 16,448,130, filed Jun. 21, 2019, titled “SYSTEMS, DEVICES, AND METHODS FOR DIGITAL REPRESENTATIONS OF MUSIC”, the content of which is incorporated herein in its entirety by reference.
The present systems, devices, and methods generally relate to computer-readable representations of music, and particularly relate to hierarchical digital file formats that include harmonic structure to enable improved computer-based algorithmic music composition.
A musical composition may be characterized by sequences of sequential, simultaneous, and/or overlapping notes that are partitioned into one or more tracks. Starting with an original musical composition, a new musical composition or “variation” can be composed by manipulating the “elements” (e.g., notes, bars, tracks, arrangement, etc.) of the original composition. As examples, different notes may be played at the original times, the original notes may be played at different times, and/or different notes may be played at different times. Further refinements can be made based on many other factors, such as changes in musical key and scale, different choices of chords, different choices of instruments, different orchestration, changes in tempo, the imposition of various audio effects, changes to the sound levels in the mix, and so on.
In order to compose a new musical composition (or variation) based on an original or previous musical composition, it is typically helpful to have a clear characterization of the elements of the original musical composition. In addition to notes, bars, tracks, and arrangements, “segments” are also important elements of a musical composition. In this context, the term “segment” (or “musical segment”) is used to refer to a particular sequence of bars (i.e., a subset of serially-adjacent bars) that represents or corresponds to a particular section or portion of a musical composition. A musical segment may include, for example, an intro, a verse, a pre-chorus, a chorus, a bridge, a middle8, a solo, or an outro. The section or portion of a musical composition that corresponds to a “segment” may be defined, for example, by strict rules of musical theory and/or based on the sound or theme of the musical composition.
Musical notation broadly refers to any application of inscribed symbols to visually represent the composition of a piece of music. The symbols provide a way of “writing down” a song so that, for example, it can be expressed and stored by a composer and later read and performed by a musician. While many different systems of musical notation have been developed throughout history, the most common form used today is sheet music.
Sheet music employs a particular set of symbols to represent a musical composition in terms of the concepts of modern musical theory. Concepts like: pitch, rhythm, tempo, chord, key, dynamics, meter, articulation, ornamentation, and many more, are all expressible in sheet music. Such concepts are so widely used in the art today that sheet music has become an almost universal language in which musicians communicate.
While it is common for human musicians to communicate musical compositions in the form of sheet music, it is notably uncommon for computers to do so. Computers typically store and communicate music in well-established digital audio file formats, such as .mid, .wav, or .mp3 (just to name a few), that are designed to facilitate communication between electronic instruments and other devices by allowing for the efficient movement of musical waveforms over computer networks. In a digital audio file format, audio data is typically encoded in one of various audio coding formats (which may be compressed or uncompressed) and either provided as a raw bitstream or, more commonly, embedded in a container or wrapper format. When the audio data corresponds to a musical composition, the audio data usually corresponds to a particular instance (e.g., a particular performance or recording) of the musical composition with all of the nuance and expression specific to that particular instance. The audio data in well-established audio file formats typically does not capture most (or any) of the higher-level musical characteristics of the musical composition that would be represented in sheet music. On the other hand, sheet music typically does not capture the nuance or expression that can characterize a particular instance of a musical composition and make it stand out among other instances of the same composition, such as small imperfections in timing or rhythm.
Sheet music and well-established audio file formats are different from one another because they are designed to accomplish different things. Sheet music visually characterizes the composition of a piece of music using the high-level concepts of musical theory, while audio file formats encode digital data corresponding to a particular instance of the piece. As a result of these differences, sheet music and well-established audio file formats each have their respective advantages and disadvantages. Sheet music is particularly well-suited for use in the composition of music and poorly suited for use in the communication of audio data between computers, while audio file formats are well-suited for use in the communication of audio data between computers but poorly suited for use in the composition of music. Sheet music itself represents a definitive and unnaturally perfect instance of a piece of music while audio file formats may capture nuance and expression not present in the corresponding sheet music. There is a need in the art for a new way of representing music that combines the best features of these two forms, namely, a computer-readable digital audio file format that is well-suited for use in the communication of audio data between computers and that also encodes both the higher-level musical concepts and the nuance and expression that characterize a musical composition. Such a digital audio file format would have broad applicability in, among other things, computer-based composition of music.
A computer-implemented method of encoding a musical composition in a digital audio format may be summarized as including: encoding respective components of the musical composition in respective data objects; encoding the data objects in a data file; and storing the data file in a non-transitory processor-readable storage medium. Encoding respective components of the musical composition in respective data objects may include encoding respective components of the musical composition in a set of hierarchically-dependent data objects. Encoding respective components of the musical composition in a set of hierarchically-dependent data objects may include: encoding a first component of the musical composition in a first data object; encoding a second component of the musical composition in a second data object, the second data object including the first data object; and encoding a third component of the musical composition in a third data object, the third data object including the second data object.
Encoding respective components of the musical composition in a set of hierarchically-dependent data objects may include: encoding parameters of each note of the musical composition in a respective note data object; and encoding parameters of each bar of the musical composition in a respective bar data object. Encoding parameters of each note of the musical composition in a respective note data object may include, for each note of the musical composition, encoding in a respective note data object at least two note parameters selected from a group consisting of: pitch, start time, duration or length, stop time, instrument, sustain, attack, reverb, and volume. Encoding parameters of each bar of the musical composition in a respective bar data object may include, for each bar of the musical composition, encoding in a respective bar data object at least two bar parameters selected from a group consisting of: bar index, time signature, beats per minute, duration or length, start time, stop time, beat timing, key, scale, chords, tracks, sequence of notes, and sequence of percussion events. Encoding respective components of the musical composition in a set of hierarchically-dependent data objects may further include encoding at least one sequence of bar data objects in an arrangement data object. Encoding respective components of the musical composition in a set of hierarchically-dependent data objects may further include encoding each segment of the musical composition in a respective segment data object, wherein: each segment data object includes a respective sequence of bar data objects; encoding at least one sequence of bar data objects in an arrangement data object includes encoding a sequence of segment data objects in the arrangement data object; and encoding each segment of the musical composition in a respective segment data object includes encoding, in at least one segment data object, at least one segment of the musical composition selected from a group consisting of: an intro of the musical composition, a verse of the musical composition, a pre-chorus of the musical composition, a chorus of the musical composition, a bridge of the musical composition, a middle8 of the musical composition, a solo of the musical composition, and an outro of the musical composition.
Each bar in at least a first set of bars of the musical composition may include a respective sequence of notes, in which case encoding parameters of each bar of the musical composition in a respective bar data object may include, for each bar in the first set of bars of the musical composition, encoding a respective sequence of note data objects in a respective bar data object. Encoding parameters of each bar of the musical composition in a respective bar data object may include encoding at least two track data objects in at least one bar data object, each track data object including a respective sequence of note data objects.
The musical composition may include a first bar. The first bar may include a first note and the first bar may consist of (e.g., include no more or less than) a first set of beats. In this case, encoding parameters of each note of the musical composition in a respective note data object may include: encoding a pitch of the first note in a first note data object; and encoding a start time of the first note in the first note data object. Furthermore, encoding parameters of each bar of the musical composition in a respective bar data object may include: encoding the first note data object in a first bar data object; and encoding, in the first bar data object, a respective start time of each beat in the first set of beats. The start time of the first note may not be synchronized with (i.e., may be “unsynchronized with”) the respective start time of any (or all) of the beats in the first set of beats.
The first bar of the musical composition may include a first sequence of notes. The first sequence of notes may include the first note, in which case encoding parameters of each note of the musical composition in a respective note data object may include encoding, in a respective note data object, a respective pitch of each note in the first sequence of notes and a respective start time of each note in the first sequence of notes.
The method may further include: generating a variation of the first bar, the variation of the first bar comprising: the first set of beats from the first bar and a second sequence of notes that is different from the first sequence of notes; and encoding the variation of the first bar in the data file. Encoding the variation of the first bar in the data file may include: encoding, in the data file, the respective start times of the beats in the first set of beats; and encoding, in the data file, a respective pitch of each note in the second sequence of notes and a respective start time of each note in the second sequence of notes. The respective start time of each note in the second sequence of notes may count from the respective start time of a respective note in the first sequence of notes and the respective start time of at least one note in the second sequence of notes may include the start time of the first note that is not synchronized with (i.e., “unsynchronized with”) the respective start time of any (or all) of the beats in the first set of beats.
The combination of: i) encoding, in the first bar data object, a respective start time of each beat in the first set of beats; and ii) encoding, in a respective note data object, a respective pitch of each note in the first sequence of notes and a respective start time of each note in the first sequence of notes, may together provide a timing relationship between the notes the beats in the first bar. In this case, encoding the variation of the first bar in the data file may include applying the timing relationship between the notes and the beats of the first bar to the notes and the beats of the variation.
Encoding the variation of the first bar data object in the data file may include replacing, in the data file, the first bar data object with the variation of the first bar data object. Encoding, in the data file, the respective start times of the beats in the first set of beats may include encoding, in the first bar data object, the respective start times of the beats in the first set of beats. Encoding, in the data file, a respective pitch of each note in the second sequence of notes and a respective start time of each note in the second sequence of notes may include encoding, in the first bar data object, the respective pitch of each note in the second sequence of notes and the respective start time of each note in the second sequence of notes.
Encoding respective components of the musical composition in respective data objects may include, for at least a first note of the musical composition, encoding a set of first note parameters in a first note data object. The set of first note parameters may include a pitch of the first note and a start time of the first note, and the method may further include encoding, in the data file, a respective start time of each beat in the musical composition, wherein the start time of the first note is not synchronized with (i.e., is “unsynchronized with”) the respective start time of any (or all) of the beats in the musical composition.
The musical composition may include notes and beats, and encoding respective components of the musical composition in respective data objects may include encoding respective start times of the notes and respective start times of the beats independently of one another in the data file.
Encoding the data objects in a data file may include encoding the data objects in a data file having a .hum file extension.
A system for encoding a musical composition in a digital audio format may be summarized as including: at least one processor and at least one non-transitory processor-readable storage medium communicatively coupled to the at least one processor, the at least one non-transitory processor-readable storage medium storing processor-executable instructions and/or data that, when executed by the at least one processor, cause the at least one processor to: encode respective components of a first instance of the musical composition in respective data objects; encode the data objects in a data file; and store the data file in at least one non-transitory processor-readable storage medium of the at least one non-transitory processor-readable storage medium. The first instance of the musical composition may include a first sequence of notes and a first set of bars, in which case the data file stored in the non-transitory processor-readable storage medium may encode: the data objects in a hierarchically-dependent structure; each note of the first instance of the musical composition in a respective note data object; and each bar of the first instance of the musical composition in a respective bar data object. The first instance of the musical composition may include a first set of beats, in which case the data file stored in the non-transitory processor-readable storage medium may encode: a respective beat start time of each beat; a respective note start time of each note; and the beat start times and the note start times independently of one another to provide a timing relationship between the beats and the notes of the first instance of the musical composition in which at least one note start time is not synchronized with any beat start time.
The processor-executable instructions and/or data stored in the non-transitory processor-readable storage medium, when executed by the at least one processor, may further cause the at least one processor to generate a variation of the first instance of the musical composition. The variation of the first instance of the musical composition may include a second sequence of notes that is different from the first sequence of notes. The timing relationship between the beats and the notes of the first instance of the musical composition may be preserved in the beats and the notes of the variation of the first instance of the musical composition.
The data file stored in the non-transitory processor-readable storage medium may have a .hum file extension.
A computer program product may be summarized as including processor-executable instructions and/or data that, when the computer program product is stored in at least one non-transitory processor-readable storage medium and executed by at least one processor communicatively coupled to the at least one non-transitory processor-readable storage medium, cause the at least one processor to: encode respective components of a musical composition in respective data objects; encode the data objects in a data file; and store the data file in at least one non-transitory processor-readable storage medium of the at least one non-transitory processor-readable storage medium. The data file stored in the non-transitory processor-readable storage medium may encode: the data objects in a hierarchically-dependent structure; each note of the musical composition in a respective note data object; and each bar of the musical composition in a respective bar data object. The first instance of the musical composition may include a first set of beats, in which case the data file stored in the non-transitory processor-readable storage medium may encode: a respective beat start time of each beat; a respective note start time of each note; and the beat start times and the note start times independently of one another to provide a timing relationship between the beats and the notes of the first instance of the musical composition in which at least one note start time is not synchronized with any beat start time.
The computer program product may further include processor-executable instructions and/or data that, when executed by the at least one processor, cause the at least one processor to generate a variation of the first instance of the musical composition. The variation of the first instance of the musical composition may include a second sequence of notes that is different from the first sequence of notes. The timing relationship between the beats and the notes of the first instance of the musical composition may be preserved in the beats and the notes of the variation of the first instance of the musical composition.
The data file stored in the non-transitory processor-readable storage medium may have a .hum file extension.
A computer-implemented method of encoding a variation of a musical composition in a digital audio format, wherein the musical composition comprises a first sequence of beats and a first sequence of notes, may be summarized as including: providing a timing relationship between the beats in the first sequence of beats and the notes in the first sequence of notes; generating the variation of the musical composition, the variation comprising: the first sequence of beats; a second sequence of notes that is different from the first sequence of notes; and the timing relationship between the first sequence of beats and the first sequence of notes applied to the first sequence of beats and the second sequence of notes; and encoding the variation of the musical composition in a digital audio file. Providing a timing relationship between the beats in the first sequence of beats and the notes in the first sequence of notes may include providing a timing relationship in which a start time of at least one note in the first sequence of notes is not synchronized with a start time of any beat in the first sequence of beats.
The method may further include encoding the musical composition in the digital audio file before generating the variation of the musical composition. Encoding the variation of the musical composition in a digital audio file may include replacing the musical composition with the variation in the digital audio file.
Generating a variation of the musical composition may include: generating the second sequence of notes that is different from the first sequence of notes; and applying the timing relationship between the first sequence of beats and the first sequence of notes to the first sequence of beats and the second sequence of notes.
A computer-implemented method of encoding a harmonic structure of a musical composition in a digital data structure, wherein the musical composition includes a sequence of bars, may be summarized as including: determining a respective harmonic fingerprint of each bar in the musical composition; comparing the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars; assigning a respective harmonic equivalence category to each bar in each respective set of harmonically equivalent bars; and encoding the respective harmonic equivalence category of each bar in the digital data structure. The digital data structure may include a set of bar data objects with each bar of the musical composition encoded in a respective bar data object. Encoding the respective harmonic equivalence category of each bar in the digital data structure may include encoding the respective harmonic equivalence category of each bar in a respective corresponding bar data object. The digital data structure may further include a set of segment data objects with each segment data object encoding a respective sequence of bar data objects, in which case the method may further include encoding a respective sequence of harmonic equivalence categories in each segment data object.
Assigning a respective harmonic equivalence category to each bar in each respective set of harmonically equivalent bars may include, for respective pairs of bars in the musical composition: assigning a first bar in the pair of bars and a second bar in the pair of bars to a same harmonic equivalence category when a comparison between the respective harmonic fingerprints of the first bar and the second bar indicates that the first bar and the second bar are harmonically equivalent to one another; and assigning the first bar in the pair of bars and the second bar in the pair of bars to different respective harmonic equivalence categories when the comparison between the respective harmonic fingerprints of the first bar and the second bar indicates that the first bar and the second bar are not harmonically equivalent to one another.
Determining a respective harmonic fingerprint of each bar in the musical composition may include determining a respective note sequence fingerprint of each bar in the musical composition. Determining a respective note sequence fingerprint of each bar in the musical composition may include, for each bar in the musical composition: sorting all notes in the bar by note start time and, for each note start time, sorting all corresponding notes by note pitch. Comparing the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars may include, for respective pairs of bars in the musical composition, determining a difference between: a maximum of a first length of a first note sequence fingerprint of a first bar in the pair of bars and a second length of a second note sequence fingerprint of a second bar in the pair of bars; and a length of a longest common note sequence between the first note sequence fingerprint and the second note sequence fingerprint.
Determining a respective harmonic fingerprint of each bar in the musical composition may include determining a respective note envelope fingerprint of each bar in the musical composition. Determining a respective note envelope fingerprint of each bar in the musical composition may include, for each bar in the musical composition: sorting all notes in the bar by note start time and, for each note start time, determining a respective highest pitch note and a respective lowest pitch note; and interpolating between respective lowest pitch notes at successive note start times and between respective highest pitch notes at successive note start times to determine an envelope for the bar. Comparing the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars may include, for respective pairs of bars in the musical composition: creating a first graphic of a first envelope of a first note envelope fingerprint of a first bar in the pair of bars, the first graphic having first pixel dimensions; creating a second graphic of a second envelope of a second note envelope fingerprint of a second bar in the pair of bars, the second graphic having the first pixel dimensions; and determining a distance between respective pixel values of the first graphic and the second graphic.
Determining a respective harmonic fingerprint of each bar in the musical composition may include determining a respective note count distribution fingerprint of each bar in the musical composition. Determining a respective note count distribution fingerprint of each bar in the musical composition may include, for each bar in the musical composition: determining a number of instances of each unique note in the bar. Comparing the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars may include: determining a mean count distribution of each unique note in the musical composition; and for respective pairs of bars in the musical composition, determining a sum of: a Kullback-Leibler Divergence between a first note count distribution fingerprint of a first bar in the pair of bars and the mean count distribution; and a Kullback-Leibler Divergence between the mean count distribution and a second note count distribution fingerprint of a second bar in the pair of bars.
Determining a respective harmonic fingerprint of each bar in the musical composition may include determining a respective note energy distribution fingerprint of each bar in the musical composition. Determining a respective note energy distribution fingerprint of each bar in the musical composition may include, for each bar in the musical composition: determining a respective note energy of each note in the bar, wherein determining a respective note energy of each note in the bar includes, for each note in the bar: determining, for each instance of the note in the bar, a respective product of note duration multiplied by note volume; and determining a sum of the respective products. Comparing the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars may include: determining a mean note energy distribution, wherein determining a mean note energy distribution includes determining a respective mean note energy of each unique note in the musical composition; and for respective pairs of bars in the musical composition, determining a sum of: a Kullback-Leibler Divergence between a first note energy distribution fingerprint of a first bar in the pair of bars and the mean note energy distribution; and a Kullback-Leibler Divergence between the mean note energy distribution and a second note energy distribution fingerprint of a second bar in the pair of bars.
Determining a respective harmonic fingerprint of each bar in the musical composition may include determining a respective audio fingerprint of each bar in the musical composition. Determining a respective audio fingerprint of each bar in the musical composition may include, for each bar in the musical composition, encoding the bar in a respective audio object. Comparing the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars may include, for respective pairs of bars in the musical composition: determining a first power spectrum of a first audio object corresponding to a first audio fingerprint of a first bar in the pair of bars; determining a second power spectrum of a second audio object corresponding to a second audio fingerprint of a second bar in the pair of bars; and determining a Euclidean distance between the first power spectrum and the second power spectrum.
Determining a respective harmonic fingerprint of each bar in the musical composition may include determining a respective graphical fingerprint of each bar in the musical composition. Determining a respective graphical fingerprint of each bar in the musical composition may include, for each bar in the musical composition: encoding the bar in a respective graphical image that visually represents musical information contained within the bar, each respective graphical image having a same number of pixels. Comparing the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars may include, for respective pairs of bars in the musical composition: determining a distance between a first graphical image corresponding to a first graphical fingerprint of a first bar in the pair of bars and a second graphical image corresponding to a second graphical fingerprint of a second bar in the pair of bars.
A system for encoding a harmonic structure of a musical composition in a digital data structure, wherein the musical composition includes a sequence of bars encoded in the digital data structure, may be summarized as including: at least one processor; and a non-transitory processor-readable storage medium communicatively coupled to the at least one processor, the non-transitory processor-readable storage medium storing processor-executable instructions and/or data that, when executed by the at least one processor, cause the at least one processor to: determine a respective harmonic fingerprint of each bar in the musical composition; compare the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars; assign a respective harmonic equivalence category to each bar in each respective set of harmonically equivalent bars; and encode the respective harmonic equivalence category of each bar in the digital data structure. The digital data structure may include a set of bar data objects with each bar of the musical composition encoded in a respective bar data object, and the processor-executable instructions and/or data that, when executed by the at least one processor, cause the at least one processor to encode the respective harmonic equivalence category of each bar in the digital data structure, may cause the at least one processor to encode the respective harmonic equivalence category of each bar in a respective corresponding bar data object.
A computer program product for encoding a harmonic structure of a musical composition in a digital data structure, wherein the musical composition includes a sequence of bars encoded in the digital data structure, may be summarized as including: processor-executable instructions and/or data that, when the computer program product is stored in a non-transitory processor-readable storage medium and executed by at least one processor communicatively coupled to the non-transitory processor-readable storage medium, cause the at least one processor to: determine a respective harmonic fingerprint of each bar in the musical composition; compare the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars; assign a respective harmonic equivalence category to each bar in each respective set of harmonically equivalent bars; and encode the respective harmonic equivalence category of each bar in the digital data structure. The digital data structure may include a set of bar data objects with each bar of the musical composition encoded in a respective bar data object, and the processor-executable instructions and/or data that, when executed by the at least one processor, cause the at least one processor to encode the respective harmonic equivalence category of each bar in the digital data structure, may cause the at least one processor to encode the respective harmonic equivalence category of each bar in a respective corresponding bar data object.
The various elements and acts depicted in the drawings are provided for illustrative purposes to support the detailed description. Unless the specific context requires otherwise, the sizes, shapes, and relative positions of the illustrated elements and acts are not necessarily shown to scale and are not necessarily intended to convey any information or limitation. In general, identical reference numbers are used to identify similar elements or acts.
The following description sets forth specific details in order to illustrate and provide an understanding of the various implementations and embodiments of the present systems, devices, and methods. A person of skill in the art will appreciate that some of the specific details described herein may be omitted or modified in alternative implementations and embodiments, and that the various implementations and embodiments described herein may be combined with each other and/or with other methods, components, materials, etc. in order to produce further implementations and embodiments.
In some instances, well-known structures and/or processes associated with computer systems and data processing have not been shown or provided in detail in order to avoid unnecessarily complicating or obscuring the descriptions of the implementations and embodiments.
Unless the specific context requires otherwise, throughout this specification and the appended claims the term “comprise” and variations thereof, such as “comprises” and “comprising,” are used in an open, inclusive sense to mean “including, but not limited to.”
Unless the specific context requires otherwise, throughout this specification and the appended claims the singular forms “a,” “an,” and “the” include plural referents. For example, reference to “an embodiment” and “the embodiment” include “embodiments” and “the embodiments,” respectively, and reference to “an implementation” and “the implementation” include “implementations” and “the implementations,” respectively. Similarly, the term “or” is generally employed in its broadest sense to mean “and/or” unless the specific context clearly dictates otherwise.
The headings and Abstract of the Disclosure are provided for convenience only and are not intended, and should not be construed, to interpret the scope or meaning of the present systems, devices, and methods.
The various embodiments described herein provide systems, devices, and methods for digital representations of music. Such digital representations include computer-readable digital audio file formats that may be readily communicated between computers and electronic instruments, and that also encode both: a) the higher-level musical concepts that fundamentally define a musical composition, and b) at least some of the nuance that gives a particular instance of the musical composition its character and expression. As will be described in more detail below, these digital audio file formats are particularly well-suited for use in computer-based composition of music, where such composition may be performed manually by a human user of a computer system, automatically by algorithms and software (e.g., artificial intelligence techniques) executed by the computer system, or by a combination of both manual (i.e., human-based) and automatic (e.g., AI-based) process steps. Algorithms and software for automatic computer-based composition of music exist in the art today but the compositions they produce tend to sound formulaic and unnatural/uninteresting to human listeners. The various implementations described herein enable computer algorithms and software to compose more sophisticated music that humans can more readily enjoy, and therefore improve the functioning of computer systems for the specific practical application of composing music.
Additionally, the various embodiments described herein provide systems, devices, and methods for encoding harmonic structure in digital representations of music. Harmonic structure is characterized by identifying tonal and/or rhythmic qualities that recur throughout a musical composition, and in some implementations a first digital representation of music that is based on musical structure may be converted into a second digital representation of music that is based on harmonic structure.
Throughout this specification and the appended claims, the term “encoding” (and variants, such as “to encode”) is often used to describe an act, step, process, or technique carried out on or with some information by a computer processor, as in for example, “encoding a musical composition in a digital file format.” Generally, “encoding” (and variants, such as “to encode”) by a computer processor means representing information as, or converting information into, a form that may be: i) further processed or manipulated by a program or application software executed by the computer processor, ii) stored by the computer processor in a non-transitory processor-readable storage medium communicatively coupled to the computer processor; and/or iii) transmitted by the computer processor. Depending on the specific context, the form into which information is encoded by a computer processor may take on a wide range of different forms, including without limitation: processor-executable instructions such as computer code or program code; program data; digital data represented using the binary number system; symbolic data represented using a finite alphabet of symbols; a structured representation containing symbolic and/or numeric data; as a graph or tree of symbols from a finite alphabet, whose nodes represent symbols and whose edges represent hierarchical relationships between symbols; and/or a transmittable digital form such as a series of impulses.
Returning to
At 101, respective components of the musical composition are encoded in respective data objects. The respective components of the musical composition may be encoded in respective data objects by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled. Examples of specific components of a musical composition that may be encoded in respective data objects are described in depth later on and may vary in different implementations, but in general such components may include concepts from modern musical theory such as any or all of: notes, percussion events, chords, bars, tracks, segments (e.g., intro, verse, pre-chorus, chorus, bridge, middle8, solo, outro, etc.), and the like.
Throughout this specification and the appended claims, reference is often made to a “data object.” Unless the specific context requires otherwise, the term “data object” is used herein to refer to a collection or set of data that is combined or amalgamated into a discretely addressable object. A data object may exist in the source code, object code, computer code, and/or program code of a computing environment where it is defined, interpreted, and manipulated, and a data object may have an analog or physical counterpart in the non-transitory processor-readable storage medium where it is stored and called or operated upon. In this case, “encoding” a component of a musical composition in a data object may include writing, by at least one processor, the component to the data object in the non-transitory processor-readable storage medium and/or forming, by at least one processor, the component(s) in the data object stored in the non-transitory processor-readable storage medium. As will be described in more detail later on, one data object may include (e.g., contain, encompass, reference, or invoke) one or more additional data object(s).
At 102, the data objects are encoded in a data file. The data objects may be encoded in the data file by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled.
Throughout this specification and the appended claims, reference is often made to a “data file.” Unless the specific context requires otherwise, the term “data file” is used herein to refer to an organized or structured collection or set of data objects (as well as any additional data that may be included, such as input data, output data, metadata, and so on) that are combined or amalgamated into a computer file. A data file may be stored in a non-transitory processor-readable storage medium, transmitted between computer systems, and used by one or more computer programs, applications, or systems. In this case, “encoding” a data object in a data file may include writing, by at least one processor, the data object to the data file and/or forming, by at least one processor, the data object in the data file.
At 103, the data file is stored in a non-transitory processor-readable storage medium. The data file may be stored in the non-transitory processor-readable storage medium by the non-transitory processor-readable storage medium itself, and/or by at least one processor communicatively coupled to the non-transitory processor-readable storage medium. The non-transitory processor-readable storage medium may be local to the at least one processor and communicatively coupled thereto by electrical communication conduits such as a computer bus and/or other tangible computer hardware, or the non-transitory processor-readable storage medium may be remote and physically-separated from the at least one processor and communicatively coupled thereto by a communication network, such as a cellular network or the internet.
Method 100 provides an illustrative example of producing a digital representation of music in which respective components of a musical composition are encoded in respective data objects within a data file. The structure of the resulting data file is unlike that of other digital audio file formats available today, and accordingly such a data file may be identified by a new type of file extension. A data file constructed in accordance with the present systems, devices, and methods may, in some implementations, be hierarchical in structure and may be given a .hum file extension. That is, in some implementations of method 100, at 102 the data objects may be encoded in a data file having a .hum file extension, and at 103 the data file having the .hum file extension may be stored in the non-transitory processor-readable storage medium. A .hum file format may stand for a Hierarchical Universal Music file format. Exemplary hierarchical aspects of the digital representations of music described herein are introduced in
Generally, acts 201, 202, and 203 of method 200 are substantially similar to acts 101, 102, and 103, respectively, of method 100 from
For example, at 101 of method 100 respective components of the musical composition are generally encoded in respective data objects. Act 201 of method 200 provides a more specific illustrative implementation in which respective components of the musical composition are encoded in a set of hierarchically-dependent data objects. The respective components of the musical composition may be encoded in a set of hierarchically-dependent data objects by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled. Throughout this specification and the appended claims, the term “hierarchically-dependent” refers to an organizational scheme in which some objects (e.g., data objects) contain, encompass, refer to, invoke, or depend on (i.e., generally and hereinafter “include”) one or more other object(s) (e.g., data object(s)) to give rise to a hierarchical structure having, for example, a top-most object and at least one bottom-most object. As will be described in more detail later on, in some implementations the resulting structure may be a tree structure or pyramid structure characterized by at least one top-most object and many bottom-most objects, with various layers of intermediate objects therebetween. In some implementations, hierarchically-dependent data objects may be hierarchically-nested. For greater certainty, in a “set of hierarchically-dependent data objects,” some data objects are dependent on other data objects and some data objects may be independent of other data objects. A person of skill in the art will appreciate that in the hierarchically-dependent “tree structures” described and illustrated herein objects at different levels (e.g., an object nested within another object and/or vertically-separated from another object in illustration) exhibit hierarchical dependence while objects at a same or common level (e.g., not nested within a particular other object and/or only horizontally-separated from a particular other object in illustration) exhibit at least some independence.
The set of hierarchically-dependent data objects may include, for example, a first data object, a second data object that includes (i.e., contains, encompasses, refers to, invokes, and/or depends on) the first data object, and a third data object that includes the second data object. In this example, encoding respective components of the musical composition in the set of hierarchically-dependent data objects at 201 may include encoding a first component of the musical composition in the first data object; encoding a second component of the musical composition in the second data object, the second data object including the first data object; and encoding a third component of the musical composition in the third data object, the third data object including the second data object. If a fourth data object is also included in the third data object and not otherwise connected to the first data object or the second data object, then the fourth data object may be said to be hierarchically-dependent on the third data object but exhibiting at least some independence from the second data object and the first data object.
In addition to specifying an exemplary hierarchically-dependent structure for the data objects, method 200 also breaks main act 201 into sub-acts 211 and 212 to provide illustrative examples of particular components of the musical composition that may be encoded in the set of hierarchically-dependent data objects. Sub-acts 211 and 212 are intended to be illustrative only and are not meant to represent either: a) an exhaustive set of all sub-acts that may be included in or performed as part of main act 201, or b) an exhaustive set of all components of the musical composition that may be encoded in the set of hierarchically-dependent data objects in accordance with the present systems, devices, and methods.
At 211, parameters of each note of the musical composition are encoded in a respective note data object. In other words, for each note in the musical composition, a corresponding note data object is defined/created and parameters of that note are encoded in the corresponding note data object. The parameters of each note of the musical composition may be encoded in a respective note data object by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled.
Throughout this specification and the appended claims, unless the specific context requires otherwise the term “note” is generally used to refer to a musical note (such as Ab, A, A #, Bb, B, C, C #, Db, D, D #, Eb, E, F, F #, Gb, G, G # (of any octave), and theoretical notes such as Cb, which is enharmonic to B) and is inclusive of rests (i.e., a note with a certain timing but no pitch or volume). A person of skill in the art will appreciate that the “parameters” of a note, or “note parameters,” may include any or all concepts used to characterize notes in modern musical theory, including without limitation: pitch, start time, stop time, duration, volume, attack, reverb, decay, sustain, and instrument (e.g., tone, timbre, relative harmonics, and the like). Thus, a “note data object” is a data object that encodes a note, including its applicable note parameters.
A musical composition may include percussion events that have no tonal pitch but are used to impart rhythm. Throughout this specification and the appended claims, unless the specific context requires otherwise the term “note” is inclusive of percussion events. A percussion event may be defined or characterized by note parameters that generally do not include a pitch and generally specify a percussion instrument as the instrument. A “note data object” as described herein may be used to encode notes and note parameters, as well as percussion events and percussion event parameters. Thus, unless the specific context requires otherwise, a “note data object” may encode a tonal note event or a rhythmic percussion event depending on the specific implementation; however, in some implementations of the present systems, devices, and methods notes and percussion events may be encoded in different data objects. In implementations in which notes and percussion events are encoded in different data objects, note parameters may be encoded in note data objects and percussion event parameters may be encoded in “percussion event data objects.” The parameters used as “percussion event parameters” may be substantially similar to the parameters used as note parameters (described above), with the exception that percussion event parameters may not include pitch (and, of course, the instrument specified for a percussion event parameter will generally be a percussion instrument). Thus, a “percussion event data object,” if and when implemented in addition to note data objects, is a data object that encodes a percussion event, including its applicable percussion event parameters.
At 212, parameters of each bar of the musical composition are encoded in a respective bar data object. In other words, for each bar in the musical composition, a corresponding bar data object is defined/created and parameters of that bar are encoded in the corresponding bar data object. The parameters of each bar of the musical composition may be encoded in a respective bar data object by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled. As will be discussed in more detail later on, each bar in at least a first set of bars of the musical composition may include a respective sequence of notes, and at 212 encoding parameters of each bar of the musical composition in a respective bar data object may include, for each bar in the first set of bars of the musical composition, encoding a respective sequence of note data objects in a respective bar data object.
Throughout this specification and the appended claims, unless the specific context requires otherwise the term “bar” is generally used to refer to a musical bar; i.e., a portion of time comprising a set number of beats from a musical composition. The number of beats in a bar depends on the time signature for the musical composition. A person of skill in the art will appreciate that the “parameters” of a bar, or “bar parameters,” may include any or all concepts used to characterize bars in modern musical theory, including without limitation: bar index, time signature, beats per minute, duration, start time, stop time, beat times, key, scale, chords, tracks, sequence of notes, and (if applicable) sequence of percussion events. Thus, a “bar data object” is a data object that encodes a bar, including its applicable bar parameters.
Sub-acts 211 and 212 of method 200 specify examples of certain components of a musical composition that may be encoded in respective hierarchically-dependent data objects at 201; namely, notes and bars, respectively. Various implementations of the present systems, devices, and methods may encode notes and bars in respective data objects, and various implementations of the present systems, devices, and methods may encode additional or alternative components of a musical composition in respective data objects. For example, an arrangement of a musical composition may be encoded in an arrangement data object. The arrangement may be represented by, for example, at least one sequence of bar data objects and encoding respective components of the musical composition in a set of hierarchically-dependent data objects at 201 of method 200 may include encoding the at least one sequence of bar data objects in the arrangement data object. As a further example, various elements or “segments” of the musical composition may be encoded in respective segment data objects. In this context, the term “segment” is used to refer to a sequence of bars (i.e., a subset of bars) that represents a particular section or portion of the musical composition, such as an intro, a verse, a pre-chorus, a chorus, a bridge, a middle8, a solo, or an outro. The section or portion of a musical composition that corresponds to a “segment” may be defined, for example, by strict rules of musical theory and/or based on the sound or theme of the musical composition. Generally, encoding respective components of the musical composition in a set of hierarchically-dependent data objects at 201 of method 200 may include encoding each segment (e.g., intro, verse, pre-chorus, chorus, bridge, middle8, solo, or outro) of the musical composition in a respective segment data object, and encoding at least one sequence of bar data objects in an arrangement data object may include encoding a sequence of segment data objects in the arrangement data object. Hierarchically, in this illustrative example the arrangement data object may include the segment data objects, each segment data object may include a respective set or sequence of bar data objects, and each bar data object may include a respective set or sequence of note data objects.
Following act 201, including corresponding sub-acts 211 and 212 in the illustrated implementation of
At 202, the hierarchically-dependent data objects are encoded in a data file similar to act 102 from method 100.
At 203, the data file is stored in a non-transitory processor-readable storage medium similar to act 103 from method 100.
As previously described, in implementations of digital representations of music that employ a set of hierarchically-dependent data objects in accordance with the present systems, devices, and methods, the resulting data structure may be a tree structure or pyramid structure characterized by at least one top-most object and many bottom-most objects, with various layers of intermediate objects therebetween.
List2{ } includes a set of rules {Rule1, Rule2, Rule3} that map the respective segment data objects to respective sequences of bar data objects. That is, Rule1 maps the segment data object seg1 to the sequence of bar data objects given by List3{bar11, bar12, bar13}; Rule2 maps the segment data object seg2 to the sequence of bar data objects given by List4{bar21, bar22}; and Rule3 maps the segment data object seg3 to the sequence of bar data objects given by List5{bar31, bar32, bar33, bar34}. The musical composition encoded in the Music[List1{ }, List2{ }] data object therefore corresponds to the following complete sequence of bars (with each bar encoded in and represented by a respective bar data object):
By decoupling the arrangement (i.e., List1) from the rules mapping the segments to sequences of bars (i.e., List2), either the arrangement or the rules can readily be modified to produce variations of the musical composition. For example, a first variation of the Music[ ] musical composition may be defined or encoded as v1Music[ ] and may comprise a simple rearrangement of the segment data objects in a new arrangement data object given by vList1{ }. If List2{ } remains unchanged and vList1{ } is programmed to encode an alternative sequence of the segment data objects vList1{seg2, seg3, seg1}, then the variation of the musical composition encoded in the v1Music[vList1{ }, List2{ }] data object corresponds to the following rearranged sequence of bars (with each bar encoded in and represented by a respective bar data object):
As is apparent in data structure 300, the various implementations of data structures described herein may include data objects and additional elements other than data objects, such as rules, lists, and other components or parameters. A person of skill in the art will appreciate that data structure 300 of
The bar11[ ] data object includes, encoded at a first level, various bar parameters such as: index (bar index), tSig (time signature), List7{tStart, tStop} (bar start time and bar stop time), bpm (beats per minute), length (bar duration), List8{chord1, . . . , chordW} (chords used in the bar, from chord1 up to chordW when there are W chords used, where W is an integer greater than 1), track1, . . . , trackX (tracks that may overlap in the bar and correspond to different instruments played over the same time period, from track1 up to trackX when there are X tracks used, where X is an integer greater than 1), key (key signature), List9{beat1, . . . , beatY} (beat start times in the bar, from beat1 up to beatY when the bar contains Y beats, where Y is an integer greater than 1), and scale. A person of skill in the art will appreciate that the foregoing list of bar parameters (corresponding to those which are illustrated in data structure 400) is used for illustrative purposes only and is not intended to be a limiting or exhaustive list of bar parameters that may be included in alternative implementations. Furthermore, a person of skill in the art will appreciate that the order in which parameters are presented in the bar11[ ] data object, and generally in any data object throughout this specification including the Figures, is used for illustrative purposes only and not intended to be limiting in any way. Different implementations of the present systems, devices, and methods may arrange lists and/or parameters in orders other than the orders used for illustrative purposes herein.
As illustrated in data structure 400, the bar parameters of the bar11[ ] data object include further data objects, some of which have their own respective hierarchically-dependent data objects. For example, the trackX[ ] data object includes List10{note1, . . . , noteZ} (a sequence of Z notes included in trackX of bar11, where Z is an integer greater than 1). The note1 data object further includes a set of note parameters: pitch (the pitch of note1, such as A # or F), vol (the volume of note1), tStart (the start time of note1), tStop (the stop time of note1), and an instr parameter specifying the instrument of note1, such as any or all of guitar, piano, percussion, vocals, and the like. In some implementations, the instr parameter may be any instrument selected from the set of all MIDI tonal instruments and/or all MIDI percussion instruments. As previously described, in some implementations percussion events may be encoded in note data objects, such as note1 [ ], that omit pitch information; however, in other implementations a bar data object, such as bar11[ ], may further include percussion event data objects in addition to note data objects, such as:
Throughout this specification and the appended claims, reference is often made to a “track.” Unless the specific context requires otherwise, the term track is used herein to refer to a collection or sequence of notes that are all “played by” the same instrument in a musical composition. For example, a musical composition that is for or by a single instrument may have only one track, but a musical composition that is for or by multiple instruments concurrently may have multiple tracks that are temporally overlaid on one another. Each respective bar of a musical composition may include multiple tracks, where each track provides the sequence of notes of a respective instrument throughout the duration of that bar. Thus, in the present systems, device, and methods, a bar data object may include at least two tracks (which may be encoded, for example, as bar parameters or as track data objects), and each respective track (or each respective track data object) may include a respective list or sequence of note data objects encoding note parameters that specify: i) note start/stop times that span the duration of the same bar; and ii) different pitches or instruments. In implementations in which note data objects and percussion event data objects are encoded separately, a track (or track data object) may include a list of percussion event data objects encoding percussion event parameters. In data structure 400, the instr parameter is defined at the note level (e.g., as shown for the exemplary note1 data object); however, in alternative implementations the instr parameter may be defined at the track level (e.g., trackX[{note1, . . . , noteZ}, instr] and applied “globally” to all notes in the track. The instr parameter of note1 is shown in a dotted circle in
In some implementations, chords may be defined by chord parameters that are encoded in chord data objects (e.g., hierarchically-nested within bar data objects). Exemplary chord parameters that may be encoded in a chord data object include, without limitation: root note of the chord, scale degrees of the chord, key, scale, chord notes, chord inversion, chord bass octave, chord timing (arpeggiation pattern), chord instrument, and/or chord volume. The key and scale of a chord (or chords) may or may not match that of the bar that contains it (or them). This allows for the use of so-called “borrowed chords.” Generally, it is possible (if desired) to have a plurality of chords in a single bar. The plurality of chords could be stacked in time or started consecutively within the temporal boundaries of the enclosing bar.
A person of skill in the art will appreciate that data structure 400 of
Data structure 300 of
In some implementations, a complete musical composition may be encoded in a single data object (e.g., a single top-level Music[ ] data object that includes any number of hierarchically-dependent data objects therein or thereunder); however, in other implementations the encoding of a complete musical composition may span multiple data objects, such as {Music1[ ], . . . , MusicN[ ]}, where N is an integer greater than 1. In this case, depending on the particular encoding, there may be an even higher-level data object that includes the set of {Music1[ ], . . . , MusicN[ ]} data objects, or the top-level of the encoded data structure may simply include a list of all Music[ ] data objects in the set of {Music1[ ], . . . , MusicN[ ]} data objects.
The various implementations described herein include systems, devices, and methods for digital audio file formats in which respective components of a musical composition are encoded in respective data objects and the respective data objects are organized in a hierarchically-dependent data structure. The structure of these hierarchical digital audio file formats has numerous advantages over established digital audio file formats (e.g., .wav, .mid, .mp3, and the like) in relation to certain applications, such as in the computer-based creation of musical compositions. For example, with various concepts from modern musical theory (e.g., notes, bars, segments, arrangements, and the like) each corresponding to respective data objects, such data objects may readily be rearranged, manipulated, and/or individually modified to produce musical variations of a composition. Furthermore, certain data objects and/or parameters within data objects (e.g., timing parameters) may be independently adjusted (e.g., misaligned relative to one another) in order to introduce nuance and expressiveness to a musical composition that may exist in a typical instance of a human performance but may not otherwise be easily transcribable in sheet music. Exemplary methods that take advantage of these capabilities of the hierarchical digital representations of music described herein are illustrated in
Method 600 includes three main acts 601, 602, and 603. Main act 601 includes two sub-acts 611 and 612; sub-act 611 includes two additional sub-acts 621 and 622; and sub-act 612 includes two additional sub-acts 631 and 632. Those of skill in the art will appreciate that in alternative implementations certain acts and/or sub-acts may be omitted and/or additional acts and/or sub-acts may be added. Those of skill in the art will also appreciate that the illustrated order of the acts and/or sub-acts is shown for exemplary purposes only and may change in alternative implementations.
Generally, acts 601, 602, and 603 of method 600 are substantially similar to acts 201, 202, and 203, respectively, of method 200 from
At 601, respective components of the musical composition are encoded in a set of hierarchically-dependent data objects similar to act 201 of method 200. Also similar to method 200, method 600 likewise breaks main act 601 into sub-acts 611 and 612 to provide illustrative examples of particular components of the musical composition that may be encoded in the set of hierarchically-dependent data objects. At 611, parameters of each note of the musical composition are encoded in a respective note data object similar to sub-act 211 of method 200, and at 612, parameters of each bar of the musical composition are encoded in a respective bar data object similar to sub-act 212 of method 200.
Sub-act 611 includes two additional sub-acts, 621 and 622. That is, in method 600, the act of encoding parameters of each note of the musical composition in a respective note data object at 611 involves performing sub-acts 621 and 622, each of which relates to the encoding of a respective specific parameter of the first note of the musical composition in a first note data object.
At 621, a pitch of the first note is encoded in the first note data object and at 622, a start time of the first note is encoded in the first note data object. Pitch and start time are respective parameters of the first note. The pitch and start time of the first note may be encoded in the first note data object by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled.
In a way similar to sub-act 611, sub-act 612 includes two additional sub-acts, 631 and 632. In method 600, the act of encoding parameters of each bar of the musical composition in a respective bar data object at 612 involves performing sub-acts 631 and 632, each of which relates to the encoding of a respective specific parameter of the first bar of the musical composition in a first bar data object.
At 631, the first note data object is encoded in the first bar data object and at 632, a respective start time of each beat in the first set of beats is encoded in the first bar data object. The first note and the start times of the beats in the first set of beats are parameters of the first bar. The first note data object and the respective start time of each beat in the first set of beats may be encoded in the first bar data object by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled.
An exemplary advantage of the present systems, devices, and methods for digital representations of music is that the note parameters (e.g., the start time of the first note) and the bar parameters (e.g., the respective start times of each beat in the first set of beats) may be defined and programmed substantially independently of one another. In conventional sheet music, the time signature of a piece defines a number of beats per bar, the beats are precisely and uniformly spaced within each bar, and the notes are precisely synchronized with or counted from the beats. That is, in conventional sheet music the start time of each note is synchronized with (e.g., temporally aligned with or snapped to) or counted from (e.g., based on or derived from) the start time of a respective beat. Sheet music transcribes an idealization of a musical composition that is rarely if ever achieved in a performance by a human musician. The imperfection of human performance typically results in slight variations, misalignments, or “unsynchronizations” between note start times and beat times which can add nuance or expression that may actually be enjoyed by the listener. Indeed, when a machine or computer plays or performs sheet music, the result is typically “too perfect” (with all note timing perfectly synchronized with or counted from all beat timing) and the performance ends up sounding unnaturally perfect and robotic. In accordance with the present systems, devices, and methods, digital representations of music in which the note timing and the beat timing are defined and programmed independently of one another readily enables computer-based generation of musical compositions that, when played back by a machine or computer, have a more nuanced and expressive sound (compared to other approaches in the art) that may be preferred by a human listener.
Method 600 provides a specific exemplary implementation of the unsynchronization feature described above. At 622, a start time of the first note is encoded in the first note data object and at 632, a respective start time of each beat in the first set of beats is encoded in the first bar data object. In method 600, the start time of the first note is deliberately not synchronized with the respective start time of any of the beats in the first set of beats.
Throughout this specification and the appended claims, reference is often made to a note (e.g., a first note) having a start time that is “not synchronized with” the start time of any beat. In this context, “synchronized” generally means aligned in time to an extent necessary to be perceived as synchronous by the human ear, and “not synchronized” or “unsynchronized” means unaligned or misaligned in time to an extent necessary to be perceived as asynchronous by the human ear. The precise quantified timing boundary between “synchronized” and “not synchronized” or “unsynchronized” may vary in different implementations and may depend on other factors, such as the note parameters being employed. As a first example, in some implementations “synchronized” may mean aligned in time within ±100 ms (e.g., if the beat start time is 01:12:000, then the corresponding note start time is somewhere in between 01:11:900-01:12:100) and “not synchronized” or “unsynchronized” may mean not aligned in time within ±100 ms (e.g., if the beat start time is 01:12:000, then the corresponding note start time is somewhere outside of 01:11:900-01:12:100). As a second example, in some implementations “synchronized” may mean aligned in time within ±50 ms (e.g., if the beat start time is 01:12:000, then the corresponding note start time is somewhere in between 01:11:950-01:12:050) and “not synchronized” or “unsynchronized” may mean not aligned in time within ±50 ms (e.g., if the beat start time is 01:12:000, then the corresponding note start time is somewhere outside of 01:11:950-01:12:050). As a third example, in some implementations “synchronized” may mean aligned in time within ±10 ms (e.g., if the beat start time is 01:12:000, then the corresponding note start time is somewhere in between 01:11:990-01:12:010) and “not synchronized” or “unsynchronized” may mean not aligned in time within ±10 ms (e.g., if the beat start time is 01:12:000, then the corresponding note start time is somewhere outside of 01:11:990-01:12:010).
In some implementations, for example when the first note has a length or duration that is at least as long as the time interval that separates adjacent beats in the first bar (e.g., the first note is a whole note, a half note, or a quarter note), the programmed “nuance” or “expressiveness” desired may be achieved when the start time of the first note is not synchronized (or “unsynchronized”) with the respective start time of any of the beats in the first set of beats. In other implementations, for example when the first note has a length or duration that is shorter than the time interval that separates adjacent beats in the first bar (e.g., the first note is an eighth note, a sixteenth note, or a thirty-second note), the “unsynchronization” feature may be extended beyond the start times of beats to include specific points in the time intervals between adjacent beats. For example, if the first note has a length or duration characterized by ½N of the time interval that separates adjacent beats in the first bar (where N is an integer and =5, such that: N=1 corresponds to an eighth note, N=2 corresponds to a sixteenth note, N=3 corresponds to a thirty-second note, N=4 corresponds to a sixty-fourth note, and N=5 corresponds to a one hundred twenty-eighth note), in some implementations the start time of the first note is advantageously not synchronized with any point in time corresponding to a ½N fraction of the time interval between adjacent beats in the first set of beats. This feature is particularly relevant when there is deliberate misalignment between note start times and beat start times in a musical composition (i.e., even transcribed in the corresponding sheet music), such as for example, when a note is deliberately positioned in a time interval between adjacent beats. A specific example of such a scenario occurs when an eighth note is positioned in between two adjacent beats in a bar in 4/4 time. In accordance with the present systems, devices, and methods, the start time of an eighth note that is deliberately positioned in between adjacent beats may advantageously be deliberately not synchronized with the point in time corresponding to ½ of the time interval between the adjacent beats. This concept may be expanded to encompass any point in time corresponding to a ½N fraction of the time interval between adjacent beats such that, for example, for a bar in 4/4 time:
Returning to method 600, at 602 the hierarchically-dependent data objects are encoded in a data file similar to act 202 from method 200, with the added detail that in the data file of method 600, the start time of the first note is not synchronized with the respective start time of any of the beats in the first set of beats.
At 603, the data file is stored in a non-transitory processor-readable storage medium similar to act 203 from method 200.
As described previously, the structure of the digital representations of music described herein gives rise to numerous advantages, particularly in the context of computer-based composition of music. Method 600 provides an example of how such structure enables deliberate timing features (e.g., a timing relationship in which a note start time is not synchronized with a beat start time) that can enhance the expressiveness and “humanness” of a computer-based composition. As an extension of this,
Method 700 includes two main acts 701 and 702. Main act 702 includes two sub-acts 721 and 722. Those of skill in the art will appreciate that in alternative implementations certain acts and/or sub-acts may be omitted and/or additional acts and/or sub-acts may be added. Those of skill in the art will also appreciate that the illustrated order of the acts and/or sub-acts is shown for exemplary purposes only and may change in alternative implementations.
At 701, a variation of the first bar is generated. The variation of the first bar comprises the same first set of beats from the original first bar but a new (i.e., a second) sequence of notes that is different from the original first sequence of notes. Thus, generating the variation of the first bar may include generating a second sequence of notes that is different from the first sequence of notes. The variation of the first bar may be generated by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled. The at least one processor may generate the variation of the first bar in response to commands and instructions input by a user, or automatically by executing one or more music composition algorithms or programs stored in a non-transitory processor-readable storage medium communicatively coupled to the at least one processor. Generally, a second sequence of notes is considered different from a first sequence of notes if at least one note parameter (e.g., pitch, volume, start time, duration/length, or stop time) of at least one note in the second sequence of notes is different from a corresponding parameter of a corresponding note in the first sequence of notes.
At 702, the variation of the first bar is encoded in the data file. The variation of the first bar may be encoded in the data file by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled. The data file may be stored in the non-transitory processor-readable storage medium.
Main act 702 includes two sub-acts 721 and 722. In method 700, the act of encoding the variation of the first bar in the data file involves performing sub-acts 721 and 722, each of which relates to encoding specific components and/or parameters of the variation in the data file.
At 721, the respective start times of the beats in the first set of beats are encoded in the data file (e.g., in a corresponding bar data object or in respective beat data objects). More explicitly, the same first set of beats with the same start times from the original instance of the first bar in the original musical composition are encoded in the data file at 721. The respective start times of the beats in the first set of beats may be encoded in the data file by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled.
At 722, a respective start time of each note in the second sequence of notes is encoded in the data file (e.g., in respective note data objects). The respective start time of each note in the second sequence of notes may be encoded in the data file by, for example, at least one processor, or by a combination of at least one processor and at least one non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled. Advantageously, the respective start time of each note in the second sequence of notes is matched to (e.g., synchronized with) or counted from the respective start time of a respective note in the first sequence of notes. This does not necessarily mean that the second sequence of notes has the same number of notes as the first sequence of notes, but only that the respective start time of each note in the second sequence of notes is determined by the respective start time of a respective note in the first sequence of notes. The term “counted from” in this context means that the start time of a second note is based on or derived from the start time of a first note, while the actual temporal alignment between the second note and the first note may depend on the other notes in the bar. For example, if an original bar in 4/4 time comprises four quarter notes each having a respective start time {t1, t2, t3, and t4} and a variation of the bar in 4/4 time also has four quarter notes, then each quarter note in the variation will have a respective start time that counts from (and is matched to) the respective start time of a respective quarter note in the original bar. That is, the respective start times of the quarter notes in the variation will also be {t1, t2, t3, and t4}. However, if the variation has a first eighth note, a second eighth note, a half note, and then a quarter note, then: the start time of the first eighth note vt1 may count from (and be matched to) the start time of the original first quarter note t1 (i.e., vt1=t1); the start time of the second eighth note vt2 may also count from the start time of the original first quarter note t1 and be offset therefrom by half of a beat (i.e., vt2=t1+(beat duration)/2)); the start time of the half note vt3 may count from (and be matched to) the start time of the original second quarter note t2 (i.e., vt3=t2); and the start time of the quarter note vt4 may count from (and be matched to) the start time of the original fourth quarter note t4 (i.e., vt4=t4).
In conventional musical theory, a variation such as the example described above would typically be thought of in terms of how note times map to beat times rather than how the variation note times map to the original note times. This is because in conventional musical theory, note times always match to (e.g., synchronize with) or count from beat times. However, as previously described (e.g., in the context of method 600 of
In conventional sheet music, note start times are essentially “snapped to” or precisely aligned with beat times and the timing relationship therebetween is therefore predictable and fixed. In the present systems, devices, and methods, note start times and beat times are encoded separately into the data structure of the digital audio file and the timing relationship therebetween may be deliberately crafted, designed, characterized, preserved, and/or modified. The fact that a first note having a start time that is not synchronized with any beat time exists in both method 600 and method 700 means that both method 600 and method 700 involve a timing relationship between notes and beats that is not standard in sheet music. In an exemplary implementation of method 700, the timing relationship between the beats and notes in the original musical composition (e.g., in the first bar of the original composition) may be preserved in the beats and notes of the variation. That is: i) encoding, in the first bar data object, a respective start time of each beat in the first set of beats at 632; and ii) encoding, in a respective note data object, a respective start time of each note in the first sequence of notes (e.g., at 622), together provide a timing relationship between the notes and the beats in the first bar of the original musical composition, and encoding the variation of the first bar in the data file at 702 may include applying the timing relationship between the notes and the beats of the first bar of the original musical composition to the notes and the beats of the variation, or “preserving” the timing relationship between the notes and beats of the original musical composition in the notes and beats of the variation. This concept of “preserving” a timing relationship is demonstrated by illustrative example in
In conventional sheet music, the timing of each quarter note 811a, 812a, 813a, and 814a would be precisely matched to (i.e., temporally aligned or synchronized with) the timing of a respective beat 801a, 802a, 803a, and 804. However, when first bar 800a is performed by a human musician, the human musician may (either intentionally or unintentionally) introduce discrepancies in the synchronization between certain notes and beats and these discrepancies may add a natural “human” feel to the performance. Examples of such timing discrepancies are illustrated in
In accordance with the present systems, devices, and methods, the timing relationship between the notes and beats of first bar 800a of the original musical composition is preserved in variation 800b. This means that the timing relationship between the notes and beats of variation 800b is characterized in substantially the same way as the timing relationship between the notes and beats of first bar 800a; that is: first note 811b is substantially synchronized with first beat 801b (i.e., beat1−time1=˜0); second note 812b is not synchronized with second beat 802b (or any other beat in variation 800b) but rather offset in front of (i.e., played before) second beat 802b by an amount A (i.e., beat2−time2=A, A>0); third note 813b is substantially synchronized with third beat 803b (i.e., beat3−time3=˜0); and fourth note 814b is not synchronized with fourth beat 804b (or any other beat in first bar 800b) but rather offset behind (i.e., played after) fourth beat 804b by an amount B (i.e., beat4−time4=B, B<0). In this way, at least some of the nuance and expressiveness (i.e., some of the “humanness”) of first bar 800a of the original musical composition may be automatically preserved or maintained in variation 800b.
In exemplary bar 800c, quarter note 811c is synchronized with beat 801c, eighth note 812c is synchronized with beat 802c, eighth note 813c is not synchronized with (i.e., unsynchronized with) halfway point 822c by an amount D1, eighth note 814c is synchronized with beat 803c, eighth note 815c is not synchronized with (i.e., unsynchronized with) halfway point 823c by an amount D2, and quarter note 816c is not synchronized with (i.e., unsynchronized with) beat 804c by an amount D3. Eighth note 813c is played in front of (i.e., before) halfway point 822c by amount D1, eighth note 815c is played behind (i.e., after) halfway point 823c by amount D2, and quarter note 816c is played in front of (i.e., before) beat 804c by amount D3. Based on this illustrative example, a person of skill in the art will appreciate how the concept of unsynchronization may be extended to smaller and smaller time intervals separating adjacent beats for smaller and smaller note durations (i.e., for larger and larger values of N).
If a musical composition is transcribed in sheet music, when a machine “performs” the musical composition the precise timing relationship between notes and beats imposed by the sheet music may cause the performance to sound unnaturally perfect and robotic to the human ear. Conversely, when a human musician performs the musical composition slight discrepancies in the timing between notes and beats can imbue the performance with a sense of feeling that is pleasing to the human listener. In accordance with the present systems, devices, and methods, a performance of a musical composition by a human musician may be analyzed (e.g., recorded and then analyzed) to extract the actual realized timing relationship between the notes and the beats (i.e., the timing relationship actually played by the musician, not necessarily the precise timing relationship represented in the corresponding sheet music), and then this same timing relationship may be preserved in any number of variations of the musical composition that are automatically generated by a computer algorithm or program.
Returning to
The various implementations described herein provide systems, devices, and methods that represent a musical composition in a digital audio file format by encoding respective components of the musical composition in respective data objects and encoding all of the data objects in at least one data file. In some of the implementations described above, many of the data objects are inter-related in a hierarchically-dependent data structure and the examples provided herein illustrate certain benefits that may be exploited in such a structure (e.g., the “unsynchronization” of notes and beats and the preservation of timing relationships across musical variations). However, while such benefits have so far been demonstrated in a hierarchically-dependent data structure, it is possible to take advantage of some benefits of the present systems, devices, and methods without requiring a hierarchically-dependent data structure.
Generally, acts 901, 902, and 903 of method 900 are substantially similar to acts 101, 102, and 103, respectively, of method 100 from
For example, at 901 respective components of the musical composition are encoded (e.g., by at least one processor) in respective data objects in a manner similar to act 101 of method 100; however, act 901 includes sub-act 911 that provides further specificity for the encoding of at least one component of the musical composition. At 911, at least a set of first note parameters for at least a first note of the musical composition are encoded (e.g., by at least one processor) in a first note data object. The set of first note parameters may include many different parameters depending on the particular implementation, but for the purposes of method 900 the set of first note parameters includes at least a pitch of the first note and a start time of the first note.
At 902, the data objects are encoded (e.g., by at least one processor) in a data file in a manner similar to act 102 of method 100; however, act 902 includes sub-act 921 that provides further specificity for the encoding of the data objects in the data file. At 921, a respective start time of each beat in the musical composition is encoded in the data file (e.g., in a first bar data object that includes the first note data object). In a similar way to that described for method 600, though without necessarily requiring the hierarchical dependency of the data objects, in method 900 the start time of the first note encoded at 911 is not synchronized with the respective start time of any of the beats encoded at 921. In this way, method 900 achieves the benefit of “unsynchronization” between notes and beats by separating note and beat timing parameters into different data objects without necessarily implementing a hierarchically-dependent data structure.
At 903, the data file is stored (e.g., by at least one processor, or by a non-transitory processor-readable storage medium) in a non-transitory processor-readable storage medium in a similar way to act 103 of method 100.
Method 900 provides an illustrative example of an implementation of the present systems, devices, and methods in which the benefit of “unsynchronization” between notes and beats is achieved without a hierarchically-dependent data structure. Even though method 900 does not necessarily require a hierarchically-dependent data structure, method 900 still involves encoding respective components of a musical composition in respective data objects (i.e., at act 901).
At 1001, a timing relationship between the beats in the first sequence of beats and the notes in the first sequence of notes is provided. In some implementations, this timing relationship may already be encoded in a data file (e.g., a digital audio file, such as but not limited to a .hum file) representing the original musical composition and “providing” the timing relationship at 1001 may simply involve reading or extracting the timing relationship from the digital audio file (e.g., by at least one processor). When providing the timing relationship at 1001 may be achieved by reading or extracting the timing relationship from a digital audio file, optional act 1011 may not be needed. However, in other implementations, a digital audio file encoding the original musical composition may not be available, in which case method 1000 may include optional act 1011 before act 1001.
At optional act 1011, the original musical composition (i.e., the musical composition of which a variation is encoded in method 1000) is encoded in a digital audio file. The original musical composition includes the first sequence of beats and the first sequence of notes, and therefore act 1011 includes encoding the timing relationship between the beats in the first sequence of beats and the notes in the first sequence of notes in the digital audio file. In summation, act 1001 generally requires that the timing relationship between the beats in the first sequence of beats in the original musical composition and the notes in the first sequence of notes in the original musical composition be available. If such timing relationship is already available in an existing encoding of the original musical composition, then method 1000 may skip optional act 1011 and begin at 1001. If such timing relationship is not already available in an existing encoding of the original musical composition, then method 1000 may begin at act 1011 where the timing relationship of the original musical composition is encoded.
As previously described, in accordance with the present systems, devices, and methods, the original musical composition may advantageously employ a timing relationship (i.e., the timing relationship that may be, optionally, encoded in a digital audio file at 1011 and read or extracted at 1001) between the beats in the first sequence of beats and the notes in the first sequence of notes in which a start time of at least one note in the first sequence of notes is not synchronized with a start time of any beat in the first sequence of beats.
At 1002, a variation of the musical composition is generated (e.g., by at least one processor, or by a combination of at least one processor and a non-transitory processor-readable storage medium with the at least one processor is communicatively coupled). Even though in method 1000 the respective components of the original musical composition and/or the variation are not necessarily encoded in respective data objects, a result of act 1002 may be substantially similar to the result of act 701 of method 700. The variation generated at 1002 may include the same first sequence of beats from the original musical composition and a new (i.e., a second) sequence of notes that is different from the original first sequence of notes. Furthermore, in accordance with the present systems, devices, and methods, the variation generated at 1002 may include (i.e., preserve or maintain) the timing relationship between the original first sequence of beats and the original first sequence of notes in the original musical composition applied to the first sequence of beats and the new second sequence of notes in the variation. Generally, generating a variation of the musical composition at 1002 may include the sub-acts of: i) generating the second sequence of notes that is different from the first sequence of notes; and ii) applying the timing relationship between the first sequence of beats and the first sequence of notes to the first sequence of beats and the second sequence of notes.
At 1003, the variation of the musical composition is encoded in a digital audio file. If method 1000 is performed without access to a digital audio file corresponding to the original musical composition (e.g., if optional act 1011 is not performed and the necessary timing relationship is provided by other means), or if the digital audio file corresponding to the original musical composition is to be kept, then the digital audio file encoded at 1003 may be separate from a digital audio file encoding of the original musical composition and method 1000 may terminate with act 1003. If method 1000 is performed with an intent to replace the original musical composition with the variation, then method 1000 may proceed from act 1003 to optional act 1012.
At optional act 1012, the encoding of the original musical composition is replaced (e.g., by at least one processor or by a combination of at least one processor and a non-transitory processor-readable storage medium with which the at least one processor is communicatively coupled) by the encoding of the variation in the digital audio file.
Returning to
In some implementations, segment information may be provided or otherwise made available with a musical composition; in other implementations, segment information may be automatically determined by employing the teachings of U.S. patent application Ser. No. 16/775,241, filed Jan. 28, 2020 and entitled “Systems, Devices, and Methods for Segmenting a Musical Composition into Musical Segments,” which is incorporated by reference herein in its entirety. In still other implementations, the systems, devices, and methods described herein for identifying harmonic structure in musical compositions may be applied to an “unsegmented” musical composition and ultimately used to influence or determine the segmentation of the musical composition. In other words, in some implementations the systems, devices, and methods described herein for identifying harmonic structure in musical compositions may be applied before a musical composition is segmented and the resulting harmonic structure may subsequently be used to determine or influence the segmentation of the musical composition.
In accordance with the present systems, devices, and methods, harmonic structure in a musical composition may be characterized by identifying tonal and/or rhythmic qualities that recur throughout the musical composition. An exemplary method for achieving this is illustrated in
At 1101, a respective harmonic fingerprint of each bar in the musical composition is determined, e.g., by at least one processor. Throughout this specification and the appended claims, the term “harmonic fingerprint” is used to refer to an aggregated, abridged, condensed, or otherwise concise representation of the musical elements in a section of music. A harmonic fingerprint may be used to characterize a musical quality of a section of music, where the section of music may be all or a portion of a musical composition. In the case of method 1100, the section of music is a bar of a musical composition and at 1101 a respective harmonic fingerprint is determined for each bar in the musical composition.
In accordance with the present systems, devices, and methods, a harmonic fingerprint may take on a wide range of different forms determined in a wide range of different ways. In different implementations of method 1100 and/or for different musical compositions, different harmonic fingerprints may advantageously be used.
In accordance with the present systems, devices, and methods, where one or more notes are used in a harmonic fingerprint (such as note sequence fingerprint 1111, note envelope fingerprint 1112, note count distribution fingerprint 1113, and note energy distribution fingerprint 1114), each note may or may not include octave information depending on the specific implementation. That is, in some implementations it may be advantageous for a note to include its octave information (e.g., Cx, where x is a natural number that denotes the octave of the particular C, such as C1, C2, C3, and so on) while in other implementations it may be advantageous for octave information to be omitted (i.e., for notes to be treated as “octaveless”).
At 1102, the respective harmonic fingerprints of respective bars in the musical composition determined at 1101 are compared (e.g., to one another) to identify sets of harmonically equivalent bars. The comparison may be performed, for example, by at least one processor. The exact manner or mechanism by which the respective harmonic fingerprints are compared may depend on the exact type or nature of the harmonic fingerprints determined at 1101.
At 1103, a respective harmonic equivalence category is assigned (e.g., by at least one processor) to each bar in each respective set of harmonically equivalent bars identified at 1102. Equivalently, each respective bar in the musical composition may be assigned (e.g., by at least one processor) to a respective harmonic equivalence category at 1103. The exact logic behind the mapping between bars and harmonic equivalence categories may vary from implementation to implementation, but as an illustrative example
Exemplary rule 1131 specifies that, as a rationale for how to assign a respective harmonic equivalence category to each bar in each respective set of harmonically equivalent bars at 1103, for each respective pair of bars in the musical composition a first bar in the pair and a second bar in the pair may be assigned to a same harmonic equivalence category when the comparison at 1102 indicates that the first bar and the second bar are harmonically equivalent to one another. Reciprocally, exemplary rule 1132 specifies that the first bar in the pair and the second bar in the pair may be assigned to different harmonic equivalence categories when the comparison at 1102 indicates that the first bar and the second bar are not harmonically equivalent to one another. Regardless of the precise logic employed, a desired result of act 1103 is a direct mapping between bars and harmonic equivalence categories. Thus, at a very high-level, method 1100 may be summarized as including: i) characterizing each respective bar in the musical composition (at 1101) in such a way (i.e., by determining a respective harmonic fingerprint thereof) that ii) musical/harmonic qualities of the bars may be compared (at 1102) to generate classes or categories of musical/harmonic similarity (i.e., harmonic equivalence categories) with iii) each bar assigned (at 1103) to a corresponding class or category of musical/harmonic similarity. These acts (i.e., acts 1101, 1102, and 1103) of method 1100 achieve a characterization of the harmonic structure of the musical composition, while act 1104 of method 1100 proceeds to incorporate such harmonic structure in a digital representation of the musical composition.
At 1104, the respective harmonic equivalence category of each bar (as assigned at 1103) is encoded in the data structure of a digital representation of the musical composition, e.g., by at least one processor. The data structure of a digital representation (i.e., the “digital data structure”) of the musical composition may, in accordance with the present systems, devices, and methods, comprise a set of bar data objects with each bar of the musical composition encoded in a respective bar data object. In this case, at 1104 the respective harmonic equivalence category of each respective bar may be encoded, by at least one processor, in a respective corresponding bar data object. For example, “harmonic equivalence category” may be included as a bar parameter encoded in each bar data object.
In implementations of method 1100 in which the digital data structure of the musical composition comprises bar data objects, the digital data structure of the musical composition may further comprise segment data objects, where each segment data object encodes a respective sequence of bar data objects. Segment data objects provide a powerful tool for characterizing and, if desired, manipulating (e.g., varying) a musical composition. For a digital data structure that includes segment data objects, the encoding at 1104 of method 1100 may further include encoding, by at least one processor, a respective sequence of harmonic equivalence categories in each segment data object. Having segment data objects that encode sequences of harmonic equivalence categories as well as sequences of bar data objects enables a musical composition to be characterized or represented in terms of either its conventional musical structure (i.e., a sequence of bars) or its harmonic structure (i.e., a sequence of harmonic equivalence categories).
In some implementations of the present systems, devices, and methods, a first digital representation of music that is based on conventional musical bar structure, such as:
Returning again to
An interpretation of the HarmonicStructure[ ] form may be as follows: when any two bars in a musical composition (whether those bars are included in the same segment or in different segments) are assigned to the same harmonic equivalence category integer value, the implementation of method 1100 has determined that those two bars are harmonically equivalent to one another with respect to, e.g., the data structure built from each bar and/or the particular harmonic fingerprint employed in the particular implementation of method 1100. Thus, in the example above, the bars: bar11, bar12, bar21, and bar33 are all found to be harmonically equivalent to one another and assigned to a same harmonic equivalence category (category 1); bar13, bar22, and bar31 are all found to be harmonically equivalent to one another and assigned to a same harmonic equivalence category (category 2); and bar32 and bar34 are found to be harmonically equivalent to one another and assigned to a same harmonic equivalence category (category 3).
As previously described, the nature or type of harmonic fingerprints determined at 1101 and compared at 1102 may vary in different implementations of method 1100, or in different iterations of the same implementation of method 1100. The six illustrative examples of harmonic fingerprints (i.e., note sequence fingerprint 1111, note envelope fingerprint 1112, note count distribution fingerprint 1113, note energy distribution fingerprint 1114, audio fingerprint 1115, and graphical fingerprint 1116) are now described in more detail for exemplary purposes, but a person of skill in the art will appreciate that alternative forms and types of harmonic fingerprints may be employed in alternative implementations of the present systems, devices, and methods.
In implementations of method 1100 for which the harmonic fingerprint determined for each bar at 1101 is a note sequence fingerprint 1111, determining a respective note sequence fingerprint of each bar at 1101 may include sorting all notes in the bar by note start time and, for each note start time, sorting all corresponding notes by note pitch. In other words, for each bar, the notes in the bar may be grouped into those that start substantially simultaneously with one another. The groups themselves may be sorted in increasing order of start time and the notes within each group may be sorted in increasing order of pitch. A convention may be adopted for enharmonic equivalents (e.g., C # and Db), such as “place the sharps before the flats.” In some implementations, the result may then be “flattened” by removing the grouping designations to produce a single list of notes referred to herein as a note sequence fingerprint. In other implementations, the result may not be flattened and instead the group designations of the notes may be preserved. A group of notes that all have the same start time may be referred to as “parallel notes.”
In order to compare note sequence fingerprints at 1102, a difference between a first note sequence fingerprint and a second note sequence fingerprint may be determined. The difference determined may include an absolute difference and/or a relative difference. As an example, determining an absolute difference between a first note sequence fingerprint and a second note sequence fingerprint may include: determining a respective length of each of the first note sequence fingerprint and the second note sequence fingerprint (length being given by, e.g., the number of notes in the note sequence fingerprint), identifying common note sequences between the first note sequence fingerprint and the second note sequence fingerprint, and determining a difference between: a) the maximum of the length of the first note sequence fingerprint and the length of the second note sequence fingerprint), and b) the length of a longest common note sequence between the first note sequence fingerprint and the second note sequence fingerprint. So, if the first note sequence fingerprint comprises 12 notes and the second note sequence fingerprint comprises 15 notes, and the largest note sequence that is common to (i.e., present in) both the first note sequence fingerprint and the second note sequence fingerprint is a particular sequence of 4 notes (e.g., A, C, A, E), then the absolute difference determined at 1102 will be 15-4=11. For absolute difference, a lower score means a higher degree of similarity between the two fingerprints being compared, so bars with a lower absolute distance in terms of note sequence fingerprint are more likely to be assigned to a same harmonic equivalence category at 1103 than are bars with a higher absolute distance in terms of note sequence fingerprint.
As an example of a relative difference between note sequence fingerprints, the absolute difference determined above may be divided by the length of the longer of the two note sequence fingerprints, i.e., 11/15 ˜0.73.
In implementations of method 1100 for which the harmonic fingerprint determined for each bar at 1101 is a note envelope fingerprint 1112, determining a respective note envelope fingerprint of each bar at 1101 may include, for each bar, characterizing a smooth polynomial curve or a piecewise series of straight line segments (i.e., an envelope function) that outlines the extremes of the notes played in the bar. When using a smooth polynomial curve, derivatives of the note gradients may, optionally, be specified at arbitrary places, which can force the interpolating polynomial to fit to the actual notes at the actual times.
As an example, the note envelope fingerprint may a bar may be determined (at 1101) by: i) sorting all notes in the bar by note start time; ii) for each note start time, determining a respective highest pitch note and a respective lowest pitch note; and iii) interpolating between (e.g., connecting between, either with a smooth polynomial curve or piecewise linear segments depending on the implementation) respective lowest pitch notes at successive note start times and between respective highest pitch notes at successive note start times. When sorting notes that correspond to percussion events, either the “pitch order” of the standard midi tunings of standard midi percussion instruments may be used, or other pitch assignments to percussion events (such as “by ear” pitch assignments provided by one or more user(s)) maybe used.
In general, a note envelope fingerprint may have the form:
The construction of the note envelope fingerprint may initially resemble the construction of a note sequence fingerprint, only instead of “flattening” the ordered arrangement of notes as in a note sequence fingerprint, the note envelope fingerprint preserves the note groupings at each note start time and characterizes upper and lower note boundaries at each note start time. Using these upper and lower note boundaries, the noteRange for the bar may be determined and converted into an equivalent noteIntegerRange by associating a unique integer with each note. For example, middle “C” may be mapped to note integer 0, with the adjacent “B” mapped to −1 and the adjacent “C #” mapped to +1). Conventional mathematical functions may be used to interpolate between adjacent (in terms of note start time) note boundaries to obtain the contours of the upper and lower boundaries of the NoteEnvelope for the bar.
Similar to note sequence fingerprints, note envelope fingerprints may adopt a convention for how to treat enharmonic equivalents (e.g., C # and Db), such as “place the sharps before the flats.” Each bar may be “time-zeroed” (e.g., mapped to a start time of 0:00:00 in the musical composition) to facilitate comparisons of the respective note envelope fingerprints at 1102.
In order to compare note envelope fingerprints at 1102, respective graphical representations of respective note envelope fingerprints may be generated and compared in terms of their pixel values. As an example, for a pair of bars in a musical composition: i) a first graphic of a first envelope corresponding to the note envelope fingerprint of a first bar may be generated, defined, or otherwise created, the first graphic having first pixel dimensions (e.g., X pixels wide by Y pixels high); ii) a second graphic of a second envelope corresponding to the note envelope fingerprint of a second bar may be generated, defined, or otherwise created, the second graphic having the same first pixel dimensions (e.g., X pixels wide by Y pixels high) as the first graphic; and iii) a “distance” may be determined between the respective pixel values of the first graphic and the second graphic. The manner or mechanism by which the distance between pixel values of the first graphic and the second graphic is determined may vary in different implementations. Some non-limiting examples of the type of “distance” that may be used to compare the pixel values of a first note envelope fingerprint and a second note envelope fingerprint at 1102 include: Euclidean Distance, Squared Euclidean Distance, Normalized Squared Euclidean Distance, Manhattan Distance, Cosine Distance, Correlation Distance, Mean Euclidean Distance, Mean Squared Euclidean Distance, Root Mean Square Distance, Mean Reciprocal Squared Euclidean Distance, Mutual Information Variation, Normalized Mutual Information Variation, Difference Normalized Entropy, Mean Pattern Intensity, Gradient Correlation, Mean Reciprocal Gradient Distance, Earth Mover Distance, and/or any combination of the foregoing. Furthermore, some implementations may compare (at 1102) respective note envelope fingerprints not in terms of images and distances between pixel values at all, but rather in terms of their constituent sequences of (time, note) pairs.
In implementations of method 1100 for which the harmonic fingerprint determined for each bar at 1101 is a note count distribution fingerprint 1113, determining a respective note count distribution fingerprint of each bar at 1101 may include, for each bar, determining a number of instances of each unique note in the bar. Depending on the implementation, enharmonic notes may be counted as equivalent or unique notes in note count distribution fingerprints. A note count distribution fingerprint may be represented as a vector whose indices correspond to different musical notes and whose values at those indices correspond to the number of occurrences of that note in the given bar (across all tracks in the bar).
In order to compare note count distribution fingerprints at 1102, many different different measures may be used depending on the specific implementation. In some implementations, a conventional Kullback-Leibler Divergence could be used at 1102 to compare respective pairs of note count distribution fingerprints. However, the conventional Kullback-Leibler Divergence provides a result that is dependent on the order in which the two note count distribution fingerprints are treated (in other words, the Kullback-Leibler Divergence between a first note count distribution fingerprint and a second note count distribution fingerprint is not necessarily the same as the Kullback-Leibler Divergence between the second note count distribution fingerprint and the first note count distribution fingerprint). This asymmetry may be undesirable in certain implementations of method 1100 because it may produce less reliable results.
In accordance with the present systems, devices, and methods, a symmetric Kullback-Leibler Divergence may be constructed and used to compare note count distribution fingerprints at 1102. The symmetric Kullback-Leibler Divergence between a first note count distribution fingerprint and a second note count distribution fingerprint may be constructed as follows: i) determine a mean count distribution of each unique note in the musical composition; and ii) determine the sum of: a) the Kullback-Leibler Divergence between the first note count distribution fingerprint and the mean count distribution; and b) the Kullback-Leibler Divergence between the mean count distribution and the second note count distribution fingerprint. This “symmetrization” of the Kullback-Leibler Divergence renders the result independent of the order in which the first note count distribution fingerprint and the second note count distribution fingerprint are treated and can therefore improve the consistency and reliably of the comparisons between note count distribution fingerprints at 1102.
In implementations of method 1100 for which the harmonic fingerprint determined for each bar at 1101 is a note energy distribution fingerprint 1114, determining a respective note energy distribution fingerprint of each bar at 1101 may include determining a respective note energy of each note in the bar and representing the note energy distribution fingerprint as a vector whose indices correspond to different musical notes and whose values at those indices correspond to the energy of that note in the given bar (across all tracks in the bar). The note energy of each note in a bar may be determined as follows: i) determine, for each instance of the note in the bar, a respective product of note duration multiplied by note volume; and ii) determine a sum of the respective products. The same at ii) becomes the note energy of the corresponding note (for the given bar). Because the units of loudness (or volume) multiplied by time do not technically correspond to conventional units of energy, a person of skill in the art will appreciate that the term “energy” is used loosely in this context to reflect the prominence or “energetic character” of notes in a bar rather than literal energy.
Note count distribution fingerprints and note energy distribution fingerprints are similar in form in that they may each be represented as vectors whose indices correspond to different notes in a bar. Therefore, note energy distribution fingerprints may be compared at 1102 using approaches that are similar to the approaches that may be employed to compare note count distribution fingerprints. For example, the symmetric Kullback-Leibler Divergence described above may be adapted for use in comparing note energy distribution fingerprints. The symmetric Kullback-Leibler Divergence between a first note energy distribution fingerprint and a second note energy distribution fingerprint may be constructed as follows: i) determine a mean note energy distribution that includes the respective mean note energy of each unique note in the musical composition; and ii) determine the sum of: a) the Kullback-Leibler Divergence between the first note energy distribution fingerprint and the mean note energy distribution; and b) the Kullback-Leibler Divergence between the mean note energy distribution and the second note energy distribution fingerprint.
In implementations of method 1100 for which the harmonic fingerprint determined for each bar at 1101 is an audio fingerprint 1115, determining a respective audio fingerprint of each bar at 1101 may include, for each bar in the musical composition, encoding the bar in a respective audio object (e.g., a respective digital audio object or a respective analog audio object). In other words, each bar data object in a .hum file representation may be mapped into a respective audio object that “sonifies” the music in the bar. The audio object may have a digital audio file format such as .wav, .mp3, or the like.
In order to compare audio fingerprints at 1102, a distance function that assesses the distance between two audio signals may be used. Many different such distance functions are available, any one of which may be used (alone or in combination with others) depending on the specific implementation. Examples of suitable distance functions that may be used to compare audio fingerprints at 1102 include, without limitation: i) distance functions based on the respective power spectra of two (or more) audio fingerprints, such as: Spectral Euclidean Distance, Spectral Itakura Saito Distance, Spectral Magnitude Phase Distortion Distance, Spectral RMS Log Distance, Spectral First Order Differential Distance, Spectral Second Order Differential Distance, and/or Cepstral Distance; and/or ii) distance functions based on values extracted from samples of the audio signals, such as: Euclidean Distance, Squared Euclidean Distance, Normalized Squared Euclidean Distance, Manhattan Distance, Cosine Distance, Correlation Distance, and/or or Root Mean Square Distance.
As an example, the Spectral Euclidean Distance between a first audio fingerprint and a second audio fingerprint (as an implementation of act 1102 of method 1100) maybe determined as follows: i) a first power spectrum of a first audio object corresponding to a first audio fingerprint is determined; ii) a second power spectrum of a second audio object corresponding to a second audio fingerprint is determined; and iii) a Euclidean distance between the first power spectrum and the second power spectrum is determined.
In implementations of method 1100 for which the harmonic fingerprint determined for each bar at 1101 is a graphical fingerprint 1116, determining a respective graphical fingerprint of each bar at 1101 may include, for each bar in the musical composition, encoding the bar in a respective graphical image that visually represents (or “visualizes”) musical information contained within the bar. In order to facilitate comparisons between graphical fingerprints at 1102, each respective graphical image may advantageously have the same number of pixels.
A bar may be encoded in a graphical image in a variety of different ways, all of which fall within the scope of the present systems, devices, and methods. As a specific example, encoding a bar in a graphical image may include constructing a two-dimensional graph in which the horizontal axis represents time (i.e, spanning the time intervals over which the various notes in the bar are played) and the vertical axis is divided into a topmost portion and a bottommost portion. The topmost portion of the vertical axis may correspond to (i.e., map different vertical values to) musical notes (e.g., spanning the range of all human-audible pitches) and the bottommost portion of the vertical axis may correspond to (i.e., map different vertical values to) different rhythm and/or percussion events (e.g., snare drum, kick drum, symbol, and the like). A respective line, curve, or other feature may be drawn on such a graph for each respective instrument that is employed in the bar, with each respective instrument represented by a different respective color. Advantageously, the respective graphical image of each bar may be time-zeroed so that, effectively, the respective graphical images of all of the bars in the musical composition have a common starting time. The respective graphical image of each bar may also be rasterized so that, effectively, the respective graphical images of all of the bars in the musical composition have the same dimension in terms of the number of pixels.
In order to compare graphical fingerprints at 1102, a distance may be determined between a first graphical image corresponding to a first graph fingerprint of a first bar and a second graphical image corresponding to a second graph fingerprint of a second bar. As is the case for many of the other types of harmonic fingerprints described, the nature or form of distance that may be determined to compare graphical fingerprints may vary in different implementations. The Earth Mover's Distance is an example of a suitable distance measure that may be used to compare graphical fingerprints in accordance with the present systems, devices, and methods. In some implementations, the Earth Mover's Distance may quantify the distance between two probability distributions over a two-dimensional plane, so if the distributions are interpreted as two different ways of piling up a certain amount of dirt over the surface, the Earth Mover's Distance may measure the minimum cost of turning one pile into the other pile, where the cost is assumed to be amount of dirt moved times the distance by which it is moved. Alternative distance functions that may be used to compare two or more graphical fingerprints at 1102 include, without limitation: Euclidean Distance, Squared Euclidean Distance, Normalized Squared Euclidean Distance, Manhattan Distance, Cosine Distance, Correlation Distance, Mean Euclidean Distance, Mean Squared Euclidean Distance, Root Mean Square Distance, Mean Reciprocal Squared Euclidean Distance, Mutual Information Variation, Normalized Mutual Information Variation, Difference Normalized Entropy, Mean Pattern Intensity, Gradient Correlation, and/or Mean Reciprocal Gradient Distance.
The various implementations described herein often make reference to “computer-based,” “computer-implemented,” “at least one processor,” “a non-transitory processor-readable storage medium,” and similar computer-oriented terms. A person of skill in the art will appreciate that the present systems, devices, and methods may be implemented using or in association with a wide range of different hardware configurations, including localized hardware configurations (e.g., a desktop computer, laptop, smartphone, or similar) and/or distributed hardware configurations that employ hardware resources located remotely relative to one another and communicatively coupled through a network, such as a cellular network or the internet. For the purpose of illustration, exemplary computer systems suitable for implementing the present systems, devices, and methods are provided in
Processor-based computer system 1200 includes at least one processor 1201, a non-transitory processor-readable storage medium or “system memory” 1202, and a system bus 1210 that communicatively couples various system components including the system memory 1202 to the processor(s) 1201. Processor-based computer system 1201 is at times referred to in the singular herein, but this is not intended to limit the implementations to a single system, since in certain implementations there will be more than one system or other networked computing device(s) involved. Non-limiting examples of commercially available processors include, but are not limited to: Core microprocessors from Intel Corporation, U.S.A., PowerPC microprocessor from IBM, ARM processors from a variety of manufacturers, Sparc microprocessors from Sun Microsystems, Inc., PA-RISC series microprocessors from Hewlett-Packard Company, and 68xxx series microprocessors from Motorola Corporation.
The processor(s) 1201 of processor-based computer system 1200 may be any logic processing unit, such as one or more central processing units (CPUs), microprocessors, digital signal processors (DSPs), application-specific integrated circuits (ASICs), field programmable gate arrays (FPGAs), and/or the like. Unless described otherwise, the construction and operation of the various blocks shown in
The system bus 1210 in the processor-based computer system 1200 may employ any known bus structures or architectures, including a memory bus with memory controller, a peripheral bus, and/or a local bus. The system memory 1202 includes read-only memory (“ROM”) 1221 and random access memory (“RAM”) 1222. A basic input/output system (“BIOS”) 1223, which may or may not form part of the ROM 1221, may contain basic routines that help transfer information between elements within processor-based computer system 1200, such as during start-up. Some implementations may employ separate buses for data, instructions and power.
Processor-based computer system 1200 (e.g., system memory 1202 thereof) may include one or more solid state memories, for instance, a Flash memory or solid state drive (SSD), which provides nonvolatile storage of processor-executable instructions, data structures, program modules and other data for processor-based computer system 1200. Although not illustrated in
Program modules in processor-based computer system 1200 may be stored in system memory 1202, such as an operating system 1224, one or more application programs 1225, program data 1226, other programs or modules 1227, and drivers 1228.
The system memory 1202 in processor-based computer system 1200 may also include one or more communications program(s) 1229, for example, a server and/or a Web client or browser for permitting processor-based computer system 1200 to access and exchange data with other systems such as user computing systems, Web sites on the Internet, corporate intranets, or other networks as described below. The communications program(s) 1229 in the depicted implementation may be markup language based, such as Hypertext Markup Language (HTML), Extensible Markup Language (XML) or Wireless Markup Language (WML), and may operate with markup languages that use syntactically delimited characters added to the data of a document to represent the structure of the document. A number of servers and/or Web clients or browsers are commercially available such as those from Google (Chrome), Mozilla (Firefox), Apple (Safari), and Microsoft (Internet Explorer).
While shown in
Processor-based computer system 1200 may include one or more interface(s) to enable and provide interactions with a user, peripheral device(s), and/or one or more additional processor-based computer system(s). As an example, processor-based computer system 1210 includes interface 1230 to enable and provide interactions with a user of processor-based computer system 1200. A user of processor-based computer system 1200 may enter commands, instructions, data, and/or information via, for example, input devices such as computer mouse 1231 and keyboard 1232. Other input devices may include a microphone, joystick, touch screen, game pad, tablet, scanner, biometric scanning device, wearable input device, and the like. These and other input devices (i.e., “I/O devices”) are communicatively coupled to processor(s) 1201 through interface 1230, which may include one or more universal serial bus (“USB”) interface(s) that communicatively couples user input to the system bus 1210, although other interfaces such as a parallel port, a game port or a wireless interface or a serial port may be used. A user of processor-based computer system 1200 may also receive information output by processor-based computer system 1200 through interface 1230, such as visual information displayed by a display 1233 and/or audio information output by one or more speaker(s) 1234. Monitor 1233 may, in some implementations, include a touch screen.
As another example of an interface, processor-based computer system 1200 includes network interface 1240 to enable processor-based computer system 1200 to operate in a networked environment using one or more of the logical connections to communicate with one or more remote computers, servers and/or devices (collectively, the “Cloud” 1241) via one or more communications channels. These logical connections may facilitate any known method of permitting computers to communicate, such as through one or more LANs and/or WANs, such as the Internet, and/or cellular communications networks. Such networking environments are well known in wired and wireless enterprise-wide computer networks, intranets, extranets, the Internet, and other types of communication networks including telecommunications networks, cellular networks, paging networks, and other mobile networks.
When used in a networking environment, network interface 1240 may include one or more wired or wireless communications interfaces, such as network interface controllers, cellular radios, WI-FI radios, and/or Bluetooth radios for establishing communications with the Cloud 1241, for instance, the Internet or a cellular network.
In a networked environment, program modules, application programs or data, or portions thereof, can be stored in a server computing system (not shown). Those skilled in the relevant art will recognize that the network connections shown in
For convenience, processor(s) 1201, system memory 1202, interface 1230, and network interface 1240 are illustrated as communicatively coupled to each other via the system bus 1210, thereby providing connectivity between the above-described components. In alternative implementations, the above-described components may be communicatively coupled in a different manner than illustrated in
In accordance with the present systems, devices, and methods, processor-based computer system 1200 may be used to implement any or all of methods 100, 200, 600, 700, 900, 1000, and/or 1100 described herein and/or encode and manipulate any or all of the data structures described herein, including without limitation data structures 300, 400, and 500. Where the descriptions of methods 100, 200, 600, 700, 900, 1000, and 1100 make reference to an act being performed by at least one processor, such act may be performed by processor(s) 1201 of computer system 1200. Where the descriptions of methods 100, 200, 600, 700, 900, 1000, and 1100 make reference an act being performed by, performed on, or otherwise involving a non-transitory processor-readable storage medium, such act may be performed by, performed on, or otherwise involve system memory 1202 of computer system 1200. Similarly, data structures 300, 400, and 500 represent illustrative examples of data structures that may be encoded by processor(s) 1201 and stored in system memory 1202 in the present systems, devices, and methods.
Computer system 1200 is an illustrative example of a system for encoding a musical composition in a digital audio format and/or for encoding a harmonic structure of a musical composition in a digital data structure, the system comprising at least one processor 1201, at least one non-transitory processor-readable storage medium 1202 communicatively coupled to the at least one processor 1201 (e.g, by system bus 1210), and the various other hardware and software components illustrated in
Using method 100 as an example, the processor-executable instructions and/or data of computer program product 1250 stored in system memory 1202 may, when executed by processor(s) 1201, cause processor(s) 1201 to encode respective components of a first instance of the musical composition in respective data objects per act 101 of method 100, encode the data objects in a data file 1251 per act 102 of method 100, and store data file 1251 in system memory 1202 per act 103 of method 100. As illustrated in
Using method 1100 as another example, the processor-executable instructions and/or data of computer program product 1250 stored in system memory 1202 may, when executed by processor(s) 1201, cause processor(s) 1201 to determine a respective harmonic fingerprint of each bar in the musical composition per act 1101 of method 1100, compare the respective harmonic fingerprints of respective bars in the musical composition to identify sets of harmonically equivalent bars per act 1102 of method 1100; assign a respective harmonic equivalence category to each bar in each respective set of harmonically equivalent bars per act 1103 of method 1100; and encode the respective harmonic equivalence category of each bar in the digital data structure 1252 per act 1104 of method 1100. Thus, digital data structure 1252 in
Throughout this specification and the appended claims, the term “computer program product” is used to refer to a package, combination, or collection of software comprising processor-executable instructions and/or data that may be accessed by (e.g., through a network such as cloud 1241) or distributed to and installed on (e.g., stored in a local non-transitory processor-readable storage medium such as system memory 1202) a computer system (e.g., computer system 1200) in order to enable certain functionality (e.g., application(s), program(s), and/or module(s)) to be executed, performed, or carried out by the computer system.
Turning again to
Similarly, the data structure 1252 that encodes the harmonic structure of the musical composition encoded in .hum file 1251 when processor(s) 1201 of system 1200 execute(s) additional or other processor-executable instructions and/or data of computer program product 1250 stored in system memory 1202 may, in accordance with the present systems, devices, and methods, encode the respective harmonic equivalence category of each bar in a respective corresponding bar data object from .hum file 1251.
Computer program product 1250, and therefore computer system 1200 when computer program product 1250 is either accessed through network 1240 or stored in system memory 1202, may also be configured to generate one or more musical variations of the first instance of the musical composition in accordance with the present systems, devices, and methods. For example, computer program product 1250 may include processor-executable instructions and/or data that, when executed by processor(s) 1201, cause processor(s) 1201 to generate a variation of the first instance of the musical composition (e.g., per act 701 of method 700 and/or act 1002 of method 1000). The variation of the first instance of the musical composition may be encoded in .hum file 1251 and stored in system memory 1202, and may or may not replace the first instance of the musical composition in .hum file 1251 depending on the specific implementation. Generally, the variation may include a second sequence of notes that is different from the first sequence of notes such that .hum file 1251 encodes a second set of note data objects (corresponding to the second sequence of notes) that is different from the first set of note data objects (corresponding to the first sequence of notes); however, in accordance with the present systems, devices, and methods, the variation encoded in .hum file 1251 may preserve the timing relationship between the beats and the notes of the first instance of the musical composition in the beats and the notes of the variation.
The various implementations described herein improve the functioning of computer systems for the specific practical application of algorithmic composition of music. Encoding various components of a musical composition in separate discrete data objects introduces a modularity that enables exceptional (relative to other approaches in the art) algorithmic control to set, vary, manipulate, and rearrange the various components of the composition. This modularity may be taken advantage of in algorithmic composition to produce new and improved software and applications for computer-based music creation/variation that output more sophisticated and enjoyable musical results. But modularity is just one example of how the present systems, devices, and methods improve the functioning of computer systems for automatic music generation. There are many other ways in which the present systems, devices, and methods advantageously improve the use of computers for generating music, including without limitation: enabling various component parameters (e.g., note timing and beat timing) to be defined and manipulated independently of one another; enabling certain parametric relationships (e.g., timing relationships) to be preserved across different components and/or parameters while other parametric relationships (e.g., note sequences) are varied; enabling musical compositions to be defined and manipulated using the higher-level concepts of musical theory with which most composers and musicians are already familiar; and enabling the harmonic structure of a musical composition to be characterized and taken into account when composing variations of the musical composition. These and other advantages are all achieved within a data file (e.g., the .hum file and/or the digital data structure that encodes the harmonic structure of the .hum file) that is well-suited for communication between computer systems and electronic instruments.
Throughout this specification and the appended claims, reference is often made to musical compositions being “automatically” generated/composed by computer-based algorithms, software, and/or artificial intelligence (AI) techniques. A person of skill in the art will appreciate that a wide range of algorithms and techniques may be employed in computer-generated music, including without limitation: algorithms based on mathematical models (e.g., stochastic processes), algorithms that characterize music as a language with a distinct grammar set and construct compositions within the corresponding grammar rules, algorithms that employ translational models to map a collection of non-musical data into a musical composition, evolutionary methods of musical composition based on genetic algorithms, and/or machine learning-based (or AI-based) algorithms that analyze prior compositions to extract patterns and rules and then apply those patterns and rules in new compositions. These and other algorithms may be advantageously adapted to exploit the features and techniques enabled by the digital representations of music described herein.
Throughout this specification and the appended claims the term “communicative” as in “communicative coupling” and in variants such as “communicatively coupled,” is generally used to refer to any engineered arrangement for transferring and/or exchanging information. For example, a communicative coupling may be achieved through a variety of different media and/or forms of communicative pathways, including without limitation: electrically conductive pathways (e.g., electrically conductive wires, electrically conductive traces), magnetic pathways (e.g., magnetic media), wireless signal transfer (e.g., radio frequency antennae), and/or optical pathways (e.g., optical fiber). Exemplary communicative couplings include, but are not limited to: electrical couplings, magnetic couplings, radio frequency couplings, and/or optical couplings.
Throughout this specification and the appended claims, infinitive verb forms are often used. Examples include, without limitation: “to encode,” “to provide,” “to store,” and the like. Unless the specific context requires otherwise, such infinitive verb forms are used in an open, inclusive sense, that is as “to, at least, encode,” “to, at least, provide,” “to, at least, store,” and so on.
This specification, including the drawings and the abstract, is not intended to be an exhaustive or limiting description of all implementations and embodiments of the present systems, devices, and methods. A person of skill in the art will appreciate that the various descriptions and drawings provided may be modified without departing from the spirit and scope of the disclosure. In particular, the teachings herein are not intended to be limited by or to the illustrative examples of computer systems and computing environments provided.
This specification provides various implementations and embodiments in the form of block diagrams, schematics, flowcharts, and examples. A person skilled in the art will understand that any function and/or operation within such block diagrams, schematics, flowcharts, or examples can be implemented, individually and/or collectively, by a wide range of hardware, software, and/or firmware. For example, the various embodiments disclosed herein, in whole or in part, can be equivalently implemented in one or more: application-specific integrated circuit(s) (i.e., ASICs); standard integrated circuit(s); computer program(s) executed by any number of computers (e.g., program(s) running on any number of computer systems); program(s) executed by any number of controllers (e.g., microcontrollers); and/or program(s) executed by any number of processors (e.g., microprocessors, central processing units, graphical processing units), as well as in firmware, and in any combination of the foregoing.
Throughout this specification and the appended claims, a “memory” or “storage medium” is a processor-readable medium that is an electronic, magnetic, optical, electromagnetic, infrared, semiconductor, or other physical device or means that contains or stores processor data, data objects, logic, instructions, and/or programs. When data, data objects, logic, instructions, and/or programs are implemented as software and stored in a memory or storage medium, such can be stored in any suitable processor-readable medium for use by any suitable processor-related instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the data, data objects, logic, instructions, and/or programs from the memory or storage medium and perform various acts or manipulations (i.e., processing steps) thereon and/or in response thereto. Thus, a “non-transitory processor-readable storage medium” can be any element that stores the data, data objects, logic, instructions, and/or programs for use by or in connection with the instruction execution system, apparatus, and/or device. As specific non-limiting examples, the processor-readable medium can be: a portable computer diskette (magnetic, compact flash card, secure digital, or the like), a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM, EEPROM, or Flash memory), a portable compact disc read-only memory (CDROM), digital tape, and/or any other non-transitory medium.
The claims of the disclosure are below. This disclosure is intended to support, enable, and illustrate the claims but is not intended to limit the scope of the claims to any specific implementations or embodiments. In general, the claims should be construed to include all possible implementations and embodiments along with the full scope of equivalents to which such claims are entitled.
Number | Name | Date | Kind |
---|---|---|---|
6798886 | Smith | Sep 2004 | B1 |
20030183065 | Leach | Oct 2003 | A1 |
20210158833 | Cooke | May 2021 | A1 |
Number | Date | Country | |
---|---|---|---|
20200402488 A1 | Dec 2020 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16448130 | Jun 2019 | US |
Child | 16775254 | US |