The present invention relates to methods and systems for displaying music for educational or entertainment.
The present disclosure enables a system for display of a visualization of sound energy in an audible input. The system may be implemented in a hardware and software combination, the comprised of an audio processing filter that may be implemented in hardware or software and that is configured to filter time-domain assessed frequency characteristics for a plurality of distinct note-frequencies in the sound energy, the time-domain assessed frequency characteristics including magnitude and duration at each distinct frequency assessed, the audio processing filter coupled to a processor configured to transform the time-domain assessed frequency characteristics for the plurality of distinct frequencies into three-dimensional coordinates, a graphics processing unit to generate graphics processing data to display a visualization of the time-domain assessed frequency characteristics wherein the visualization includes a first surface comprised of a first set of coordinates transformed from a first set of the plurality of distinct frequencies and a second surface comprised of a second set of coordinates transformed from a second set of the plurality of distinct frequencies.
Aspects may be implemented in a system wherein the first and or second set of coordinates are positioned on or form a plane or cone.
Additional aspects may be implemented in a system wherein the plurality of distinct frequencies correspond to musical notes and the first set of the plurality of distinct frequencies comprises a first octave and the second set of the plurality of distinct frequencies comprises a second octave, and the musical notes may be Western 12-note scale notes, a Middle Eastern scale tones, an Indian scale tones, and Chinese Scale tones, a custom scale, or a combination thereof.
Additional aspects of the disclosure include transformation of sound energy to three-dimensional coordinates create a matrix or visualization of the sound energy wherein each set of three-dimensional coordinates includes a magnitude coordinate that is proportional to the magnitude at each distinct frequency assessed, a radial coordinate that is proportional to the time elapsed between successive coordinates, a first angle coordinate common to the first set of coordinates and a second angle coordinate common to the second set of coordinates.
The present disclosure also enables a method of producing a three-dimensional visualization of an audio input. In the method, and for each of a plurality of distinct frequencies, the method includes sampling the audio input to determine a relative magnitude characteristic proportional to the sound energy present at the distinct frequency sampled and sequence of occurrence characteristic, transforming each sample of relative magnitude characteristic and sequence of occurrence characteristic into a set of coordinates for which each of the plurality of distinct frequencies includes a unique polar angle coordinate and azimuth angle coordinate combination, converting each set of coordinates into graphics display data, and producing a graphical three-dimensional visualization of the graphics display data.
Aspects of the method disclosed also enables including in each set of coordinates also includes a relative magnitude coordinate that corresponds to the magnitude of sound energy at each distinct frequency assessed by the system, or each set of coordinates may alternately comprise a magnitude coordinate proportional to the magnitude of the sound energy in the audio input at each distinct frequency, and a radial coordinate proportional to the occurrence of the sample relative to other samples.
Aspects of the method enable visualizations wherein a first range of the plurality of distinct frequencies have a common first polar angle coordinate and a second range of the plurality of frequencies have a common second polar angle coordinate, and the graphical three-dimensional visualization comprises a first surface composed of the coordinates from the first range of the plurality of distinct frequencies and a second surface composed of the coordinates from the second range of distinct frequencies. Additional aspects include assigning a unique azimuthal angle coordinate for each distinct frequency within the first range of the distinct frequencies, and the same azimuth angle coordinate to an integer multiple of the distinct frequency in the second range of the distinct frequencies.
Additional aspects enabled by the disclosure include graphical user interface control of the system to create a visual display of the audio energy in an audio source, the system configured to transform the audio energy at a plurality of distinct frequencies into coordinates for a three-dimensional display of the audio energy. The interface may include a control to define a time interval between successive coordinates for the three-dimensional display, a control for selecting a music scale selected from Western 12-note scale notes, a Middle Eastern scale tones, an Indian scale tones, and Chinese Scale tones, a custom scale, or a combination thereof.
Numerous advantages and features of the present invention will become readily apparent from the following detailed description of the invention and the embodiments thereof, from the claims and from the accompanying drawings.
The objects, features and advantages of the present invention will be more readily appreciated upon reference to the following disclosure when considered in conjunction with the accompanying drawings, wherein reference numerals are used to identify the components in the various views.
The figures illustrate an embodiment of a three-dimensional visualization system for audible inputs or audio data incorporating aspects of the invention. The features of the disclosed embodiment(s) should not be construed as limiting any aspect of the invention, however. The aspects of the invention disclosed may be scaled or modified for all types of analog or digital inputs and it is contemplated that adaptions of the teachings of this disclosure enable visualization of captured analog audio, streamed, or stored digital audio, interpreted sheet music, or of the output from any instrument or input device capable of creating MIDI output data. The embodiments disclosed comprise hardware and software systems for visualization of audio content, which may be live audio, recorded audio, audio encoded in any digital protocol or format, and other audio or musical formats capable of being interpreted by a software or hardware translator or interpreter. In preferred embodiments, the systems enable visualization of frequencies within the audible spectrum so that the systems facilitate visual analysis of the frequencies for study or entertainment purposes.
An embodiment of a system may comprise an audio processing filter 20 configured to filter time-domain assessed audible characteristics for a plurality of distinct frequencies 25, and a process to transform the time-domain assessed audible characteristics for the plurality of distinct frequencies 25 into graphics processing data 27 that may be rendered for display or visualization of the assessed audible characteristics. As one example, the assessed audible characteristics may generate graphics processing data 27 to generate a three-dimensional visualization that comprises a radial geometric image 304. One manner of capturing and creating assessed audible characteristics comprises sampling the audio and processing the representative data by audio processing filter 20 that comprises a plurality of narrow bandpass filters 222, each tuned to filter the specific frequency that is to be assessed.
In the embodiment illustrated in
In the illustrated conceptual block diagram in
In yet another preferred embodiment, the two-dimensional visualization system design is adapted according to the description and enables a system for three-dimensional visualization of time domain based audible characteristics for a plurality of distinct frequencies by mapping or transforming the characteristics to spherical coordinates that may be converted to graphics processing data 27 for display. In one such three dimensional visualization embodiment, an audio processing filter 20 (see
Moreover, as is evident from the Table 1 frequency values in a Western scale, the higher octaves of each note are integer multiples of the lower octave(s). In certain preferred embodiments, the system will transform the each note or distinct frequency to have the same azimuth angle coordinate as the other octaves of the same note, but with different polar angle coordinates. As one example, an embodiment according to the above preference will transform a “D” note (i.e., frequency=36.71 Hz) of the third octave to have a first azimuth angle coordinate (e.g., φ1) and a first polar angle coordinate (e.g., θ1), and will transform a “D” note of the fourth octave (i.e., frequency=73.42 Hz) to have the same first azimuth angle coordinate (e.g., φ1) but a different or second polar angle coordinate (e.g., θ2). 1 https://piano-ology.com/scales-middle-eastern-theory-ear-training/#:˜:text=The%20scale%20structure%20of%20the,keys%20are%20letter%20names.
In one example of a hardware-software system implementing the present disclosure, the system may comprise a host system and at least one Digital Signal Processor (DSP) subsystem that incorporates at least one digital to analog converter, DSP program memory, DSP data memory, a sampling clock, anti-aliasing and anti-imaging filters and a DSP computing engine controllable by DSP algorithm software and accessible via a host interface to allow the DSP subsystem to receive commands from and communicate with the host system in which the DSP subsystem is embedded. The host system may incorporate a graphical user interface (GUI) as described further below that enables the programming of the DSP algorithm software via the host interface by loading the audio processing filters 202 and look-up tables 230 from either the host system memory or DSP program memory for subsequent processing of sound energy inputs by the DSP computing engine. A user may direct the host system resources and DSP subsystem resources via a GUI as described further below to sample sound energy at a system audio input and process the time-domain assessed audible characteristics for each octave 10 and for each of the plurality of distinct frequencies 206 indicated by the user or by preset music scales preprogrammed in the DSP algorithm software. The time-domain assessed audible characteristics transformed by the look-up tables 230 and then written to system host memory and that may be read out as graphics processing data 27 for display on a monitor or display and a three-dimensional visualization of the sound energy.
Additional examples of three-dimensional visualization created by the system described are illustrated in
In yet another preferred embodiment, the sound energy in an audio or audible input comprised of plurality of ranges of frequencies (i.e., an audio clip or stream of music containing multiple octaves of notes) may be transformed, such as by mathematical calculation or by mapping, to three-dimensional (e.g., spherical) coordinates from a periodic function that swings, oscillates, transitions, or ranges between a first and second polar angles around “r” as a frequency-range-axis or octave-axis and the coordinate position is determined by the intersection of the periodic function and the time interval “r”, referred to herein as a “range coordinate visualization or matrix.” An illustration of the concept and a flow diagram for building a matrix or three-dimensional visualization is illustrated in
In another embodiment option, the system may be used to create spherical or other three-dimensional coordinates to visualize a Musical Instrument Digital Interface (MIDI) stream or file comprised of MIDI messages. A system block diagram for decoding and visualizing MIDI messages is illustrated in
The host hardware-software system may parse the MIDI message data in host system memory 522 to read timing information, note numbers, and volume information, and other MIDI message data and process the MIDI message data according to a MIDI Look up Table 530 to generate three dimensional coordinates according to the description herein. An intensity may be correlated to and proportional to the velocity information associated with each MIDI message. Accordingly, coordinates generated by a software system according to the description will include a radial coordinate “r” (distance to origin) that is related to or proportional to the MIDI timing information encoded in the MIDI message, an amplitude or loudness coordinate “A” that is proportional to the velocity information included in the MIDI message, an azimuthal coordinate proportional to the note number in the MIDI message, and a polar coordinate dependent on the octave of the note expressed in the MIDI message.
A user of the GUI may select a point coordinate visualization 280 of MIDI messages. As before with sound energy, it is preferred that MIDI messages of different octaves generate coordinates that display alternate surfaces. In this case, the software system generates coordinates for MIDI notes in a first range of frequencies (i.e., that correspond to a first octave) generates GPU data that correspond to a first surface, and MIDI notes in in a second range of frequencies (i.e., a second or different octave) generate GPU data that correspond to a second surface. Further, and as before, transformed spherical coordinates of the MIDI notes in each octave of the scale may be mapped with distinct angles relative to the azimuthal axis but with the same angle with respect to the polar axis such that the MIDI notes in each octave are included either in a plane that includes the azimuthal axis or on the surface of a cone originating from the intersection of the polar and azimuthal axis. In the event that a user selects a range coordinate visualization 282, the MIDI notes in a first range of frequencies that correspond to a first octave may have a first transformed polar angle coordinate “θn-2” between a first polar angle range limit, “θa”, and a second polar angle range limit, “θb”, and the transformed spherical coordinates for a second range of frequencies that correspond to a second octave may have a second transformed polar angle “θn-1” between a third polar angle range limit, “θb”, and a fourth polar angle range limit, “θc”. Thus, for each MIDI message, whether sampled or read, there is a time-domain assessed frequency characteristic that may be transformed, such as by mathematical calculation or mapping, to a spherical coordinates, and more preferably, a plurality of assessed frequency characteristics may correspond to tones in a recognized music scale and the MIDI notes in each octave of the scale may be mapped with distinct angles relative to the azimuthal axis but with the same angle with respect to the polar axis such that the MIDI notes in each octave are included either in a plane that includes the azimuthal axis or on the surface of a cone originating from the intersection of the polar and azimuthal axis. Finally, despite that two octaves are described above, it should be obvious that the system hardware and software may be scaled according to the description to enable the visualization of as many as nine octaves of notes or frequencies.
The host software and hardware system may also be scaled to include as many MIDI channels as needed for the complexity of the visualization desired. In such an option, the host hardware-software system to create three-dimensional visualizations includes dedicated software algorithms that enable deconstruction of received MIDI channel(s) messages that is written to host system memory that is subsequently read by the application software to scrape the information to generate coordinates according to the description in this specification. The MIDI converter or interpreter may transform the MIDI message by reading the timestamp of the MIDI Track data (byte 1 of the music data) for each MIDI event to create or transform the radial coordinate for each MIDI event in the MIDI message, the Status byte (byte 2) and bytes 3 may be read to determine the MIDI data note-pitch, and byte 4 to determine the volume of the note.
While various embodiments have been described above, it should be understood that they have been presented by way of example only, and not limitation. Thus, the breadth and scope of a preferred embodiment should not be limited by any of the above-described exemplary embodiments but should be defined only in accordance with the following claims and their equivalents.
| Number | Name | Date | Kind |
|---|---|---|---|
| 5075880 | Moses | Dec 1991 | A |
| 5092216 | Wadhams | Mar 1992 | A |
| 5225618 | Wadhams | Jul 1993 | A |
| 5467401 | Nagamitsu | Nov 1995 | A |
| 5497435 | Berger | Mar 1996 | A |
| 5901708 | Chang | May 1999 | A |
| 5902948 | Haruyama | May 1999 | A |
| 6031172 | Papadopoulos | Feb 2000 | A |
| 6323408 | Liu | Nov 2001 | B1 |
| 6330486 | Padula | Dec 2001 | B1 |
| 6411289 | Zimmerman | Jun 2002 | B1 |
| 6429863 | LoPiccolo | Aug 2002 | B1 |
| 6686529 | Kim | Feb 2004 | B2 |
| 6831220 | Varme | Dec 2004 | B2 |
| 6995310 | Knapp | Feb 2006 | B1 |
| 7820900 | Lemons | Oct 2010 | B2 |
| 7928306 | Lemons | Apr 2011 | B2 |
| 7932454 | Lemons | Apr 2011 | B2 |
| 7932455 | Lemons | Apr 2011 | B2 |
| 7947888 | Lemons | May 2011 | B2 |
| 7994409 | Lemons | Aug 2011 | B2 |
| 8018459 | Lemons | Sep 2011 | B2 |
| 8073701 | Lemons | Dec 2011 | B2 |
| 8088985 | Bowser | Jan 2012 | B1 |
| 10755683 | Baltazor | Aug 2020 | B1 |
| 10775941 | Heath | Sep 2020 | B2 |
| 11756515 | Pereverzev | Sep 2023 | B1 |
| 20030222977 | Yoshino | Dec 2003 | A1 |
| 20040237756 | Forbes | Dec 2004 | A1 |
| 20050190199 | Brown | Sep 2005 | A1 |
| 20060107826 | Knapp | May 2006 | A1 |
| 20060274144 | Landschaft | Dec 2006 | A1 |
| 20070009115 | Reining | Jan 2007 | A1 |
| 20070143268 | Chosokabe | Jun 2007 | A1 |
| 20070243915 | Egozy | Oct 2007 | A1 |
| 20080002844 | Chin | Jan 2008 | A1 |
| 20080022842 | Lemons | Jan 2008 | A1 |
| 20080143751 | Chosokabe | Jun 2008 | A1 |
| 20080245211 | Lemons | Oct 2008 | A1 |
| 20080245212 | Lemons | Oct 2008 | A1 |
| 20080259083 | Lemons | Oct 2008 | A1 |
| 20080264238 | Lemons | Oct 2008 | A1 |
| 20080264241 | Lemons | Oct 2008 | A1 |
| 20080270904 | Lemons | Oct 2008 | A1 |
| 20080274443 | Lemons | Nov 2008 | A1 |
| 20080314228 | Dreyfuss | Dec 2008 | A1 |
| 20090164225 | Cho | Jun 2009 | A1 |
| 20090223349 | Lemons | Sep 2009 | A1 |
| 20100110072 | Nakayama | May 2010 | A1 |
| 20100137049 | Epstein | Jun 2010 | A1 |
| 20110096073 | Adhikari | Apr 2011 | A1 |
| 20110187718 | Diara | Aug 2011 | A1 |
| 20110191674 | Rawley | Aug 2011 | A1 |
| 20110273438 | Liu | Nov 2011 | A1 |
| 20120097012 | Kurihara | Apr 2012 | A1 |
| 20130000463 | Grover | Jan 2013 | A1 |
| 20130148812 | Corteel | Jun 2013 | A1 |
| 20130214485 | Dupree | Aug 2013 | A1 |
| 20140310011 | Biswas | Oct 2014 | A1 |
| 20150071447 | Sen | Mar 2015 | A1 |
| 20150332690 | Kim | Nov 2015 | A1 |
| 20160042657 | Granger | Feb 2016 | A1 |
| 20160042740 | Jiang | Feb 2016 | A1 |
| 20160203805 | Strachan | Jul 2016 | A1 |
| 20170105081 | Jin | Apr 2017 | A1 |
| 20180047372 | Scallie | Feb 2018 | A1 |
| 20180174559 | Elson | Jun 2018 | A1 |
| 20180197551 | Mcdowell | Jul 2018 | A1 |
| 20180270600 | Boehm | Sep 2018 | A1 |
| 20180359594 | Maeno | Dec 2018 | A1 |
| 20200120326 | Deshpande | Apr 2020 | A1 |
| 20210043172 | Yamagiwa | Feb 2021 | A1 |
| 20210289309 | Herre | Sep 2021 | A1 |
| 20210314724 | Murata | Oct 2021 | A1 |
| 20220029721 | Jackson | Jan 2022 | A1 |
| 20220030374 | Suyama | Jan 2022 | A1 |
| 20220208212 | Liao | Jun 2022 | A1 |
| 20220357417 | Kim | Nov 2022 | A1 |
| 20220377489 | Herre | Nov 2022 | A1 |
| 20230123105 | Iezzi | Apr 2023 | A1 |
| 20230300532 | Spittle | Sep 2023 | A1 |
| 20230421977 | Araujo-Simon | Dec 2023 | A1 |
| Number | Date | Country | |
|---|---|---|---|
| 20240070941 A1 | Feb 2024 | US |