US20200142926A1 - Artificial intelligence methodology to automatically generate interactive play along songs - Google Patents
Artificial intelligence methodology to automatically generate interactive play along songs Download PDFInfo
- Publication number
- US20200142926A1 US20200142926A1 US16/732,021 US201916732021A US2020142926A1 US 20200142926 A1 US20200142926 A1 US 20200142926A1 US 201916732021 A US201916732021 A US 201916732021A US 2020142926 A1 US2020142926 A1 US 2020142926A1
- Authority
- US
- United States
- Prior art keywords
- song
- tempo
- audio file
- notes
- specified
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/36—Accompaniment arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/63—Querying
- G06F16/638—Presentation of query results
- G06F16/639—Presentation of query results using playlists
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
- G10H1/0025—Automatic or semi-automatic music composition, e.g. producing random music, applying rules from music theory or modifying a musical piece
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0033—Recording/reproducing or transmission of music for electrophonic musical instruments
- G10H1/0041—Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
- G10H1/0058—Transmission between separate instruments or between individual components of a musical system
- G10H1/0066—Transmission between separate instruments or between individual components of a musical system using a MIDI interface
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/18—Selecting circuits
- G10H1/20—Selecting circuits for transposition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/031—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
- G10H2210/066—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for pitch analysis as part of wider processing for musical purposes, e.g. transcription, musical performance evaluation; Pitch recognition, e.g. in polyphonic sounds; Estimation or use of missing fundamental
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/031—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
- G10H2210/076—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction of timing, tempo; Beat detection
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/031—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
- G10H2210/081—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for automatic key or tonality recognition, e.g. using musical rules or a knowledge base
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/395—Special musical scales, i.e. other than the 12- interval equally tempered scale; Special input devices therefor
- G10H2210/525—Diatonic scales, e.g. aeolian, ionian or major, dorian, locrian, lydian, mixolydian, phrygian, i.e. seven note, octave-repeating musical scales comprising five whole steps and two half steps for each octave, in which the two half steps are separated from each other by either two or three whole steps
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/011—Files or data streams containing coded musical information, e.g. for transmission
- G10H2240/046—File format, i.e. specific or non-standard musical file format used in or adapted for electrophonic musical instruments, e.g. in wavetables
- G10H2240/056—MIDI or other note-oriented file format
Definitions
- the present disclosure relates to the use of Artificial Intelligence to automatically create and generate a multi instrument/sound/note interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like.
- Interactive play-along songs can be played by a multi-media platform, such as available from Beamz Interactive, Inc. of Scottsdale Ariz. Such a platform is described in commonly assigned U.S. Pat. Nos. 7,504,577, 7,858,870, 9,542,919 and 9,646,588, the teachings of each which are incorporated herein by reference in their entirety.
- the Beamz Cross Platform i.e. IOS, Android, PC, MAC, AR, MR, VR, etc.
- Interactive Music Engine is an award winning rich multi-media technology and platform configured to allow a user to selectively trigger various triggers each associated with a subset of a musical composition, to both compose and play sympathetic music. Different subsets/Midi files of a musical composition can be selectively assigned to the triggers.
- the platform includes triggers, and an electronic processor generating a control signal as a function of the triggers selected by a user.
- the electronic processor also generates the control signal as a function of a plurality of music programs. Each music program comprises sound elements comprising a subset of a musical composition, and the music programs are correlated to each other.
- the electronic processor generates audio signals indicative of audible musical sounds as a function of the control signal.
- This disclosure comprises a method and system of using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like, that can later be played on an interactive music engine.
- FIG. 1 illustrates a diagram of a system for multilayered media playback
- FIG. 2 illustrates the Standard Midi numbering scheme used by an application engine
- FIG. 3 illustrates a Virtual Song map prepared by the application engine
- FIG. 4 illustrates the Scale based templates that supply generic notes for further transpositions, and illustrates two possible Scale Note Templates
- FIG. 5 illustrates how a Global Key Offset transposes the preliminary Scale Note pool to become a Seed Note Pool in the scanned song (root) key;
- FIG. 6 illustrates how Global Regional Offsets are used to dynamically transpose Seed Notes to accommodate key changes during song play
- FIG. 7 illustrates two selectable Instruments and their definitions
- FIG. 8 illustrates how the preliminary seed note pool is expanded by the application engine to fill an Instrument's specific Octave Range
- FIG. 9 illustrates how Offsets are dynamically applied by application engine 1007 to an Instrument's preliminary seed note pool to produce a playable sympathetic note pool in real time
- FIG. 10 shows an example of a 13 step Melody Pattern that will play “Mary Had a Little Lamb” melody in any Key and on any Instrument;
- FIG. 11 illustrates the main processing flow executable by the application engine configured to generate an interactive play-along song according to this disclosure.
- This disclosure is directed to using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like.
- the generated interactive play-along song can be played by a multi-media platform, such as available from Beamz Interactive, Inc. of Scottsdale Ariz.
- a multi-media platform such as available from Beamz Interactive, Inc. of Scottsdale Ariz.
- Such a platform is described in commonly assigned U.S. Pat. Nos. 7,504,577, 7,858,870, 9,542,919 and 9,646,588, the teachings of which are incorporated herein by reference in their entirety.
- These methods and illustrations show interactive music content being generated for direct use with the Beamz Interactive Music Engine.
- all song properties and generated note pools can be used by other music engines as long as they can accommodate Tempo and Key changes throughout a song; and they can use pools of sympathetic notes that are appropriate to the key changes as the song plays.
- the Beamz Interactive Music Engine includes an electronic processor generating a control signal as a function of trigger objects selected by a user, the electronic processor also generating the control signal as a function of a plurality of music programs, wherein each said music program comprises sound elements comprising a subset of a musical composition, and the music programs are correlated to each other, the electronic processor generating an audio signals indicative of audible musical sounds as a function of the control signal.
- FIG. 1 illustrates a diagram of a system 1000 for multilayered media playback.
- the system 1000 can be implemented in electronic device 1001 , and embodied as one of a computer, a smart phone, a tablet, a touchscreen computer, and the like having a display 1004 .
- Application engine 1007 is operable on an electronic processor 1006 and receives one or more inputs 1005 from the multiple trigger objects 1003 within the application environment 1001 .
- Application engine 1007 controls playback of media files 1011 that are combined to form a multilayered media file based on one or more of trigger inputs 1005 , and definition file 1010 via sound engine 1008 .
- the media files 1011 can be one or more MIDI files, samples such as .wav and .mp3 files, video files in a plurality of formats, and/or any other audio or video file format.
- the trigger 1003 is held broken for a substantially short period of time, such as with a threshold for the short period of time of 0.5 seconds or less.
- the application engine 1007 can use the short break to trigger a one-shot, play a single note in a streamed sequence, or start and stop a loop.
- the trigger 1003 is held broken for a longer period of time, such as with a threshold for the longer period of time of 0.5 seconds or more. Additional thresholds may be used for long trigger breaks with each threshold associated with a different action to be taken by the application engine 1007 .
- the application engine 1007 can use a long trigger break to Pulse (stream) notes.
- Processor 1006 is configured such that visual outputs from application engine 1007 are displayed on display 1004 and output from sound engine 1008 is played on speaker 1009 .
- the combination of application engine 1007 and sound engine 1008 form an application on the processor 1006 .
- the processor 1006 is configured to selectively associate a music programs with each of the plurality of trigger objects.
- the processor 1006 is configured such that when one of the trigger objects 1003 is in a first state for a prolonged period of time successive said audible musical sounds are generated, such that, for instance, the musical program associated with the trigger object continues to play uninterrupted, along with any other music programs that are playing in response the associated trigger object being triggered.
- Display 1004 displays the application environment 1002 .
- trigger-specific visual output from application engine 1007 can be shown to alter the display properties or attributes of any element in the display 1004 .
- FIG. 2 illustrates the Standard Midi numbering scheme used by application engine 1007 according to the methods of this disclosure. All illustrations & examples herein are based on the Standard Midi Note Numbering scheme.
- Midi is based on a chromatic music scale that breaks an octave into 12 notes (semitones) shown at 11 . There are 128 possible Midi notes 11 spanning 10 octaves, shown at 12 .
- the lowest possible Midi note 11 is “C” in Script #0 shown at 101
- the highest Midi note 11 is “G” in Script #10 shown at 102 .
- each Midi note 11 is identified by its Key followed by the Octave 12 in which the Key falls.
- Midi note C5 represents Middle “C” in Script #5 shown at 103 .
- this standard Midi note numbering scheme is used as a foundation for numerically generating play-along musical notes that will always be sympathetic or compatible with the analyzed audio file/song in real time.
- FIG. 3 illustrates a Virtual Song map 20 .
- This Virtual Song map 20 is a graphic (Music Staff) illustration that is prepared by application engine 1007 executing on processor 1006 from scanning an imported audio file/song file according to this disclosure.
- This example shows the data derived by application engine 1007 from performing Tempo & Chromatic analysis on a sample 10 bar (audio) audio file/song that was recorded in the key of C major at an approximate tempo of 120 BPM.
- the application engine 1007 performs Tempo Analysis on the audio file/song to place Tempo value markers 21 in the Virtual Song map 20 to establish or change the tempo at that point in the audio file/song.
- Tempo value marker 21 When a Tempo value marker 21 is encountered during playback, the current Tempo is set to the value contained in the Tempo value marker 21 .
- the application engine 1007 performs Chromatic Analysis to add Regional Key markers 22 to the Virtual Song map 20 to indicate chromatic changes and their location within the audio file/song.
- Each Regional Key marker 22 contains the chromatic global Region Key Offset 23 needed to transpose Seed Notes 42 from the detected Root Key 26 to Region specific Seed Notes 52 in the new key.
- a Regional Key marker 22 is placed at that location in the Virtual Song map 20 with an offset value 23 of +5.
- the global Region Key Offset 23 is set to the value contained in the marker 22 .
- the application engine 1007 creates Song Sections (Chromatic Regions) shown at 28 .
- Each chromatic region 28 becomes a Song Section (subset of song) in the song's definition as a whole and it marks a key change in the audio file/song.
- the combined lengths 25 of all the Song Sections 28 constitute the song's main loop length 29 .
- All generated play-along notes begin from preprogrammed scale notes which are generated from standard templates, in a standard key, within a standard octave.
- the offsets described herein are used to transpose these notes to a different key as required by the audio file/song.
- Offsets are expressed as semitone values—the number of semitones a given (root) note needs to be transposed into the target note.
- the offset amount is added to the MIDI note number for the root note to obtain the MIDI note number of the target note.
- audio transposition software is used to apply an offset value to transpose a sampled root note to a sample in the frequency of the target key.
- FIG. 4 illustrates the Scale based templates that supply generic notes for further transpositions, and illustrates two possible Scale Note Templates.
- the scanned key data is used to select an appropriate seed note template 31 for the scanned audio file/song.
- a Global Root Key Offset 24 ( FIG. 5 ) is established that, when applied, transposes the templated scale note pool 32 from the key of C to the detected key for the song.
- Scale note templates supply the preliminary note pool 32 that is musically sympathetic to the overall scanned audio file/song. Their associated Midi note numbers 33 are used to mathematically transpose them to the proper key and octave as needed.
- FIG. 5 illustrates how a Global Key Offset transposes the preliminary Scale Note pool 32 to become a Seed Note Pool 42 in the scanned song (root) key.
- the scanned key data is used to determine a Global Root Key Offset 24 which is applied to transpose the preliminary sympathetic note pool 32 from the key of C to the detected key for the audio file/song.
- the chromatic scanning determined the audio file/song to be in the key of G Major and the Global Root Key Offset 24 is set to +7. Adding this offset to the Midi note numbers from the C Major template 33 transposes the seed note pool to the key of G Major 42 , 43 .
- FIG. 6 illustrates how Global Regional Offsets 23 are used to dynamically transpose Seed Notes to accommodate key changes during song play.
- the current Regional Offset 23 is set to the offset value contained in the marker 23 in real time. This ensures that all triggered notes will be in the proper key for what the song is currently playing.
- Region #1 plays in the key of C major while Region #2 is transposed to the key of F major.
- FIG. 7 illustrates two selectable Instruments and their definitions.
- each instrument 61 must be assigned to a Midi Synthesizer Patch 63 within a downloadable sound set 62 .
- the ideal Scripte Range for each instrument is determined from the Low 64 & High 65 Octave values. In FIG. 7 , the octave range for the upright bass is two octaves; the octave range for the electric piano is five octaves.
- FIG. 8 illustrates how the preliminary seed note pool 42 is expanded by application engine 1007 to fill an Instrument's specific Octave Range.
- the Scripte Range 66 is calculated from the Low Scripte 64 and High Script 65 values.
- Octave Range Offsets 67 are established for each Scripte contained in the Scripte Range 66 .
- FIG. 9 illustrates how Offsets are dynamically applied by application engine 1007 to an Instrument's preliminary seed note pool to produce a playable sympathetic note pool in real time.
- the Global Key Offset 24 is added to the Midi Note numbers 33 of the appropriate Scale Note template 31 which transposes them into Seed Notes 42 in the Detected Root Key 26 of the scanned song. (See FIG. 5 ).
- Instrument Script Offsets 67 are added to the Midi Note numbers 72 of the Preliminary Seed Notes 70 . This transposes them to be within the Instrument's specified Octave Ranges 64 - 65 .
- Octave Offsets 67 are used to generate Preliminary Seed Notes 71 for all octaves that fall within the range specified by the Low Scripte 64 and High Scripte 65 properties from the Instrument definition 61 .
- the Global Region Offset 84 is set to the offset value contained in the marker 23 . This ensures that all triggered notes will be in the proper key for what the song is currently playing.
- Each note in the note pool 86 has its own Index ID, 87 which is not affected by any of the Offset values. This index establishes a way to retrieve and play the same melodic pattern across all Regions of the song. See FIG. 10 .
- FIG. 10 illustrates how Melody Patterns work.
- Melody patterns establish a sequential list of notes (melodies) that can be retrieved from an Instrument's Sympathetic Note Pool 86 and played when a user triggers them.
- Melody patterns are predefined within software executing in the application engine 1007 . They are required for playback and can be either selected or established as defaults for an Instrument.
- the same Melody Pattern can be used across all chromatic regions in a song by dynamically adjusting the Regional Offset 84 .
- Any Melody Pattern can be used with any Instrument.
- Melody Patterns do not directly reference the pattern notes by their Key designation 86 . Instead, the notes are referenced by their Index ID 87 within the Instrument's Sympathetic Note Pool ( FIG. 9 ). This makes a Melody Pattern transparent to the Key or Octave to which it applies.
- FIG. 10 shows an example of a 13 step Melody Pattern that will play “Mary Had a Little Lamb” melody in any Key and on any Instrument.
- FIG. 11 illustrates the main processing flow 1100 executable by application engine 1007 configured to generate an interactive play-along song according to this disclosure.
- the application engine 1007 receives a user selection of an audio file/song to import the audio file/song from either a music streaming service or from a personal music library, and the like.
- the imported audio file/song must be in a standard audio format such as .wav or .mp3, or the like.
- a preliminary Virtual Song map 20 is established showing the length of the overall song, its measure alignment grid, tempo setting markers, and where they occur in the audio file/song as it is played.
- a link to the imported file 27 is saved as part of the song definitions in definition file 1010 .
- the length of the imported song 28 is saved as part of the song definitions.
- the imported audio file/song is software scanned by application engine 1007 to determine its beat points (tempo), and its beat/measure alignments (time signature).
- the results of the scan are used by application engine 1007 to prepare a Virtual Song map 20 for the entire song by placing Tempo markers into the map at the appropriate locations along its length, as previously discussed, as shown in FIG. 3 .
- the imported audio file is software analyzed by application engine 1007 to determine its overall key 26 .
- the audio file is also scanned along its length to determine key changes and mark where they occur.
- Regional Key indicator markers 22 are placed in the virtual Song Map 20 at the appropriate locations along its length.
- a Global Key Offset 24 is established. This Global offset transposes all generic Scale Notes 32 from the generic key of C to Seed Notes 42 in the detected key. For example, if the detected key is F, the Global Key Offset 24 would be + 5 .
- the user interface/display 1004 presents the user with a menu of possible play-along instruments 61 or instrument combinations.
- the menu is organized in ways to simplify selections such as organizing the instrument choices by Genre.
- an Instrument-specific Base Pool of notes 71 is established by application engine 1007 for each selected instrument.
- each Instrument selection comes with a predetermined (default) method to trigger it (Trigger Type Pulse; One-shot etc).
- Trigger Type Pulse Trigger Type Pulse; One-shot etc.
- the user may accept or change the triggering characteristics.
- each Instrument selection comes with a default selection of playback templates shown in FIG. 10 that the user may accept or change.
- Playback templates are Melodic in nature and they dictate the melody sequence of sympathetic notes being played.
- step 1109 upon completion, all song selections & properties are saved by the application engine 1007 in an XML format that can be read, edited, and played, such as by Beamz Interactive software offered by Beam Interactive, Inc. of Scottsdale Ariz.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Auxiliary Devices For Music (AREA)
- Electrophonic Musical Instruments (AREA)
Abstract
A method and system of using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like, that can later be played on an interactive music engine.
Description
- This application is a continuation of U.S. patent application U.S. Ser. No. 16/036,352, now issued U.S. Pat. No. 10,534,811, filed Jan. 29, 2018 entitled ARTIFICIAL INTELLIGENCE METHODOLOGY TO AUTOMATICALLY GENERATE INTERACTIVE PLAY ALONG SONGS, the teachings of which are incorporated herein in their entirety.
- The present disclosure relates to the use of Artificial Intelligence to automatically create and generate a multi instrument/sound/note interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like.
- Interactive play-along songs can be played by a multi-media platform, such as available from Beamz Interactive, Inc. of Scottsdale Ariz. Such a platform is described in commonly assigned U.S. Pat. Nos. 7,504,577, 7,858,870, 9,542,919 and 9,646,588, the teachings of each which are incorporated herein by reference in their entirety.
- The Beamz Cross Platform (i.e. IOS, Android, PC, MAC, AR, MR, VR, etc.) Interactive Music Engine is an award winning rich multi-media technology and platform configured to allow a user to selectively trigger various triggers each associated with a subset of a musical composition, to both compose and play sympathetic music. Different subsets/Midi files of a musical composition can be selectively assigned to the triggers. The platform includes triggers, and an electronic processor generating a control signal as a function of the triggers selected by a user. The electronic processor also generates the control signal as a function of a plurality of music programs. Each music program comprises sound elements comprising a subset of a musical composition, and the music programs are correlated to each other. The electronic processor generates audio signals indicative of audible musical sounds as a function of the control signal.
- There is a need to use Artificial Intelligence to automatically create and generate a multi instrument/sound/note interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like, that can later be played on an interactive music engine.
- This disclosure comprises a method and system of using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like, that can later be played on an interactive music engine.
- The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the disclosure and together with the description serve to explain the principles of the disclosure.
-
FIG. 1 illustrates a diagram of a system for multilayered media playback; -
FIG. 2 illustrates the Standard Midi numbering scheme used by an application engine; -
FIG. 3 illustrates a Virtual Song map prepared by the application engine; -
FIG. 4 illustrates the Scale based templates that supply generic notes for further transpositions, and illustrates two possible Scale Note Templates; -
FIG. 5 illustrates how a Global Key Offset transposes the preliminary Scale Note pool to become a Seed Note Pool in the scanned song (root) key; -
FIG. 6 illustrates how Global Regional Offsets are used to dynamically transpose Seed Notes to accommodate key changes during song play; -
FIG. 7 illustrates two selectable Instruments and their definitions; -
FIG. 8 illustrates how the preliminary seed note pool is expanded by the application engine to fill an Instrument's specific Octave Range; -
FIG. 9 illustrates how Offsets are dynamically applied byapplication engine 1007 to an Instrument's preliminary seed note pool to produce a playable sympathetic note pool in real time; -
FIG. 10 shows an example of a 13 step Melody Pattern that will play “Mary Had a Little Lamb” melody in any Key and on any Instrument; and -
FIG. 11 illustrates the main processing flow executable by the application engine configured to generate an interactive play-along song according to this disclosure. - Interactive Music Engine
- This disclosure is directed to using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like. The generated interactive play-along song can be played by a multi-media platform, such as available from Beamz Interactive, Inc. of Scottsdale Ariz. Such a platform is described in commonly assigned U.S. Pat. Nos. 7,504,577, 7,858,870, 9,542,919 and 9,646,588, the teachings of which are incorporated herein by reference in their entirety. These methods and illustrations show interactive music content being generated for direct use with the Beamz Interactive Music Engine. However, all song properties and generated note pools can be used by other music engines as long as they can accommodate Tempo and Key changes throughout a song; and they can use pools of sympathetic notes that are appropriate to the key changes as the song plays.
- The Beamz Interactive Music Engine includes an electronic processor generating a control signal as a function of trigger objects selected by a user, the electronic processor also generating the control signal as a function of a plurality of music programs, wherein each said music program comprises sound elements comprising a subset of a musical composition, and the music programs are correlated to each other, the electronic processor generating an audio signals indicative of audible musical sounds as a function of the control signal.
-
FIG. 1 illustrates a diagram of asystem 1000 for multilayered media playback. - The
system 1000 can be implemented inelectronic device 1001, and embodied as one of a computer, a smart phone, a tablet, a touchscreen computer, and the like having adisplay 1004. -
Application engine 1007 is operable on anelectronic processor 1006 and receives one ormore inputs 1005 from themultiple trigger objects 1003 within theapplication environment 1001. -
Application engine 1007 controls playback ofmedia files 1011 that are combined to form a multilayered media file based on one or more oftrigger inputs 1005, anddefinition file 1010 viasound engine 1008. Themedia files 1011 can be one or more MIDI files, samples such as .wav and .mp3 files, video files in a plurality of formats, and/or any other audio or video file format. - With a short trigger break, the
trigger 1003 is held broken for a substantially short period of time, such as with a threshold for the short period of time of 0.5 seconds or less. Theapplication engine 1007 can use the short break to trigger a one-shot, play a single note in a streamed sequence, or start and stop a loop. - With a long trigger break, the
trigger 1003 is held broken for a longer period of time, such as with a threshold for the longer period of time of 0.5 seconds or more. Additional thresholds may be used for long trigger breaks with each threshold associated with a different action to be taken by theapplication engine 1007. - The
application engine 1007 can use a long trigger break to Pulse (stream) notes. -
Processor 1006 is configured such that visual outputs fromapplication engine 1007 are displayed ondisplay 1004 and output fromsound engine 1008 is played onspeaker 1009. - The combination of
application engine 1007 andsound engine 1008 form an application on theprocessor 1006. Theprocessor 1006 is configured to selectively associate a music programs with each of the plurality of trigger objects. Theprocessor 1006 is configured such that when one of thetrigger objects 1003 is in a first state for a prolonged period of time successive said audible musical sounds are generated, such that, for instance, the musical program associated with the trigger object continues to play uninterrupted, along with any other music programs that are playing in response the associated trigger object being triggered. -
Display 1004 displays theapplication environment 1002. - When a
trigger object 1003 is triggered by a user, trigger-specific visual output fromapplication engine 1007 can be shown to alter the display properties or attributes of any element in thedisplay 1004. -
FIG. 2 illustrates the Standard Midi numbering scheme used byapplication engine 1007 according to the methods of this disclosure. All illustrations & examples herein are based on the Standard Midi Note Numbering scheme. - Midi is based on a chromatic music scale that breaks an octave into 12 notes (semitones) shown at 11. There are 128
possible Midi notes 11 spanning 10 octaves, shown at 12. - The lowest
possible Midi note 11 is “C” in Octave #0 shown at 101, thehighest Midi note 11 is “G” in Octave #10 shown at 102. - As illustrated in
FIG. 2 , eachMidi note 11 is identified by its Key followed by the Octave 12 in which the Key falls. For example, Midi note C5 represents Middle “C” in Octave #5 shown at 103. - All Midi notes 11 are numbered at 13 sequentially from the lowest to the highest (0-127)
- By using a combination of generic Offset Templates, this standard Midi note numbering scheme is used as a foundation for numerically generating play-along musical notes that will always be sympathetic or compatible with the analyzed audio file/song in real time.
-
FIG. 3 illustrates aVirtual Song map 20. ThisVirtual Song map 20 is a graphic (Music Staff) illustration that is prepared byapplication engine 1007 executing onprocessor 1006 from scanning an imported audio file/song file according to this disclosure. - When the song is played by
application engine 1007 andsong engine 1008, it follows thisVirtual Song map 20, responding to the markers as they are encountered. - This example shows the data derived by
application engine 1007 from performing Tempo & Chromatic analysis on asample 10 bar (audio) audio file/song that was recorded in the key of C major at an approximate tempo of 120 BPM. - The
application engine 1007 performs Tempo Analysis on the audio file/song to placeTempo value markers 21 in theVirtual Song map 20 to establish or change the tempo at that point in the audio file/song. When aTempo value marker 21 is encountered during playback, the current Tempo is set to the value contained in theTempo value marker 21. - The
application engine 1007 performs Chromatic Analysis to add RegionalKey markers 22 to theVirtual Song map 20 to indicate chromatic changes and their location within the audio file/song. - Each
Regional Key marker 22 contains the chromatic global Region Key Offset 23 needed to transposeSeed Notes 42 from the detectedRoot Key 26 to Regionspecific Seed Notes 52 in the new key. - For example, if the audio file/song is detected to be in C, and the scan detects that, during play, the key changes to F, a
Regional Key marker 22 is placed at that location in theVirtual Song map 20 with an offsetvalue 23 of +5. - When a
Regional Key marker 22 is encountered during playback of the audio file/song, the global Region Key Offset 23 is set to the value contained in themarker 22. - The
application engine 1007 creates Song Sections (Chromatic Regions) shown at 28. Eachchromatic region 28 becomes a Song Section (subset of song) in the song's definition as a whole and it marks a key change in the audio file/song. - In this example, there are 6
Song Sections 28—each having their own properties.Key 22, Regional (key) offset 23, andLength 25. - The combined
lengths 25 of all theSong Sections 28 constitute the song'smain loop length 29. - Use of Transposition Offsets
- The following methods for generating sympathetic note pools are based on the use and application of multiple “Offsets”.
- All generated play-along notes begin from preprogrammed scale notes which are generated from standard templates, in a standard key, within a standard octave.
- The offsets described herein are used to transpose these notes to a different key as required by the audio file/song.
- Offsets are expressed as semitone values—the number of semitones a given (root) note needs to be transposed into the target note.
- For the sake of simplicity, all illustrations and methods described herein use the standard MIDI note numbering scheme to easily show how notes can be transposed by semitones.
- However, these transposition offset methods apply to both MIDI notes and note samples as well.
- With root notes in MIDI format, the offset amount is added to the MIDI note number for the root note to obtain the MIDI note number of the target note.
- With sampled notes, audio transposition software is used to apply an offset value to transpose a sampled root note to a sample in the frequency of the target key.
-
FIG. 4 illustrates the Scale based templates that supply generic notes for further transpositions, and illustrates two possible Scale Note Templates. - After chromatic scanning is completed by
application engine 1007, the scanned key data is used to select an appropriateseed note template 31 for the scanned audio file/song. A Global Root Key Offset 24 (FIG. 5 ) is established that, when applied, transposes the templatedscale note pool 32 from the key of C to the detected key for the song. - All Note Templates are based on the standard Midi note numbering scheme (
FIG. 2 ) and are prepared in the key of C with a Root Note of CO 101 (lowest possible Midi note). - Scale note templates supply the
preliminary note pool 32 that is musically sympathetic to the overall scanned audio file/song. Their associatedMidi note numbers 33 are used to mathematically transpose them to the proper key and octave as needed. -
FIG. 5 illustrates how a Global Key Offset transposes the preliminaryScale Note pool 32 to become aSeed Note Pool 42 in the scanned song (root) key. - After chromatic scanning is complete, the scanned key data is used to determine a Global Root Key Offset 24 which is applied to transpose the preliminary
sympathetic note pool 32 from the key of C to the detected key for the audio file/song. - In this illustration, the chromatic scanning determined the audio file/song to be in the key of G Major and the Global Root Key Offset 24 is set to +7. Adding this offset to the Midi note numbers from the
C Major template 33 transposes the seed note pool to the key ofG Major - FOR THE SAKE OF SIMPLICITY, ALL OTHER FIGURES ARE ILLUSTRATED WITH AN EXAMPLE SONG IN THE KEY OF C. (GLOBAL KEY OFFSET=0)
-
FIG. 6 illustrates how Global Regional Offsets 23 are used to dynamically transpose Seed Notes to accommodate key changes during song play. - As the audio file/song plays, and (chromatic) Song Section markers are encountered, the current Regional Offset 23 is set to the offset value contained in the
marker 23 in real time. This ensures that all triggered notes will be in the proper key for what the song is currently playing. - In this example,
Region # 1 plays in the key of C major whileRegion # 2 is transposed to the key of F major. -
FIG. 7 illustrates two selectable Instruments and their definitions. - All predefined instrument choices contain information necessary to use that instrument in a musical composition. If the notes to be played are MIDI, each
instrument 61 must be assigned to aMidi Synthesizer Patch 63 within a downloadable sound set 62. The ideal Octave Range for each instrument is determined from the Low 64 &High 65 Octave values. InFIG. 7 , the octave range for the upright bass is two octaves; the octave range for the electric piano is five octaves. -
FIG. 8 illustrates how the preliminaryseed note pool 42 is expanded byapplication engine 1007 to fill an Instrument's specific Octave Range. - For each Instrument selected by the user:
- The
Octave Range 66 is calculated from theLow Octave 64 andHigh Octave 65 values. - Octave Range Offsets 67 are established for each Octave contained in the
Octave Range 66. - For example, all
Seed Note Pools 42 are always generated forOctave # 0. - If an instrument sounds best in
octave # 5, the offset value (in chromatic semi-tones) would be +60 (5×12 semis). - Using the Octave Range Offsets 67, one set of
preliminary Seed Notes 71 is generated for each octave contained in theOctave Range 66. -
FIG. 9 illustrates how Offsets are dynamically applied byapplication engine 1007 to an Instrument's preliminary seed note pool to produce a playable sympathetic note pool in real time. - All of an Instrument's play-notes start out as standard Scale Notes contained in a
Scale Note Template 31. The appropriate template is indicated by chromatic analysis of the imported song. Chromatic analysis also sets the Global Key Offset 24 (SeeFIG. 5 ). - The Global Key Offset 24 is added to the Midi Note numbers 33 of the appropriate
Scale Note template 31 which transposes them intoSeed Notes 42 in the DetectedRoot Key 26 of the scanned song. (SeeFIG. 5 ). - Instrument Octave Offsets 67 are added to the Midi Note numbers 72 of the
Preliminary Seed Notes 70. This transposes them to be within the Instrument's specified Octave Ranges 64-65. - Octave Offsets 67 are used to generate
Preliminary Seed Notes 71 for all octaves that fall within the range specified by theLow Octave 64 andHigh Octave 65 properties from theInstrument definition 61. - As the
application engine 1007 plays the audio file/song in real time, and (chromatic) Song Section markers are encountered, the Global Region Offset 84 is set to the offset value contained in themarker 23. This ensures that all triggered notes will be in the proper key for what the song is currently playing. - Each note in the
note pool 86 has its own Index ID, 87 which is not affected by any of the Offset values. This index establishes a way to retrieve and play the same melodic pattern across all Regions of the song. SeeFIG. 10 . -
FIG. 10 illustrates how Melody Patterns work. Melody patterns establish a sequential list of notes (melodies) that can be retrieved from an Instrument'sSympathetic Note Pool 86 and played when a user triggers them. - Melody patterns are predefined within software executing in the
application engine 1007. They are required for playback and can be either selected or established as defaults for an Instrument. - Melody Patterns can have an unlimited number of steps.
- The same Melody Pattern can be used across all chromatic regions in a song by dynamically adjusting the Regional Offset 84.
- Any Melody Pattern can be used with any Instrument.
- To provide flexibility across all keys, Melody Patterns do not directly reference the pattern notes by their
Key designation 86. Instead, the notes are referenced by theirIndex ID 87 within the Instrument's Sympathetic Note Pool (FIG. 9 ). This makes a Melody Pattern transparent to the Key or Octave to which it applies. -
FIG. 10 shows an example of a 13 step Melody Pattern that will play “Mary Had a Little Lamb” melody in any Key and on any Instrument. -
FIG. 11 illustrates themain processing flow 1100 executable byapplication engine 1007 configured to generate an interactive play-along song according to this disclosure. - At
step 1101 theapplication engine 1007 receives a user selection of an audio file/song to import the audio file/song from either a music streaming service or from a personal music library, and the like. - The imported audio file/song must be in a standard audio format such as .wav or .mp3, or the like. A preliminary
Virtual Song map 20 is established showing the length of the overall song, its measure alignment grid, tempo setting markers, and where they occur in the audio file/song as it is played. A link to the importedfile 27 is saved as part of the song definitions indefinition file 1010. The length of the importedsong 28 is saved as part of the song definitions. - At
step 1102 the imported audio file/song is software scanned byapplication engine 1007 to determine its beat points (tempo), and its beat/measure alignments (time signature). The results of the scan are used byapplication engine 1007 to prepare aVirtual Song map 20 for the entire song by placing Tempo markers into the map at the appropriate locations along its length, as previously discussed, as shown inFIG. 3 . - At
step 1103 the imported audio file is software analyzed byapplication engine 1007 to determine itsoverall key 26. - The audio file is also scanned along its length to determine key changes and mark where they occur.
- Regional
Key indicator markers 22 are placed in thevirtual Song Map 20 at the appropriate locations along its length. - All predesigned Key/
Scale templates 31 contain a genericScale Note pool 32 in the key of C inOctave # 0. - Based on the detected key for the audio file/song, a Global Key Offset 24 is established. This Global offset transposes all
generic Scale Notes 32 from the generic key of C toSeed Notes 42 in the detected key. For example, if the detected key is F, the Global Key Offset 24 would be +5. - At step 1104 a basic pool of Sympathetic Seed Notes for the song is established by the
application engine 1007 applying the Global Key offset to the scale notes contained in the Generic Key/Scale template. (Template Note) 32+(Global Key Offset) 24=Seed Note Pool 42. - At
step 1105 the user interface/display 1004 presents the user with a menu of possible play-alonginstruments 61 or instrument combinations. The menu is organized in ways to simplify selections such as organizing the instrument choices by Genre. - At
step 1106 an Instrument-specific Base Pool ofnotes 71 is established byapplication engine 1007 for each selected instrument. - At
step 1107 each Instrument selection comes with a predetermined (default) method to trigger it (Trigger Type Pulse; One-shot etc). Upon user selection of an Instrument, the user may accept or change the triggering characteristics. - At
step 1108 each Instrument selection comes with a default selection of playback templates shown inFIG. 10 that the user may accept or change. Playback templates are Melodic in nature and they dictate the melody sequence of sympathetic notes being played. - At
step 1109, upon completion, all song selections & properties are saved by theapplication engine 1007 in an XML format that can be read, edited, and played, such as by Beamz Interactive software offered by Beam Interactive, Inc. of Scottsdale Ariz. - The appended claims set forth novel and inventive aspects of the subject matter described above, but the claims may also encompass additional subject matter not specifically recited in detail. For example, certain features, elements, or aspects may be omitted from the claims if not necessary to distinguish the novel and inventive features from what is already known to a person having ordinary skill in the art. Features, elements, and aspects described herein may also be combined or replaced by alternative features serving the same, equivalent, or similar purpose without departing from the scope of the invention defined by the appended claims.
Claims (18)
1. A non-transitory computer readable medium including instructions operable by an electronic processor, comprising instructions for:
receiving an audio file from a remote source;
directly analyzing the audio file to map in audio form and create a virtual song definition configured for interactive playback; and
performing chromatic analysis across a length of the audio file to create a song section in the virtual song definition, where each song section contains a regional key offset that is configured to transpose seed notes from a detected root key to regional specific seed notes for the song section in real-time.
2. The non-transitory computer readable medium as specified in claim 1 , further comprising instructions for performing tempo analysis on the length of the audio file to create a tempo region in the virtual song definition, where each tempo region contains a tempo value that establishes a playback tempo for the region.
3. The non-transitory computer readable medium as specified in claim 2 , further comprising instructions such that when a tempo value marker is encountered during playback of the virtual song definition, a current tempo is set to a value contained in the tempo value marker.
4. The non-transitory computer readable medium as specified in claim 1 , further including instructions to use pools of sympathetic notes that correspond to key changes as the audio file plays.
5. The non-transitory computer readable medium as specified in claim 4 , further including instructions for associating an instrument to one said pool of sympathetic notes.
6. The non-transitory computer readable medium as specified in claim 5 , further including instructions for associating a plurality of instruments to respective pools of sympathetic notes.
7. The non-transitory computer readable medium as specified in claim 5 , further comprising instructions for assigning a melody template to each instrument's pool of sympathetic notes to govern how notes are retrieved and played.
8. The non-transitory computer readable medium as specified in claim 1 , further comprising instructions for saving audio file definitions in a definition file.
9. The non-transitory computer readable medium as specified in claim 8 , wherein definition file is an XML file.
10. A system, comprising;
a processor configured to:
receive an audio file from a remote source;
directly analyze the audio file to map the audio file and create a virtual song definition map configured for interactive playback; and
perform chromatic analysis across a length of the audio file to create song section objects in the virtual song definition, where each song section contains a regional key offset that is configured to transpose seed notes from a detected root key to regional specific seed notes for the song section in real-time.
11. The system as specified in claim 10 , wherein the processor is configured to:
perform tempo analysis on the length of the audio file to create tempo region objects in the virtual song definition, where each tempo region contains a tempo value that establishes a playback tempo for the region; and
perform tempo analysis on the audio file to create the virtual song map and place tempo value markers in the virtual song map to establish or change the tempo at that point in the audio file.
12. The system as specified in claim 11 , wherein the processor is configured such that when a tempo value marker is encountered during playback of the virtual song definition, a current tempo is set to a value contained in the tempo value marker.
13. The system as specified in claim 10 , wherein the processor is configured to use pools of sympathetic notes that correspond to key changes as the audio file plays.
14. The system as specified in claim 13 , wherein the processor is configured to associate an instrument to one said pool of sympathetic notes.
15. The system as specified in claim 14 , wherein the processor is configured to associate a plurality of instruments to respective pools of sympathetic notes.
16. The system as specified in claim 14 , wherein the processor is configured to assign a melody template to each instrument's pool of sympathetic notes to govern how notes are retrieved and played.
17. The system as specified in claim 10 , wherein the processor is configured to save audio file definitions in a definition file.
18. The system as specified in claim 17 , wherein definition file is an XML file.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/732,021 US20200142926A1 (en) | 2018-01-29 | 2019-12-31 | Artificial intelligence methodology to automatically generate interactive play along songs |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201862623131P | 2018-01-29 | 2018-01-29 | |
US16/036,052 US10534811B2 (en) | 2018-01-29 | 2018-07-16 | Artificial intelligence methodology to automatically generate interactive play along songs |
US16/732,021 US20200142926A1 (en) | 2018-01-29 | 2019-12-31 | Artificial intelligence methodology to automatically generate interactive play along songs |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/036,052 Continuation US10534811B2 (en) | 2018-01-29 | 2018-07-16 | Artificial intelligence methodology to automatically generate interactive play along songs |
Publications (1)
Publication Number | Publication Date |
---|---|
US20200142926A1 true US20200142926A1 (en) | 2020-05-07 |
Family
ID=67392149
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/036,052 Expired - Fee Related US10534811B2 (en) | 2018-01-29 | 2018-07-16 | Artificial intelligence methodology to automatically generate interactive play along songs |
US16/732,021 Abandoned US20200142926A1 (en) | 2018-01-29 | 2019-12-31 | Artificial intelligence methodology to automatically generate interactive play along songs |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/036,052 Expired - Fee Related US10534811B2 (en) | 2018-01-29 | 2018-07-16 | Artificial intelligence methodology to automatically generate interactive play along songs |
Country Status (1)
Country | Link |
---|---|
US (2) | US10534811B2 (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110958351A (en) * | 2019-11-26 | 2020-04-03 | 东莞市鼎酷科技有限公司 | Message prompt tone processing method and device, storage medium and terminal |
US20210303618A1 (en) * | 2020-03-31 | 2021-09-30 | Aries Adaptive Media, LLC | Processes and systems for mixing audio tracks according to a template |
CN111930289B (en) * | 2020-09-09 | 2021-05-07 | 智者四海(北京)技术有限公司 | Method and system for processing pictures and texts |
CN116764240A (en) * | 2022-03-10 | 2023-09-19 | 腾讯科技(深圳)有限公司 | Performance recording method, device, equipment and storage medium in virtual scene |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010039872A1 (en) * | 2000-05-11 | 2001-11-15 | Cliff David Trevor | Automatic compilation of songs |
US20030094093A1 (en) * | 2001-05-04 | 2003-05-22 | David Smith | Music performance system |
US20080271592A1 (en) * | 2003-08-20 | 2008-11-06 | David Joseph Beckford | System, computer program and method for quantifying and analyzing musical intellectual property |
US20090048694A1 (en) * | 2005-07-01 | 2009-02-19 | Pioneer Corporation | Computer program, information reproduction device, and method |
US7667125B2 (en) * | 2007-02-01 | 2010-02-23 | Museami, Inc. | Music transcription |
US7956276B2 (en) * | 2006-12-04 | 2011-06-07 | Sony Corporation | Method of distributing mashup data, mashup method, server apparatus for mashup data, and mashup apparatus |
US8115090B2 (en) * | 2006-11-28 | 2012-02-14 | Sony Corporation | Mashup data file, mashup apparatus, and content creation method |
US20140018947A1 (en) * | 2012-07-16 | 2014-01-16 | SongFlutter, Inc. | System and Method for Combining Two or More Songs in a Queue |
US9111519B1 (en) * | 2011-10-26 | 2015-08-18 | Mixwolf LLC | System and method for generating cuepoints for mixing song data |
-
2018
- 2018-07-16 US US16/036,052 patent/US10534811B2/en not_active Expired - Fee Related
-
2019
- 2019-12-31 US US16/732,021 patent/US20200142926A1/en not_active Abandoned
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010039872A1 (en) * | 2000-05-11 | 2001-11-15 | Cliff David Trevor | Automatic compilation of songs |
US20030094093A1 (en) * | 2001-05-04 | 2003-05-22 | David Smith | Music performance system |
US20080271592A1 (en) * | 2003-08-20 | 2008-11-06 | David Joseph Beckford | System, computer program and method for quantifying and analyzing musical intellectual property |
US20090048694A1 (en) * | 2005-07-01 | 2009-02-19 | Pioneer Corporation | Computer program, information reproduction device, and method |
US8115090B2 (en) * | 2006-11-28 | 2012-02-14 | Sony Corporation | Mashup data file, mashup apparatus, and content creation method |
US7956276B2 (en) * | 2006-12-04 | 2011-06-07 | Sony Corporation | Method of distributing mashup data, mashup method, server apparatus for mashup data, and mashup apparatus |
US7667125B2 (en) * | 2007-02-01 | 2010-02-23 | Museami, Inc. | Music transcription |
US9111519B1 (en) * | 2011-10-26 | 2015-08-18 | Mixwolf LLC | System and method for generating cuepoints for mixing song data |
US20140018947A1 (en) * | 2012-07-16 | 2014-01-16 | SongFlutter, Inc. | System and Method for Combining Two or More Songs in a Queue |
Also Published As
Publication number | Publication date |
---|---|
US10534811B2 (en) | 2020-01-14 |
US20190236209A1 (en) | 2019-08-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20200142926A1 (en) | Artificial intelligence methodology to automatically generate interactive play along songs | |
US7743092B2 (en) | Method for recognizing and distributing music | |
JP4111004B2 (en) | Performance practice device and performance practice program | |
US8035020B2 (en) | Collaborative music creation | |
US7601904B2 (en) | Interactive tool and appertaining method for creating a graphical music display | |
US6924425B2 (en) | Method and apparatus for storing a multipart audio performance with interactive playback | |
ES2539813T3 (en) | Music transcription | |
US7741554B2 (en) | Apparatus and method for automatically creating music piece data | |
US7572968B2 (en) | Electronic musical instrument | |
AU784788B2 (en) | Array or equipment for composing | |
WO2006112584A1 (en) | Music composing device | |
US20020144587A1 (en) | Virtual music system | |
JP5229998B2 (en) | Code name detection device and code name detection program | |
KR100664677B1 (en) | Method for generating music contents using handheld terminal | |
US20020144588A1 (en) | Multimedia data file | |
JP4274272B2 (en) | Arpeggio performance device | |
US6046396A (en) | Stringed musical instrument performance information composing apparatus and method | |
JP3915807B2 (en) | Automatic performance determination device and program | |
JP2016142967A (en) | Accompaniment training apparatus and accompaniment training program | |
JP4219652B2 (en) | A singing practice support system for a karaoke device that controls the main melody volume at the relevant location based on the pitch error measured immediately before repeat performance | |
JP7419830B2 (en) | Accompaniment sound generation device, electronic musical instrument, accompaniment sound generation method, and accompaniment sound generation program | |
JP3949544B2 (en) | Karaoke device that displays error of singing voice pitch on bar graph | |
JP3879524B2 (en) | Waveform generation method, performance data processing method, and waveform selection device | |
WO2022172732A1 (en) | Information processing system, electronic musical instrument, information processing method, and machine learning system | |
JP2018146716A (en) | Training device, training program, and training method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
AS | Assignment |
Owner name: TOPDOWN LICENSING LLC, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BEAMZ IP, LLC;REEL/FRAME:053029/0492 Effective date: 20200528 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |