US20200142926A1 - Artificial intelligence methodology to automatically generate interactive play along songs - Google Patents

Artificial intelligence methodology to automatically generate interactive play along songs Download PDF

Info

Publication number
US20200142926A1
US20200142926A1 US16/732,021 US201916732021A US2020142926A1 US 20200142926 A1 US20200142926 A1 US 20200142926A1 US 201916732021 A US201916732021 A US 201916732021A US 2020142926 A1 US2020142926 A1 US 2020142926A1
Authority
US
United States
Prior art keywords
song
tempo
audio file
notes
specified
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/732,021
Inventor
Gary Bencar
Todd Nystrom
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
TOPDOWN LICENSING LLC
Original Assignee
Beamz Ip LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beamz Ip LLC filed Critical Beamz Ip LLC
Priority to US16/732,021 priority Critical patent/US20200142926A1/en
Publication of US20200142926A1 publication Critical patent/US20200142926A1/en
Assigned to TOPDOWN LICENSING LLC reassignment TOPDOWN LICENSING LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BEAMZ IP, LLC
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/638Presentation of query results
    • G06F16/639Presentation of query results using playlists
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • G10H1/0025Automatic or semi-automatic music composition, e.g. producing random music, applying rules from music theory or modifying a musical piece
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0033Recording/reproducing or transmission of music for electrophonic musical instruments
    • G10H1/0041Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
    • G10H1/0058Transmission between separate instruments or between individual components of a musical system
    • G10H1/0066Transmission between separate instruments or between individual components of a musical system using a MIDI interface
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/18Selecting circuits
    • G10H1/20Selecting circuits for transposition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/066Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for pitch analysis as part of wider processing for musical purposes, e.g. transcription, musical performance evaluation; Pitch recognition, e.g. in polyphonic sounds; Estimation or use of missing fundamental
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/076Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction of timing, tempo; Beat detection
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/081Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for automatic key or tonality recognition, e.g. using musical rules or a knowledge base
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/395Special musical scales, i.e. other than the 12- interval equally tempered scale; Special input devices therefor
    • G10H2210/525Diatonic scales, e.g. aeolian, ionian or major, dorian, locrian, lydian, mixolydian, phrygian, i.e. seven note, octave-repeating musical scales comprising five whole steps and two half steps for each octave, in which the two half steps are separated from each other by either two or three whole steps
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/011Files or data streams containing coded musical information, e.g. for transmission
    • G10H2240/046File format, i.e. specific or non-standard musical file format used in or adapted for electrophonic musical instruments, e.g. in wavetables
    • G10H2240/056MIDI or other note-oriented file format

Definitions

  • the present disclosure relates to the use of Artificial Intelligence to automatically create and generate a multi instrument/sound/note interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like.
  • Interactive play-along songs can be played by a multi-media platform, such as available from Beamz Interactive, Inc. of Scottsdale Ariz. Such a platform is described in commonly assigned U.S. Pat. Nos. 7,504,577, 7,858,870, 9,542,919 and 9,646,588, the teachings of each which are incorporated herein by reference in their entirety.
  • the Beamz Cross Platform i.e. IOS, Android, PC, MAC, AR, MR, VR, etc.
  • Interactive Music Engine is an award winning rich multi-media technology and platform configured to allow a user to selectively trigger various triggers each associated with a subset of a musical composition, to both compose and play sympathetic music. Different subsets/Midi files of a musical composition can be selectively assigned to the triggers.
  • the platform includes triggers, and an electronic processor generating a control signal as a function of the triggers selected by a user.
  • the electronic processor also generates the control signal as a function of a plurality of music programs. Each music program comprises sound elements comprising a subset of a musical composition, and the music programs are correlated to each other.
  • the electronic processor generates audio signals indicative of audible musical sounds as a function of the control signal.
  • This disclosure comprises a method and system of using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like, that can later be played on an interactive music engine.
  • FIG. 1 illustrates a diagram of a system for multilayered media playback
  • FIG. 2 illustrates the Standard Midi numbering scheme used by an application engine
  • FIG. 3 illustrates a Virtual Song map prepared by the application engine
  • FIG. 4 illustrates the Scale based templates that supply generic notes for further transpositions, and illustrates two possible Scale Note Templates
  • FIG. 5 illustrates how a Global Key Offset transposes the preliminary Scale Note pool to become a Seed Note Pool in the scanned song (root) key;
  • FIG. 6 illustrates how Global Regional Offsets are used to dynamically transpose Seed Notes to accommodate key changes during song play
  • FIG. 7 illustrates two selectable Instruments and their definitions
  • FIG. 8 illustrates how the preliminary seed note pool is expanded by the application engine to fill an Instrument's specific Octave Range
  • FIG. 9 illustrates how Offsets are dynamically applied by application engine 1007 to an Instrument's preliminary seed note pool to produce a playable sympathetic note pool in real time
  • FIG. 10 shows an example of a 13 step Melody Pattern that will play “Mary Had a Little Lamb” melody in any Key and on any Instrument;
  • FIG. 11 illustrates the main processing flow executable by the application engine configured to generate an interactive play-along song according to this disclosure.
  • This disclosure is directed to using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like.
  • the generated interactive play-along song can be played by a multi-media platform, such as available from Beamz Interactive, Inc. of Scottsdale Ariz.
  • a multi-media platform such as available from Beamz Interactive, Inc. of Scottsdale Ariz.
  • Such a platform is described in commonly assigned U.S. Pat. Nos. 7,504,577, 7,858,870, 9,542,919 and 9,646,588, the teachings of which are incorporated herein by reference in their entirety.
  • These methods and illustrations show interactive music content being generated for direct use with the Beamz Interactive Music Engine.
  • all song properties and generated note pools can be used by other music engines as long as they can accommodate Tempo and Key changes throughout a song; and they can use pools of sympathetic notes that are appropriate to the key changes as the song plays.
  • the Beamz Interactive Music Engine includes an electronic processor generating a control signal as a function of trigger objects selected by a user, the electronic processor also generating the control signal as a function of a plurality of music programs, wherein each said music program comprises sound elements comprising a subset of a musical composition, and the music programs are correlated to each other, the electronic processor generating an audio signals indicative of audible musical sounds as a function of the control signal.
  • FIG. 1 illustrates a diagram of a system 1000 for multilayered media playback.
  • the system 1000 can be implemented in electronic device 1001 , and embodied as one of a computer, a smart phone, a tablet, a touchscreen computer, and the like having a display 1004 .
  • Application engine 1007 is operable on an electronic processor 1006 and receives one or more inputs 1005 from the multiple trigger objects 1003 within the application environment 1001 .
  • Application engine 1007 controls playback of media files 1011 that are combined to form a multilayered media file based on one or more of trigger inputs 1005 , and definition file 1010 via sound engine 1008 .
  • the media files 1011 can be one or more MIDI files, samples such as .wav and .mp3 files, video files in a plurality of formats, and/or any other audio or video file format.
  • the trigger 1003 is held broken for a substantially short period of time, such as with a threshold for the short period of time of 0.5 seconds or less.
  • the application engine 1007 can use the short break to trigger a one-shot, play a single note in a streamed sequence, or start and stop a loop.
  • the trigger 1003 is held broken for a longer period of time, such as with a threshold for the longer period of time of 0.5 seconds or more. Additional thresholds may be used for long trigger breaks with each threshold associated with a different action to be taken by the application engine 1007 .
  • the application engine 1007 can use a long trigger break to Pulse (stream) notes.
  • Processor 1006 is configured such that visual outputs from application engine 1007 are displayed on display 1004 and output from sound engine 1008 is played on speaker 1009 .
  • the combination of application engine 1007 and sound engine 1008 form an application on the processor 1006 .
  • the processor 1006 is configured to selectively associate a music programs with each of the plurality of trigger objects.
  • the processor 1006 is configured such that when one of the trigger objects 1003 is in a first state for a prolonged period of time successive said audible musical sounds are generated, such that, for instance, the musical program associated with the trigger object continues to play uninterrupted, along with any other music programs that are playing in response the associated trigger object being triggered.
  • Display 1004 displays the application environment 1002 .
  • trigger-specific visual output from application engine 1007 can be shown to alter the display properties or attributes of any element in the display 1004 .
  • FIG. 2 illustrates the Standard Midi numbering scheme used by application engine 1007 according to the methods of this disclosure. All illustrations & examples herein are based on the Standard Midi Note Numbering scheme.
  • Midi is based on a chromatic music scale that breaks an octave into 12 notes (semitones) shown at 11 . There are 128 possible Midi notes 11 spanning 10 octaves, shown at 12 .
  • the lowest possible Midi note 11 is “C” in Script #0 shown at 101
  • the highest Midi note 11 is “G” in Script #10 shown at 102 .
  • each Midi note 11 is identified by its Key followed by the Octave 12 in which the Key falls.
  • Midi note C5 represents Middle “C” in Script #5 shown at 103 .
  • this standard Midi note numbering scheme is used as a foundation for numerically generating play-along musical notes that will always be sympathetic or compatible with the analyzed audio file/song in real time.
  • FIG. 3 illustrates a Virtual Song map 20 .
  • This Virtual Song map 20 is a graphic (Music Staff) illustration that is prepared by application engine 1007 executing on processor 1006 from scanning an imported audio file/song file according to this disclosure.
  • This example shows the data derived by application engine 1007 from performing Tempo & Chromatic analysis on a sample 10 bar (audio) audio file/song that was recorded in the key of C major at an approximate tempo of 120 BPM.
  • the application engine 1007 performs Tempo Analysis on the audio file/song to place Tempo value markers 21 in the Virtual Song map 20 to establish or change the tempo at that point in the audio file/song.
  • Tempo value marker 21 When a Tempo value marker 21 is encountered during playback, the current Tempo is set to the value contained in the Tempo value marker 21 .
  • the application engine 1007 performs Chromatic Analysis to add Regional Key markers 22 to the Virtual Song map 20 to indicate chromatic changes and their location within the audio file/song.
  • Each Regional Key marker 22 contains the chromatic global Region Key Offset 23 needed to transpose Seed Notes 42 from the detected Root Key 26 to Region specific Seed Notes 52 in the new key.
  • a Regional Key marker 22 is placed at that location in the Virtual Song map 20 with an offset value 23 of +5.
  • the global Region Key Offset 23 is set to the value contained in the marker 22 .
  • the application engine 1007 creates Song Sections (Chromatic Regions) shown at 28 .
  • Each chromatic region 28 becomes a Song Section (subset of song) in the song's definition as a whole and it marks a key change in the audio file/song.
  • the combined lengths 25 of all the Song Sections 28 constitute the song's main loop length 29 .
  • All generated play-along notes begin from preprogrammed scale notes which are generated from standard templates, in a standard key, within a standard octave.
  • the offsets described herein are used to transpose these notes to a different key as required by the audio file/song.
  • Offsets are expressed as semitone values—the number of semitones a given (root) note needs to be transposed into the target note.
  • the offset amount is added to the MIDI note number for the root note to obtain the MIDI note number of the target note.
  • audio transposition software is used to apply an offset value to transpose a sampled root note to a sample in the frequency of the target key.
  • FIG. 4 illustrates the Scale based templates that supply generic notes for further transpositions, and illustrates two possible Scale Note Templates.
  • the scanned key data is used to select an appropriate seed note template 31 for the scanned audio file/song.
  • a Global Root Key Offset 24 ( FIG. 5 ) is established that, when applied, transposes the templated scale note pool 32 from the key of C to the detected key for the song.
  • Scale note templates supply the preliminary note pool 32 that is musically sympathetic to the overall scanned audio file/song. Their associated Midi note numbers 33 are used to mathematically transpose them to the proper key and octave as needed.
  • FIG. 5 illustrates how a Global Key Offset transposes the preliminary Scale Note pool 32 to become a Seed Note Pool 42 in the scanned song (root) key.
  • the scanned key data is used to determine a Global Root Key Offset 24 which is applied to transpose the preliminary sympathetic note pool 32 from the key of C to the detected key for the audio file/song.
  • the chromatic scanning determined the audio file/song to be in the key of G Major and the Global Root Key Offset 24 is set to +7. Adding this offset to the Midi note numbers from the C Major template 33 transposes the seed note pool to the key of G Major 42 , 43 .
  • FIG. 6 illustrates how Global Regional Offsets 23 are used to dynamically transpose Seed Notes to accommodate key changes during song play.
  • the current Regional Offset 23 is set to the offset value contained in the marker 23 in real time. This ensures that all triggered notes will be in the proper key for what the song is currently playing.
  • Region #1 plays in the key of C major while Region #2 is transposed to the key of F major.
  • FIG. 7 illustrates two selectable Instruments and their definitions.
  • each instrument 61 must be assigned to a Midi Synthesizer Patch 63 within a downloadable sound set 62 .
  • the ideal Scripte Range for each instrument is determined from the Low 64 & High 65 Octave values. In FIG. 7 , the octave range for the upright bass is two octaves; the octave range for the electric piano is five octaves.
  • FIG. 8 illustrates how the preliminary seed note pool 42 is expanded by application engine 1007 to fill an Instrument's specific Octave Range.
  • the Scripte Range 66 is calculated from the Low Scripte 64 and High Script 65 values.
  • Octave Range Offsets 67 are established for each Scripte contained in the Scripte Range 66 .
  • FIG. 9 illustrates how Offsets are dynamically applied by application engine 1007 to an Instrument's preliminary seed note pool to produce a playable sympathetic note pool in real time.
  • the Global Key Offset 24 is added to the Midi Note numbers 33 of the appropriate Scale Note template 31 which transposes them into Seed Notes 42 in the Detected Root Key 26 of the scanned song. (See FIG. 5 ).
  • Instrument Script Offsets 67 are added to the Midi Note numbers 72 of the Preliminary Seed Notes 70 . This transposes them to be within the Instrument's specified Octave Ranges 64 - 65 .
  • Octave Offsets 67 are used to generate Preliminary Seed Notes 71 for all octaves that fall within the range specified by the Low Scripte 64 and High Scripte 65 properties from the Instrument definition 61 .
  • the Global Region Offset 84 is set to the offset value contained in the marker 23 . This ensures that all triggered notes will be in the proper key for what the song is currently playing.
  • Each note in the note pool 86 has its own Index ID, 87 which is not affected by any of the Offset values. This index establishes a way to retrieve and play the same melodic pattern across all Regions of the song. See FIG. 10 .
  • FIG. 10 illustrates how Melody Patterns work.
  • Melody patterns establish a sequential list of notes (melodies) that can be retrieved from an Instrument's Sympathetic Note Pool 86 and played when a user triggers them.
  • Melody patterns are predefined within software executing in the application engine 1007 . They are required for playback and can be either selected or established as defaults for an Instrument.
  • the same Melody Pattern can be used across all chromatic regions in a song by dynamically adjusting the Regional Offset 84 .
  • Any Melody Pattern can be used with any Instrument.
  • Melody Patterns do not directly reference the pattern notes by their Key designation 86 . Instead, the notes are referenced by their Index ID 87 within the Instrument's Sympathetic Note Pool ( FIG. 9 ). This makes a Melody Pattern transparent to the Key or Octave to which it applies.
  • FIG. 10 shows an example of a 13 step Melody Pattern that will play “Mary Had a Little Lamb” melody in any Key and on any Instrument.
  • FIG. 11 illustrates the main processing flow 1100 executable by application engine 1007 configured to generate an interactive play-along song according to this disclosure.
  • the application engine 1007 receives a user selection of an audio file/song to import the audio file/song from either a music streaming service or from a personal music library, and the like.
  • the imported audio file/song must be in a standard audio format such as .wav or .mp3, or the like.
  • a preliminary Virtual Song map 20 is established showing the length of the overall song, its measure alignment grid, tempo setting markers, and where they occur in the audio file/song as it is played.
  • a link to the imported file 27 is saved as part of the song definitions in definition file 1010 .
  • the length of the imported song 28 is saved as part of the song definitions.
  • the imported audio file/song is software scanned by application engine 1007 to determine its beat points (tempo), and its beat/measure alignments (time signature).
  • the results of the scan are used by application engine 1007 to prepare a Virtual Song map 20 for the entire song by placing Tempo markers into the map at the appropriate locations along its length, as previously discussed, as shown in FIG. 3 .
  • the imported audio file is software analyzed by application engine 1007 to determine its overall key 26 .
  • the audio file is also scanned along its length to determine key changes and mark where they occur.
  • Regional Key indicator markers 22 are placed in the virtual Song Map 20 at the appropriate locations along its length.
  • a Global Key Offset 24 is established. This Global offset transposes all generic Scale Notes 32 from the generic key of C to Seed Notes 42 in the detected key. For example, if the detected key is F, the Global Key Offset 24 would be + 5 .
  • the user interface/display 1004 presents the user with a menu of possible play-along instruments 61 or instrument combinations.
  • the menu is organized in ways to simplify selections such as organizing the instrument choices by Genre.
  • an Instrument-specific Base Pool of notes 71 is established by application engine 1007 for each selected instrument.
  • each Instrument selection comes with a predetermined (default) method to trigger it (Trigger Type Pulse; One-shot etc).
  • Trigger Type Pulse Trigger Type Pulse; One-shot etc.
  • the user may accept or change the triggering characteristics.
  • each Instrument selection comes with a default selection of playback templates shown in FIG. 10 that the user may accept or change.
  • Playback templates are Melodic in nature and they dictate the melody sequence of sympathetic notes being played.
  • step 1109 upon completion, all song selections & properties are saved by the application engine 1007 in an XML format that can be read, edited, and played, such as by Beamz Interactive software offered by Beam Interactive, Inc. of Scottsdale Ariz.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Acoustics & Sound (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Auxiliary Devices For Music (AREA)
  • Electrophonic Musical Instruments (AREA)

Abstract

A method and system of using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like, that can later be played on an interactive music engine.

Description

    PRIORITY
  • This application is a continuation of U.S. patent application U.S. Ser. No. 16/036,352, now issued U.S. Pat. No. 10,534,811, filed Jan. 29, 2018 entitled ARTIFICIAL INTELLIGENCE METHODOLOGY TO AUTOMATICALLY GENERATE INTERACTIVE PLAY ALONG SONGS, the teachings of which are incorporated herein in their entirety.
  • TECHNICAL FIELD
  • The present disclosure relates to the use of Artificial Intelligence to automatically create and generate a multi instrument/sound/note interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like.
  • BACKGROUND
  • Interactive play-along songs can be played by a multi-media platform, such as available from Beamz Interactive, Inc. of Scottsdale Ariz. Such a platform is described in commonly assigned U.S. Pat. Nos. 7,504,577, 7,858,870, 9,542,919 and 9,646,588, the teachings of each which are incorporated herein by reference in their entirety.
  • The Beamz Cross Platform (i.e. IOS, Android, PC, MAC, AR, MR, VR, etc.) Interactive Music Engine is an award winning rich multi-media technology and platform configured to allow a user to selectively trigger various triggers each associated with a subset of a musical composition, to both compose and play sympathetic music. Different subsets/Midi files of a musical composition can be selectively assigned to the triggers. The platform includes triggers, and an electronic processor generating a control signal as a function of the triggers selected by a user. The electronic processor also generates the control signal as a function of a plurality of music programs. Each music program comprises sound elements comprising a subset of a musical composition, and the music programs are correlated to each other. The electronic processor generates audio signals indicative of audible musical sounds as a function of the control signal.
  • There is a need to use Artificial Intelligence to automatically create and generate a multi instrument/sound/note interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like, that can later be played on an interactive music engine.
  • SUMMARY
  • This disclosure comprises a method and system of using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like, that can later be played on an interactive music engine.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the disclosure and together with the description serve to explain the principles of the disclosure.
  • FIG. 1 illustrates a diagram of a system for multilayered media playback;
  • FIG. 2 illustrates the Standard Midi numbering scheme used by an application engine;
  • FIG. 3 illustrates a Virtual Song map prepared by the application engine;
  • FIG. 4 illustrates the Scale based templates that supply generic notes for further transpositions, and illustrates two possible Scale Note Templates;
  • FIG. 5 illustrates how a Global Key Offset transposes the preliminary Scale Note pool to become a Seed Note Pool in the scanned song (root) key;
  • FIG. 6 illustrates how Global Regional Offsets are used to dynamically transpose Seed Notes to accommodate key changes during song play;
  • FIG. 7 illustrates two selectable Instruments and their definitions;
  • FIG. 8 illustrates how the preliminary seed note pool is expanded by the application engine to fill an Instrument's specific Octave Range;
  • FIG. 9 illustrates how Offsets are dynamically applied by application engine 1007 to an Instrument's preliminary seed note pool to produce a playable sympathetic note pool in real time;
  • FIG. 10 shows an example of a 13 step Melody Pattern that will play “Mary Had a Little Lamb” melody in any Key and on any Instrument; and
  • FIG. 11 illustrates the main processing flow executable by the application engine configured to generate an interactive play-along song according to this disclosure.
  • DETAILED DESCRIPTION
  • Interactive Music Engine
  • This disclosure is directed to using Artificial Intelligence to automatically create and generate an interactive play-along song from a selected audio file/song imported from either a music streaming service, a personal music library and the like. The generated interactive play-along song can be played by a multi-media platform, such as available from Beamz Interactive, Inc. of Scottsdale Ariz. Such a platform is described in commonly assigned U.S. Pat. Nos. 7,504,577, 7,858,870, 9,542,919 and 9,646,588, the teachings of which are incorporated herein by reference in their entirety. These methods and illustrations show interactive music content being generated for direct use with the Beamz Interactive Music Engine. However, all song properties and generated note pools can be used by other music engines as long as they can accommodate Tempo and Key changes throughout a song; and they can use pools of sympathetic notes that are appropriate to the key changes as the song plays.
  • The Beamz Interactive Music Engine includes an electronic processor generating a control signal as a function of trigger objects selected by a user, the electronic processor also generating the control signal as a function of a plurality of music programs, wherein each said music program comprises sound elements comprising a subset of a musical composition, and the music programs are correlated to each other, the electronic processor generating an audio signals indicative of audible musical sounds as a function of the control signal.
  • FIG. 1 illustrates a diagram of a system 1000 for multilayered media playback.
  • The system 1000 can be implemented in electronic device 1001, and embodied as one of a computer, a smart phone, a tablet, a touchscreen computer, and the like having a display 1004.
  • Application engine 1007 is operable on an electronic processor 1006 and receives one or more inputs 1005 from the multiple trigger objects 1003 within the application environment 1001.
  • Application engine 1007 controls playback of media files 1011 that are combined to form a multilayered media file based on one or more of trigger inputs 1005, and definition file 1010 via sound engine 1008. The media files 1011 can be one or more MIDI files, samples such as .wav and .mp3 files, video files in a plurality of formats, and/or any other audio or video file format.
  • With a short trigger break, the trigger 1003 is held broken for a substantially short period of time, such as with a threshold for the short period of time of 0.5 seconds or less. The application engine 1007 can use the short break to trigger a one-shot, play a single note in a streamed sequence, or start and stop a loop.
  • With a long trigger break, the trigger 1003 is held broken for a longer period of time, such as with a threshold for the longer period of time of 0.5 seconds or more. Additional thresholds may be used for long trigger breaks with each threshold associated with a different action to be taken by the application engine 1007.
  • The application engine 1007 can use a long trigger break to Pulse (stream) notes.
  • Processor 1006 is configured such that visual outputs from application engine 1007 are displayed on display 1004 and output from sound engine 1008 is played on speaker 1009.
  • The combination of application engine 1007 and sound engine 1008 form an application on the processor 1006. The processor 1006 is configured to selectively associate a music programs with each of the plurality of trigger objects. The processor 1006 is configured such that when one of the trigger objects 1003 is in a first state for a prolonged period of time successive said audible musical sounds are generated, such that, for instance, the musical program associated with the trigger object continues to play uninterrupted, along with any other music programs that are playing in response the associated trigger object being triggered.
  • Display 1004 displays the application environment 1002.
  • When a trigger object 1003 is triggered by a user, trigger-specific visual output from application engine 1007 can be shown to alter the display properties or attributes of any element in the display 1004.
  • FIG. 2 illustrates the Standard Midi numbering scheme used by application engine 1007 according to the methods of this disclosure. All illustrations & examples herein are based on the Standard Midi Note Numbering scheme.
  • Midi is based on a chromatic music scale that breaks an octave into 12 notes (semitones) shown at 11. There are 128 possible Midi notes 11 spanning 10 octaves, shown at 12.
  • The lowest possible Midi note 11 is “C” in Octave #0 shown at 101, the highest Midi note 11 is “G” in Octave #10 shown at 102.
  • As illustrated in FIG. 2, each Midi note 11 is identified by its Key followed by the Octave 12 in which the Key falls. For example, Midi note C5 represents Middle “C” in Octave #5 shown at 103.
  • All Midi notes 11 are numbered at 13 sequentially from the lowest to the highest (0-127)
  • By using a combination of generic Offset Templates, this standard Midi note numbering scheme is used as a foundation for numerically generating play-along musical notes that will always be sympathetic or compatible with the analyzed audio file/song in real time.
  • FIG. 3 illustrates a Virtual Song map 20. This Virtual Song map 20 is a graphic (Music Staff) illustration that is prepared by application engine 1007 executing on processor 1006 from scanning an imported audio file/song file according to this disclosure.
  • When the song is played by application engine 1007 and song engine 1008, it follows this Virtual Song map 20, responding to the markers as they are encountered.
  • This example shows the data derived by application engine 1007 from performing Tempo & Chromatic analysis on a sample 10 bar (audio) audio file/song that was recorded in the key of C major at an approximate tempo of 120 BPM.
  • The application engine 1007 performs Tempo Analysis on the audio file/song to place Tempo value markers 21 in the Virtual Song map 20 to establish or change the tempo at that point in the audio file/song. When a Tempo value marker 21 is encountered during playback, the current Tempo is set to the value contained in the Tempo value marker 21.
  • The application engine 1007 performs Chromatic Analysis to add Regional Key markers 22 to the Virtual Song map 20 to indicate chromatic changes and their location within the audio file/song.
  • Each Regional Key marker 22 contains the chromatic global Region Key Offset 23 needed to transpose Seed Notes 42 from the detected Root Key 26 to Region specific Seed Notes 52 in the new key.
  • For example, if the audio file/song is detected to be in C, and the scan detects that, during play, the key changes to F, a Regional Key marker 22 is placed at that location in the Virtual Song map 20 with an offset value 23 of +5.
  • When a Regional Key marker 22 is encountered during playback of the audio file/song, the global Region Key Offset 23 is set to the value contained in the marker 22.
  • The application engine 1007 creates Song Sections (Chromatic Regions) shown at 28. Each chromatic region 28 becomes a Song Section (subset of song) in the song's definition as a whole and it marks a key change in the audio file/song.
  • In this example, there are 6 Song Sections 28—each having their own properties. Key 22, Regional (key) offset 23, and Length 25.
  • The combined lengths 25 of all the Song Sections 28 constitute the song's main loop length 29.
  • Use of Transposition Offsets
  • The following methods for generating sympathetic note pools are based on the use and application of multiple “Offsets”.
  • All generated play-along notes begin from preprogrammed scale notes which are generated from standard templates, in a standard key, within a standard octave.
  • The offsets described herein are used to transpose these notes to a different key as required by the audio file/song.
  • Offsets are expressed as semitone values—the number of semitones a given (root) note needs to be transposed into the target note.
  • For the sake of simplicity, all illustrations and methods described herein use the standard MIDI note numbering scheme to easily show how notes can be transposed by semitones.
  • However, these transposition offset methods apply to both MIDI notes and note samples as well.
  • With root notes in MIDI format, the offset amount is added to the MIDI note number for the root note to obtain the MIDI note number of the target note.
  • With sampled notes, audio transposition software is used to apply an offset value to transpose a sampled root note to a sample in the frequency of the target key.
  • FIG. 4 illustrates the Scale based templates that supply generic notes for further transpositions, and illustrates two possible Scale Note Templates.
  • After chromatic scanning is completed by application engine 1007, the scanned key data is used to select an appropriate seed note template 31 for the scanned audio file/song. A Global Root Key Offset 24 (FIG. 5) is established that, when applied, transposes the templated scale note pool 32 from the key of C to the detected key for the song.
  • All Note Templates are based on the standard Midi note numbering scheme (FIG. 2) and are prepared in the key of C with a Root Note of CO 101 (lowest possible Midi note).
  • Scale note templates supply the preliminary note pool 32 that is musically sympathetic to the overall scanned audio file/song. Their associated Midi note numbers 33 are used to mathematically transpose them to the proper key and octave as needed.
  • FIG. 5 illustrates how a Global Key Offset transposes the preliminary Scale Note pool 32 to become a Seed Note Pool 42 in the scanned song (root) key.
  • After chromatic scanning is complete, the scanned key data is used to determine a Global Root Key Offset 24 which is applied to transpose the preliminary sympathetic note pool 32 from the key of C to the detected key for the audio file/song.
  • In this illustration, the chromatic scanning determined the audio file/song to be in the key of G Major and the Global Root Key Offset 24 is set to +7. Adding this offset to the Midi note numbers from the C Major template 33 transposes the seed note pool to the key of G Major 42, 43.
  • FOR THE SAKE OF SIMPLICITY, ALL OTHER FIGURES ARE ILLUSTRATED WITH AN EXAMPLE SONG IN THE KEY OF C. (GLOBAL KEY OFFSET=0)
  • FIG. 6 illustrates how Global Regional Offsets 23 are used to dynamically transpose Seed Notes to accommodate key changes during song play.
  • As the audio file/song plays, and (chromatic) Song Section markers are encountered, the current Regional Offset 23 is set to the offset value contained in the marker 23 in real time. This ensures that all triggered notes will be in the proper key for what the song is currently playing.
  • In this example, Region #1 plays in the key of C major while Region #2 is transposed to the key of F major.
  • FIG. 7 illustrates two selectable Instruments and their definitions.
  • All predefined instrument choices contain information necessary to use that instrument in a musical composition. If the notes to be played are MIDI, each instrument 61 must be assigned to a Midi Synthesizer Patch 63 within a downloadable sound set 62. The ideal Octave Range for each instrument is determined from the Low 64 & High 65 Octave values. In FIG. 7, the octave range for the upright bass is two octaves; the octave range for the electric piano is five octaves.
  • FIG. 8 illustrates how the preliminary seed note pool 42 is expanded by application engine 1007 to fill an Instrument's specific Octave Range.
  • For each Instrument selected by the user:
  • The Octave Range 66 is calculated from the Low Octave 64 and High Octave 65 values.
  • Octave Range Offsets 67 are established for each Octave contained in the Octave Range 66.
  • For example, all Seed Note Pools 42 are always generated for Octave #0.
  • If an instrument sounds best in octave #5, the offset value (in chromatic semi-tones) would be +60 (5×12 semis).
  • Using the Octave Range Offsets 67, one set of preliminary Seed Notes 71 is generated for each octave contained in the Octave Range 66.
  • FIG. 9 illustrates how Offsets are dynamically applied by application engine 1007 to an Instrument's preliminary seed note pool to produce a playable sympathetic note pool in real time.
  • All of an Instrument's play-notes start out as standard Scale Notes contained in a Scale Note Template 31. The appropriate template is indicated by chromatic analysis of the imported song. Chromatic analysis also sets the Global Key Offset 24 (See FIG. 5).
  • The Global Key Offset 24 is added to the Midi Note numbers 33 of the appropriate Scale Note template 31 which transposes them into Seed Notes 42 in the Detected Root Key 26 of the scanned song. (See FIG. 5).
  • Instrument Octave Offsets 67 are added to the Midi Note numbers 72 of the Preliminary Seed Notes 70. This transposes them to be within the Instrument's specified Octave Ranges 64-65.
  • Octave Offsets 67 are used to generate Preliminary Seed Notes 71 for all octaves that fall within the range specified by the Low Octave 64 and High Octave 65 properties from the Instrument definition 61.
  • As the application engine 1007 plays the audio file/song in real time, and (chromatic) Song Section markers are encountered, the Global Region Offset 84 is set to the offset value contained in the marker 23. This ensures that all triggered notes will be in the proper key for what the song is currently playing.
  • Each note in the note pool 86 has its own Index ID, 87 which is not affected by any of the Offset values. This index establishes a way to retrieve and play the same melodic pattern across all Regions of the song. See FIG. 10.
  • FIG. 10 illustrates how Melody Patterns work. Melody patterns establish a sequential list of notes (melodies) that can be retrieved from an Instrument's Sympathetic Note Pool 86 and played when a user triggers them.
  • Melody patterns are predefined within software executing in the application engine 1007. They are required for playback and can be either selected or established as defaults for an Instrument.
  • Melody Patterns can have an unlimited number of steps.
  • The same Melody Pattern can be used across all chromatic regions in a song by dynamically adjusting the Regional Offset 84.
  • Any Melody Pattern can be used with any Instrument.
  • To provide flexibility across all keys, Melody Patterns do not directly reference the pattern notes by their Key designation 86. Instead, the notes are referenced by their Index ID 87 within the Instrument's Sympathetic Note Pool (FIG. 9). This makes a Melody Pattern transparent to the Key or Octave to which it applies.
  • FIG. 10 shows an example of a 13 step Melody Pattern that will play “Mary Had a Little Lamb” melody in any Key and on any Instrument.
  • FIG. 11 illustrates the main processing flow 1100 executable by application engine 1007 configured to generate an interactive play-along song according to this disclosure.
  • At step 1101 the application engine 1007 receives a user selection of an audio file/song to import the audio file/song from either a music streaming service or from a personal music library, and the like.
  • The imported audio file/song must be in a standard audio format such as .wav or .mp3, or the like. A preliminary Virtual Song map 20 is established showing the length of the overall song, its measure alignment grid, tempo setting markers, and where they occur in the audio file/song as it is played. A link to the imported file 27 is saved as part of the song definitions in definition file 1010. The length of the imported song 28 is saved as part of the song definitions.
  • At step 1102 the imported audio file/song is software scanned by application engine 1007 to determine its beat points (tempo), and its beat/measure alignments (time signature). The results of the scan are used by application engine 1007 to prepare a Virtual Song map 20 for the entire song by placing Tempo markers into the map at the appropriate locations along its length, as previously discussed, as shown in FIG. 3.
  • At step 1103 the imported audio file is software analyzed by application engine 1007 to determine its overall key 26.
  • The audio file is also scanned along its length to determine key changes and mark where they occur.
  • Regional Key indicator markers 22 are placed in the virtual Song Map 20 at the appropriate locations along its length.
  • All predesigned Key/Scale templates 31 contain a generic Scale Note pool 32 in the key of C in Octave #0.
  • Based on the detected key for the audio file/song, a Global Key Offset 24 is established. This Global offset transposes all generic Scale Notes 32 from the generic key of C to Seed Notes 42 in the detected key. For example, if the detected key is F, the Global Key Offset 24 would be +5.
  • At step 1104 a basic pool of Sympathetic Seed Notes for the song is established by the application engine 1007 applying the Global Key offset to the scale notes contained in the Generic Key/Scale template. (Template Note) 32+(Global Key Offset) 24=Seed Note Pool 42.
  • At step 1105 the user interface/display 1004 presents the user with a menu of possible play-along instruments 61 or instrument combinations. The menu is organized in ways to simplify selections such as organizing the instrument choices by Genre.
  • At step 1106 an Instrument-specific Base Pool of notes 71 is established by application engine 1007 for each selected instrument.
  • At step 1107 each Instrument selection comes with a predetermined (default) method to trigger it (Trigger Type Pulse; One-shot etc). Upon user selection of an Instrument, the user may accept or change the triggering characteristics.
  • At step 1108 each Instrument selection comes with a default selection of playback templates shown in FIG. 10 that the user may accept or change. Playback templates are Melodic in nature and they dictate the melody sequence of sympathetic notes being played.
  • At step 1109, upon completion, all song selections & properties are saved by the application engine 1007 in an XML format that can be read, edited, and played, such as by Beamz Interactive software offered by Beam Interactive, Inc. of Scottsdale Ariz.
  • The appended claims set forth novel and inventive aspects of the subject matter described above, but the claims may also encompass additional subject matter not specifically recited in detail. For example, certain features, elements, or aspects may be omitted from the claims if not necessary to distinguish the novel and inventive features from what is already known to a person having ordinary skill in the art. Features, elements, and aspects described herein may also be combined or replaced by alternative features serving the same, equivalent, or similar purpose without departing from the scope of the invention defined by the appended claims.

Claims (18)

1. A non-transitory computer readable medium including instructions operable by an electronic processor, comprising instructions for:
receiving an audio file from a remote source;
directly analyzing the audio file to map in audio form and create a virtual song definition configured for interactive playback; and
performing chromatic analysis across a length of the audio file to create a song section in the virtual song definition, where each song section contains a regional key offset that is configured to transpose seed notes from a detected root key to regional specific seed notes for the song section in real-time.
2. The non-transitory computer readable medium as specified in claim 1, further comprising instructions for performing tempo analysis on the length of the audio file to create a tempo region in the virtual song definition, where each tempo region contains a tempo value that establishes a playback tempo for the region.
3. The non-transitory computer readable medium as specified in claim 2, further comprising instructions such that when a tempo value marker is encountered during playback of the virtual song definition, a current tempo is set to a value contained in the tempo value marker.
4. The non-transitory computer readable medium as specified in claim 1, further including instructions to use pools of sympathetic notes that correspond to key changes as the audio file plays.
5. The non-transitory computer readable medium as specified in claim 4, further including instructions for associating an instrument to one said pool of sympathetic notes.
6. The non-transitory computer readable medium as specified in claim 5, further including instructions for associating a plurality of instruments to respective pools of sympathetic notes.
7. The non-transitory computer readable medium as specified in claim 5, further comprising instructions for assigning a melody template to each instrument's pool of sympathetic notes to govern how notes are retrieved and played.
8. The non-transitory computer readable medium as specified in claim 1, further comprising instructions for saving audio file definitions in a definition file.
9. The non-transitory computer readable medium as specified in claim 8, wherein definition file is an XML file.
10. A system, comprising;
a processor configured to:
receive an audio file from a remote source;
directly analyze the audio file to map the audio file and create a virtual song definition map configured for interactive playback; and
perform chromatic analysis across a length of the audio file to create song section objects in the virtual song definition, where each song section contains a regional key offset that is configured to transpose seed notes from a detected root key to regional specific seed notes for the song section in real-time.
11. The system as specified in claim 10, wherein the processor is configured to:
perform tempo analysis on the length of the audio file to create tempo region objects in the virtual song definition, where each tempo region contains a tempo value that establishes a playback tempo for the region; and
perform tempo analysis on the audio file to create the virtual song map and place tempo value markers in the virtual song map to establish or change the tempo at that point in the audio file.
12. The system as specified in claim 11, wherein the processor is configured such that when a tempo value marker is encountered during playback of the virtual song definition, a current tempo is set to a value contained in the tempo value marker.
13. The system as specified in claim 10, wherein the processor is configured to use pools of sympathetic notes that correspond to key changes as the audio file plays.
14. The system as specified in claim 13, wherein the processor is configured to associate an instrument to one said pool of sympathetic notes.
15. The system as specified in claim 14, wherein the processor is configured to associate a plurality of instruments to respective pools of sympathetic notes.
16. The system as specified in claim 14, wherein the processor is configured to assign a melody template to each instrument's pool of sympathetic notes to govern how notes are retrieved and played.
17. The system as specified in claim 10, wherein the processor is configured to save audio file definitions in a definition file.
18. The system as specified in claim 17, wherein definition file is an XML file.
US16/732,021 2018-01-29 2019-12-31 Artificial intelligence methodology to automatically generate interactive play along songs Abandoned US20200142926A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/732,021 US20200142926A1 (en) 2018-01-29 2019-12-31 Artificial intelligence methodology to automatically generate interactive play along songs

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201862623131P 2018-01-29 2018-01-29
US16/036,052 US10534811B2 (en) 2018-01-29 2018-07-16 Artificial intelligence methodology to automatically generate interactive play along songs
US16/732,021 US20200142926A1 (en) 2018-01-29 2019-12-31 Artificial intelligence methodology to automatically generate interactive play along songs

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/036,052 Continuation US10534811B2 (en) 2018-01-29 2018-07-16 Artificial intelligence methodology to automatically generate interactive play along songs

Publications (1)

Publication Number Publication Date
US20200142926A1 true US20200142926A1 (en) 2020-05-07

Family

ID=67392149

Family Applications (2)

Application Number Title Priority Date Filing Date
US16/036,052 Expired - Fee Related US10534811B2 (en) 2018-01-29 2018-07-16 Artificial intelligence methodology to automatically generate interactive play along songs
US16/732,021 Abandoned US20200142926A1 (en) 2018-01-29 2019-12-31 Artificial intelligence methodology to automatically generate interactive play along songs

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US16/036,052 Expired - Fee Related US10534811B2 (en) 2018-01-29 2018-07-16 Artificial intelligence methodology to automatically generate interactive play along songs

Country Status (1)

Country Link
US (2) US10534811B2 (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110958351A (en) * 2019-11-26 2020-04-03 东莞市鼎酷科技有限公司 Message prompt tone processing method and device, storage medium and terminal
WO2021202760A1 (en) * 2020-03-31 2021-10-07 Aries Adaptive Media, LLC Processes and systems for mixing audio tracks according to a template
CN111930289B (en) * 2020-09-09 2021-05-07 智者四海(北京)技术有限公司 Method and system for processing pictures and texts
CN116764240A (en) * 2022-03-10 2023-09-19 腾讯科技(深圳)有限公司 Performance recording method, device, equipment and storage medium in virtual scene

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20010039872A1 (en) * 2000-05-11 2001-11-15 Cliff David Trevor Automatic compilation of songs
US20030094093A1 (en) * 2001-05-04 2003-05-22 David Smith Music performance system
US20080271592A1 (en) * 2003-08-20 2008-11-06 David Joseph Beckford System, computer program and method for quantifying and analyzing musical intellectual property
US20090048694A1 (en) * 2005-07-01 2009-02-19 Pioneer Corporation Computer program, information reproduction device, and method
US7667125B2 (en) * 2007-02-01 2010-02-23 Museami, Inc. Music transcription
US7956276B2 (en) * 2006-12-04 2011-06-07 Sony Corporation Method of distributing mashup data, mashup method, server apparatus for mashup data, and mashup apparatus
US8115090B2 (en) * 2006-11-28 2012-02-14 Sony Corporation Mashup data file, mashup apparatus, and content creation method
US20140018947A1 (en) * 2012-07-16 2014-01-16 SongFlutter, Inc. System and Method for Combining Two or More Songs in a Queue
US9111519B1 (en) * 2011-10-26 2015-08-18 Mixwolf LLC System and method for generating cuepoints for mixing song data

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20010039872A1 (en) * 2000-05-11 2001-11-15 Cliff David Trevor Automatic compilation of songs
US20030094093A1 (en) * 2001-05-04 2003-05-22 David Smith Music performance system
US20080271592A1 (en) * 2003-08-20 2008-11-06 David Joseph Beckford System, computer program and method for quantifying and analyzing musical intellectual property
US20090048694A1 (en) * 2005-07-01 2009-02-19 Pioneer Corporation Computer program, information reproduction device, and method
US8115090B2 (en) * 2006-11-28 2012-02-14 Sony Corporation Mashup data file, mashup apparatus, and content creation method
US7956276B2 (en) * 2006-12-04 2011-06-07 Sony Corporation Method of distributing mashup data, mashup method, server apparatus for mashup data, and mashup apparatus
US7667125B2 (en) * 2007-02-01 2010-02-23 Museami, Inc. Music transcription
US9111519B1 (en) * 2011-10-26 2015-08-18 Mixwolf LLC System and method for generating cuepoints for mixing song data
US20140018947A1 (en) * 2012-07-16 2014-01-16 SongFlutter, Inc. System and Method for Combining Two or More Songs in a Queue

Also Published As

Publication number Publication date
US10534811B2 (en) 2020-01-14
US20190236209A1 (en) 2019-08-01

Similar Documents

Publication Publication Date Title
US20200142926A1 (en) Artificial intelligence methodology to automatically generate interactive play along songs
US7743092B2 (en) Method for recognizing and distributing music
JP4111004B2 (en) Performance practice device and performance practice program
US8035020B2 (en) Collaborative music creation
US7601904B2 (en) Interactive tool and appertaining method for creating a graphical music display
US6924425B2 (en) Method and apparatus for storing a multipart audio performance with interactive playback
ES2539813T3 (en) Music transcription
US7741554B2 (en) Apparatus and method for automatically creating music piece data
US7572968B2 (en) Electronic musical instrument
AU784788B2 (en) Array or equipment for composing
WO2006112584A1 (en) Music composing device
US20020144587A1 (en) Virtual music system
JP5229998B2 (en) Code name detection device and code name detection program
KR100664677B1 (en) Method for generating music contents using handheld terminal
US20020144588A1 (en) Multimedia data file
JP4274272B2 (en) Arpeggio performance device
US6046396A (en) Stringed musical instrument performance information composing apparatus and method
JP3915807B2 (en) Automatic performance determination device and program
JP2016142967A (en) Accompaniment training apparatus and accompaniment training program
JP4219652B2 (en) A singing practice support system for a karaoke device that controls the main melody volume at the relevant location based on the pitch error measured immediately before repeat performance
JP7419830B2 (en) Accompaniment sound generation device, electronic musical instrument, accompaniment sound generation method, and accompaniment sound generation program
JP3879524B2 (en) Waveform generation method, performance data processing method, and waveform selection device
WO2022172732A1 (en) Information processing system, electronic musical instrument, information processing method, and machine learning system
JP2004093601A (en) Singing practice support system for karaoke device
JP2018146716A (en) Training device, training program, and training method

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

AS Assignment

Owner name: TOPDOWN LICENSING LLC, TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BEAMZ IP, LLC;REEL/FRAME:053029/0492

Effective date: 20200528

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION