EP0944033B1 - Melody retrieval system and method - Google Patents

Melody retrieval system and method Download PDF

Info

Publication number
EP0944033B1
EP0944033B1 EP99302171A EP99302171A EP0944033B1 EP 0944033 B1 EP0944033 B1 EP 0944033B1 EP 99302171 A EP99302171 A EP 99302171A EP 99302171 A EP99302171 A EP 99302171A EP 0944033 B1 EP0944033 B1 EP 0944033B1
Authority
EP
European Patent Office
Prior art keywords
span
pitch
relative
values
music
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP99302171A
Other languages
German (de)
French (fr)
Other versions
EP0944033A1 (en
Inventor
Tomonari Sonoda
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP10111273A external-priority patent/JPH11272274A/en
Priority claimed from JP10378084A external-priority patent/JP2000187671A/en
Application filed by Individual filed Critical Individual
Publication of EP0944033A1 publication Critical patent/EP0944033A1/en
Application granted granted Critical
Publication of EP0944033B1 publication Critical patent/EP0944033B1/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0033Recording/reproducing or transmission of music for electrophonic musical instruments
    • G10H1/0041Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/632Query formulation
    • G06F16/634Query by example, e.g. query by humming
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/121Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods
    • G10H2240/131Library retrieval, i.e. searching a database or selecting a specific musical piece, segment, pattern, rule or parameter set
    • G10H2240/141Library retrieval matching, i.e. any of the steps of matching an inputted segment or phrase with musical database contents, e.g. query by humming, singing or playing; the steps may include, e.g. musical analysis of the input, musical feature extraction, query formulation, or details of the retrieval process
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/171Transmission of musical instrument data, control or status information; Transmission, remote access or control of music data for electrophonic musical instruments
    • G10H2240/201Physical layer or hardware aspects of transmission to or from an electrophonic musical instrument, e.g. voltage levels, bit streams, code words or symbols over a physical link connecting network nodes or instruments
    • G10H2240/241Telephone transmission, i.e. using twisted pair telephone lines or any type of telephone network
    • G10H2240/251Mobile telephone transmission, i.e. transmitting, accessing or controlling music data wirelessly via a wireless or mobile telephone receiver, analogue or digital, e.g. DECT, GSM, UMTS

Definitions

  • the present invention relates to a melody retrieval system and method, and more particularly to a WWW-based melody retrieval system with thresholds determined by using distribution of pitch and span of notes from a music database.
  • the input melody can include various errors which are caused by the uncertainty of the user's memory or by the user's singing ability. Those errors are usually tolerated by using several thresholds and converting the input melody into approximate relative-value sequences of pitch change (the pitch difference between adjacent notes) and of span change (the inter-onset interval ratio of adjacent notes).
  • an approximate relative-pitch value is expressed by using three characters ⁇ U(p), D(own), and E(qual) ⁇ , and the sequence "do-re-mi-mi-re-do” is converted into "X-U-U-E-D-D-" (wherein "X” indicates the first note which does not have a relative value). These approximate sequences are then utilized in the matching process which compares the input with all songs in a database.
  • the pitch information is mainly used as a search clue while the span information is not used effectively, and it is difficult to improve the matching accuracy by using only the pitch information.
  • the invention may provide:
  • a media retrieval system which can retrieve media with music by melody.
  • FIG. 3 a first embodiment of a melody retrieval system.
  • sequences of relative-pitch difference and relative-span ratio are first obtained from sequences of pitch and span values of each piece of music in a database (S101), and a histogram of the relative-pitch values and a histogram of the relative-span values of the database are made by using the distribution of all notes in the database (S102).
  • the relative-pitch value is the normalized pitch difference between the adjacent notes in which a semitone difference is normalized to the value of 100.
  • the relative-span value is expressed by percentage.
  • the total frequency of values of a histogram of pitch is expressed by Sum1
  • that of a histogram of span is expressed by Sum2, respectively.
  • the number of categories for the histogram of relative-pitch values is expressed by Category_Num1
  • that of relative-span values is expressed by Category_Num2 respectively.
  • the relative-pitch difference and relative-span ratio are respectively converted into the approximate relative-pitch values and relative-span values by categorizing those values in several parts like U, E, D in FIG. 1 or S, E, L in FIG. 2 which are separated by the thresholds previously calculated (S104).
  • Each number of category is 3 in FIGS. 1 and 2.
  • a song is inputted by a user singing, humming, whistling or similar into a microphone 10 as preparation for retrieval (S105).
  • the song is subjected to an A/D conversion (S106), and the A/D converted signal is written into a memory 16, through which the A/D converted signal is processed by a processor 14 in accordance with a program of the memory 16 to detect a voiced sound from the A/D conversion signal (S107).
  • a fundamental frequency of each frame is defined from the detected voiced sound (S108).
  • An onset time of the voiced sound is divided off as an onset time of each note, a time difference (number of frames) with an onset time of the next note is determined as the span of the note, and the maximum value among the fundamental frequencies of each note contained during its span is defined as the highest pitch values (S109).
  • the relative pitch and span values (S110) of each note in the input song are calculated from the obtained pitch and span from the previous note.
  • the relative values are then transmitted to a song retrieval system A, and this transmittal may take place over the WWW network.
  • the relative-pitch values and relative-span values of the user's input melody are respectively converted into approximate relative-pitch values and approximate relative-span values by using the same thresholds of (S104) and categorizing relative-pitch values and relative-span values in several parts like U, E, D in FIG. 1 or S, E, L in FIG. 2 (Sill).
  • S104 the same thresholds of
  • S112 a distance between the pitch and span of each melody in the music database and the input is calculated
  • the titles of pieces of music which are sorted by the order of short distance previously calculated, are shown on a display 20 of an input device B (S113).
  • the input device B comprises a microphone 10, an A/D converter 12, a processor 14, a memory 16, an input and output unit 18, a display 20, a speaker 22 and a D/A converter 24.
  • the input and output unit 18 may be connected to a WWW network (not shown).
  • Received from the song retrieval system A is the retrieved result, which is shown in the display 20, and the desired melody among the retrieved songs can be heard from the speaker 22.
  • the title of the desired song can be easily and accurately retrieved by inputting the melody by singing into the microphone 10.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Mathematical Physics (AREA)
  • Acoustics & Sound (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Auxiliary Devices For Music (AREA)
  • Electrophonic Musical Instruments (AREA)

Description

The present invention relates to a melody retrieval system and method, and more particularly to a WWW-based melody retrieval system with thresholds determined by using distribution of pitch and span of notes from a music database.
One of the problems in building a melody retrieval system, where a melody is input by a user singing, is that the input melody can include various errors which are caused by the uncertainty of the user's memory or by the user's singing ability. Those errors are usually tolerated by using several thresholds and converting the input melody into approximate relative-value sequences of pitch change (the pitch difference between adjacent notes) and of span change (the inter-onset interval ratio of adjacent notes).
For instance, an approximate relative-pitch value is expressed by using three characters {U(p), D(own), and E(qual)}, and the sequence "do-re-mi-mi-re-do" is converted into "X-U-U-E-D-D-" (wherein "X" indicates the first note which does not have a relative value). These approximate sequences are then utilized in the matching process which compares the input with all songs in a database.
Prior melody retrieval systems [Kageyama et al. 1993; Ghias et al. 1995] utilized static heuristic thresholds to obtain these approximate sequences. Since it is difficult to determine the optimum thresholds for all songs, those systems did not take advantage of the span information which can potentially be useful, and mainly used pitch information as a search clue.
It was therefore difficult to improve the matching accuracy by using only the pitch information.
It was also difficult to reduce the number of answer candidates, because several songs in a large database tended to have the same patterns of the approximate relative-values of pitch and span sequences.
Moreover, the public use of music databases over computer networks such as the World Wide Web (WWW) has not been considered.
In conventional music retrieval systems, the pitch information is mainly used as a search clue while the span information is not used effectively, and it is difficult to improve the matching accuracy by using only the pitch information.
According to a first aspect of the invention, there is provided a melody retrieval method as claimed in claim 1.
According to a second aspect of the invention, there is provided a melody retrieval system as claimed in claim 4.
In its preferred aspects, the invention may provide:
  • a WWW-based melody retrieval system on which a user can input a melody by singing;
  • a WWW-based melody retrieval system with thresholds determined by using distribution of pitch and span of notes from a music database accurately;
  • a WWW-based melody retrieval system with thresholds determined by using distribution of pitch and span of notes extracted from a music database whereby an accurate song's title can be obtained quickly;
  • a WWW-based melody retrieval system with thresholds determined by using distribution of pitch and span of notes from a music database whereby an approximate relative-value sequences of pitch and span with the maximum quantity of information can be obtained dynamically from music database and user's input melody;
  • a WWW-based melody retrieval system which can convert relative-pitch or relative-span information of notes into approximate relative-values by categorisation whereby the approximate values of each category appear with equal frequency in a database;
  • a WWW-based melody retrieval system which can determine arbitrary number of categories for converting relative-pitch or relative-span information of notes into approximate relative-values;
  • a WWW-based melody retrieval system using the coarse-to-fine matching whereby the number of answer candidates can be reduced by considering the trade-off between coarse matching and fine matching;
  • a WWW-based melody retrieval system which utilizes a sung melody as a query and is reliable so that it is robust enough to deal with anonymous users's inputs;
  • a WWW-based melody retrieval device which can be mounted either in a desired position or article whereby a melody retrieval can be carried out easily from the remotely located music databases over a WWW network;
  • a WWW-based melody retrieval device which can be set in a car whereby a car driver can retrieve music by singing, thus enabling the car driver to keep safe driving since there is no need for him to use his hands for retrieving or selecting music;
  • a WWW-based melody retrieval device which can be set in a karaoke player whereby a user can retrieve songs by singing;
  • a WWW-based melody retrieval device which can be set in a record shop whereby a user can retrieve music by singing over the network even if the record shop does not have the user's desired music CDs;
  • a WWW-based melody retrieval device which can be set in either a cellular phone or a portable device whereby a user can retrieve music anywhere;
  • a WWW-based melody retrieval device on which a plurality of users can input their desired melodies by singing at the same time; and
  • a device for retrieving the song's title from a music database over a WWW network whereby the matching accuracy of a music retrieval is high enough for melody retrieval services over a WWW network.
  • In a further preferred aspect of the invention, there is provided a media retrieval system which can retrieve media with music by melody.
    A preferred embodiment of the invention will now be described by way of example only and with reference to the accompanying drawings, in which:
  • FIG. 1 is a histogram of the relative-pitch values in a melody retrieval system with thresholds determined by using distribution of relative-pitch values of notes extracted from a database, wherein the distributed three categories of U(p), E(qual) and D(own) for pitch determining the thresholds are illustrated;
  • FIG. 2 is a histogram of the relative-span values in the melody retrieval system with thresholds determined by using distribution of relative-span values of notes extracted from a database, wherein the distributed three categories of L(onger), E(qual) and S(horter) for span determining the thresholds are illustrated;
  • FIG. 3 is a flow chart illustrating the operation of the melody retrieval system, and
  • FIG. 4 is a melody retrieval terminal device which may be used with the melody retrieval system.
  • Referring now to the accompanying drawings, in which like numerals designate like parts throughout the several views thereof, there is shown in FIG. 3, a first embodiment of a melody retrieval system.
    In preparing for retrieval, sequences of relative-pitch difference and relative-span ratio are first obtained from sequences of pitch and span values of each piece of music in a database (S101), and a histogram of the relative-pitch values and a histogram of the relative-span values of the database are made by using the distribution of all notes in the database (S102).
    In FIG. 1, the relative-pitch value is the normalized pitch difference between the adjacent notes in which a semitone difference is normalized to the value of 100.
    In FIG. 2, the relative-span value is expressed by percentage.
    In making these histograms, the total frequency of values of a histogram of pitch is expressed by Sum1, and that of a histogram of span is expressed by Sum2, respectively. The number of categories for the histogram of relative-pitch values is expressed by Category_Num1, and that of relative-span values is expressed by Category_Num2 respectively.
    In this case, thresholds for relative-pitch values are determined so that each category, which will be divided by the thresholds, equally has M1 (=Sum1 / Category_Num1) values in it. Thresholds for relative-span values are similarly determined so that each category equally has M2 (=Sum2 / Category_Num2) values in it (S103). Subsequently the relative-pitch difference and relative-span ratio are respectively converted into the approximate relative-pitch values and relative-span values by categorizing those values in several parts like U, E, D in FIG. 1 or S, E, L in FIG. 2 which are separated by the thresholds previously calculated (S104).
    Each number of category is 3 in FIGS. 1 and 2.
    A song is inputted by a user singing, humming, whistling or similar into a microphone 10 as preparation for retrieval (S105). The song is subjected to an A/D conversion (S106), and the A/D converted signal is written into a memory 16, through which the A/D converted signal is processed by a processor 14 in accordance with a program of the memory 16 to detect a voiced sound from the A/D conversion signal (S107). A fundamental frequency of each frame is defined from the detected voiced sound (S108).
    An onset time of the voiced sound is divided off as an onset time of each note, a time difference (number of frames) with an onset time of the next note is determined as the span of the note, and the maximum value among the fundamental frequencies of each note contained during its span is defined as the highest pitch values (S109).
    The relative pitch and span values (S110) of each note in the input song are calculated from the obtained pitch and span from the previous note. The relative values are then transmitted to a song retrieval system A, and this transmittal may take place over the WWW network.
    As shown in FIGS. 3 and 4, the relative-pitch values and relative-span values of the user's input melody are respectively converted into approximate relative-pitch values and approximate relative-span values by using the same thresholds of (S104) and categorizing relative-pitch values and relative-span values in several parts like U, E, D in FIG. 1 or S, E, L in FIG. 2 (Sill). These approximate values are compared with the value of each melody in the music database, a distance between the pitch and span of each melody in the music database and the input is calculated (S112), and the titles of pieces of music, which are sorted by the order of short distance previously calculated, are shown on a display 20 of an input device B (S113).
    As particularly shown in FIG. 4, the input device B comprises a microphone 10, an A/D converter 12, a processor 14, a memory 16, an input and output unit 18, a display 20, a speaker 22 and a D/A converter 24.
    It can be understood from FIG. 4 that the input and output unit 18 may be connected to a WWW network (not shown).
    Received from the song retrieval system A is the retrieved result, which is shown in the display 20, and the desired melody among the retrieved songs can be heard from the speaker 22.
    As described in the foregoing paragraphs, the title of the desired song can be easily and accurately retrieved by inputting the melody by singing into the microphone 10.

    Claims (10)

    1. A melody retrieval method, comprising the steps of:
      obtaining sequences of relative-pitch difference and relative-span ratio from sequences of pitch and span values of each piece of music in a database;
      making a histogram of the relative-pitch values of the database by using the distribution of all notes in the entire database;
      making a histogram of the relative-span values of the database by using the distribution of all notes in the entire database;
      determining thresholds for relative pitch so that each category, which will be divided by the thresholds, equally having M1 (=Sum1 / Category_Num1) values in it, and thresholds for relative-span values are similarly determined so that each category equally has M2 (=Sum2 / Category_Num2) values in it;
      converting the relative-pitch difference and relative-span ratio into the approximate relative-pitch values and approximate relative-span values respectively;
      inputting a song by humming, singing or similar into a microphone as preparation for retrieval;
      subjecting the inputted song to an A/D conversion;
      detecting a voiced sound from a sound signal after A/D conversion;
      defining a fundamental frequency of each frame from the detected voiced sound;
      dividing off an onset time of the voiced sound as an onset time of each note;
      determining a time difference (number of frames) as the span of the note;
      defining the maximum value among the fundamental frequencies of each note contained during its span as the highest pitch value;
      calculating the relative pitch and span values from the obtained pitch and span from the previous sound;
      converting the relative pitch and span information of the user's input melody into approximate relative-pitch values and approximate relative-span values by using the thresholds previously obtained;
      comparing the approximate relative-pitch values and approximate relative-span values with the approximate relative-pitch values and approximate relative-span values of each melody in the music database;
      calculating a distance between the pitch and span of each melody in the music database and the input;
      retrieving at least one piece of music based on the calculated distances; and
      presenting information about the retrieved music.
    2. A melody retrieval method as claimed in claim 1, wherein the steps concerning the database and the comparison of the coarse values are carried out at a first location, and said steps concerning the processing of the input song take place at a second location, the relative pitch and span values of the input song being transmitted from said second location to said first location, and the information about the retrieved music being transmitted from said first location to said second location, preferably via the World Wide Web.
    3. A melody retrieval method as claimed in any preceding claim, further including a step of:
      retrieving a piece of media with music by melody retrieval.
    4. A melody retrieval system, comprising:
      a database of pieces of music;
      first means for obtaining sequences of relative-pitch difference and relative-span ratio from sequences of pitch and span values of each piece of music, making a histogram of the relative-pitch values of the database by using the distribution of all notes in the entire database, making a histogram of the relative-span values of the database by using the distribution of all notes in the entire database, determining thresholds for relative pitch so that each category, which will be divided by the thresholds, equally having M1 (=Sum1 / Category_Num1) values in it, and thresholds for relative-span values are similarly determined so that each category equally has M2 (=Sum2 / Category_Num2) values in it, and converting the relative-pitch difference and relative-span ratio into the approximate relative-pitch values and approximate relative-span values respectively;
      input means for inputting a song;
      processing means for subjecting the inputted song to an A/D conversion, detecting a voiced sound from a sound signal after A/D conversion, defining a fundamental frequency of each frame from the detected voiced sound, dividing off an onset time of the voiced sound as an onset time of each note, determining a time difference (number of frames) as the span of the note, defining the maximum value among the fundamental frequencies of each note contained during its span as the highest pitch value, and calculating the relative pitch and span values from the obtained pitch and span from the previous sound;
      second means for converting the relative pitch and span information of the user's input melody into approximate relative-pitch values and approximate relative-span values by using the thresholds previously obtained, comparing the approximate relative-pitch values and approximate relative-span values with the approximate relative-pitch values and approximate relative-span values of each melody in the music database, calculating a distance between the pitch and span of each melody in the music database and the input, and retrieving at least one piece of music based on the calculated distances; and
      presentation means for presenting information about the retrieved music.
    5. A melody retrieval system as claimed in claim 4, wherein said first means and said processing means are in different locations, first means and said processing means being connected to allow transmission of information therebetween, preferably via the World Wide Web.
    6. A melody retrieval system as claimed in claim 5 or claim 6, wherein said input means, said processing means and said presentation means are provided in an input device comprising a microphone, an A/D converter, an operating unit (operator) for treating a digital signal, an input and output unit for either transmitting or receiving the digital signal over the WWW network, a memory storing a program, a display showing the retrieved title of the database songs, a D/A converter, and an output unit for playing the retrieved song.
    7. Computer software on a carrier allowing any of said first means, said processing means and said second means in claim 4 to carry out the actions specified in that claim.
    8. A melody retrieval method comprising the steps of:
      analysing each piece of music in a database of pieces of music and categorizing each piece of music according to the pitch and span of its notes;
      analysing an input piece of music and categorizing it according to the pitch and span of its notes;
      comparing the categorized input piece of music to the categorized pieces of music in the database; and
      retrieving at least one piece of music from the database based on the results of the comparison.
    9. A melody retrieval system comprising;
      a database of pieces of music;
      first means for analysing each piece of music in a database of pieces of music and categorizing each piece of music according to the pitch and span of its notes;
      second means for analysing an input piece of music and categorizing it according to the pitch and span of its notes;
      third means for comparing the categorized input piece of music to the categorized pieces of music in the database; and
      fourth means for retrieving at least one piece of music from the database based on the results of the comparison.
    10. Computer software on a carrier, allowing any of the said means in claim 9 to carry out the actions specified in that claim.
    EP99302171A 1998-03-19 1999-03-19 Melody retrieval system and method Expired - Lifetime EP0944033B1 (en)

    Applications Claiming Priority (4)

    Application Number Priority Date Filing Date Title
    JP11127398 1998-03-19
    JP10111273A JPH11272274A (en) 1998-03-19 1998-03-19 Method for retrieving piece of music by use of singing voice
    JP37808498 1998-12-21
    JP10378084A JP2000187671A (en) 1998-12-21 1998-12-21 Music retrieval system with singing voice using network and singing voice input terminal equipment to be used at the time of retrieval

    Publications (2)

    Publication Number Publication Date
    EP0944033A1 EP0944033A1 (en) 1999-09-22
    EP0944033B1 true EP0944033B1 (en) 2003-05-28

    Family

    ID=26450702

    Family Applications (1)

    Application Number Title Priority Date Filing Date
    EP99302171A Expired - Lifetime EP0944033B1 (en) 1998-03-19 1999-03-19 Melody retrieval system and method

    Country Status (3)

    Country Link
    US (1) US6121530A (en)
    EP (1) EP0944033B1 (en)
    DE (1) DE69908226T2 (en)

    Families Citing this family (111)

    * Cited by examiner, † Cited by third party
    Publication number Priority date Publication date Assignee Title
    US6345104B1 (en) 1994-03-17 2002-02-05 Digimarc Corporation Digital watermarks and methods for security documents
    US7313251B2 (en) * 1993-11-18 2007-12-25 Digimarc Corporation Method and system for managing and controlling electronic media
    US8094949B1 (en) 1994-10-21 2012-01-10 Digimarc Corporation Music methods and systems
    US6560349B1 (en) 1994-10-21 2003-05-06 Digimarc Corporation Audio monitoring using steganographic information
    US7224819B2 (en) 1995-05-08 2007-05-29 Digimarc Corporation Integrating digital watermarks in multimedia content
    US6760463B2 (en) * 1995-05-08 2004-07-06 Digimarc Corporation Watermarking methods and media
    US7805500B2 (en) 1995-05-08 2010-09-28 Digimarc Corporation Network linking methods and apparatus
    US8429205B2 (en) 1995-07-27 2013-04-23 Digimarc Corporation Associating data with media signals in media signal systems through auxiliary data steganographically embedded in the media signals
    US7562392B1 (en) 1999-05-19 2009-07-14 Digimarc Corporation Methods of interacting with audio and ambient music
    US6411725B1 (en) 1995-07-27 2002-06-25 Digimarc Corporation Watermark enabled video objects
    US7711564B2 (en) 1995-07-27 2010-05-04 Digimarc Corporation Connected audio and other media objects
    US6829368B2 (en) 2000-01-26 2004-12-07 Digimarc Corporation Establishing and interacting with on-line media collections using identifiers in media signals
    US6505160B1 (en) * 1995-07-27 2003-01-07 Digimarc Corporation Connected audio and other media objects
    US6965682B1 (en) * 1999-05-19 2005-11-15 Digimarc Corp Data transmission by watermark proxy
    US8180844B1 (en) 2000-03-18 2012-05-15 Digimarc Corporation System for linking from objects to remote resources
    US7505605B2 (en) 1996-04-25 2009-03-17 Digimarc Corporation Portable devices and methods employing digital watermarking
    US20030056103A1 (en) 2000-12-18 2003-03-20 Levy Kenneth L. Audio/video commerce application architectural framework
    US7930546B2 (en) * 1996-05-16 2011-04-19 Digimarc Corporation Methods, systems, and sub-combinations useful in media identification
    US7689532B1 (en) 2000-07-20 2010-03-30 Digimarc Corporation Using embedded data with file sharing
    US8332478B2 (en) 1998-10-01 2012-12-11 Digimarc Corporation Context sensitive connected content
    US20070100757A1 (en) * 1999-05-19 2007-05-03 Rhoads Geoffrey B Content Protection Arrangements
    US7185201B2 (en) * 1999-05-19 2007-02-27 Digimarc Corporation Content identifiers triggering corresponding responses
    US7760905B2 (en) * 1999-06-29 2010-07-20 Digimarc Corporation Wireless mobile phone with content processing
    US7406214B2 (en) * 1999-05-19 2008-07-29 Digimarc Corporation Methods and devices employing optical sensors and/or steganography
    US20070055884A1 (en) 1999-05-19 2007-03-08 Rhoads Geoffrey B User control and activation of watermark enabled objects
    US7565294B2 (en) 1999-05-19 2009-07-21 Digimarc Corporation Methods and systems employing digital content
    US8095796B2 (en) 1999-05-19 2012-01-10 Digimarc Corporation Content identifiers
    US8752118B1 (en) 1999-05-19 2014-06-10 Digimarc Corporation Audio and video content-based methods
    US20020032734A1 (en) 2000-07-26 2002-03-14 Rhoads Geoffrey B. Collateral data combined with user characteristics to select web site
    US6355869B1 (en) * 1999-08-19 2002-03-12 Duane Mitton Method and system for creating musical scores from musical recordings
    US6868394B1 (en) * 1999-09-21 2005-03-15 Daniel Mele Method and apparatus for simplified artist-buyer transactions
    US7194752B1 (en) 1999-10-19 2007-03-20 Iceberg Industries, Llc Method and apparatus for automatically recognizing input audio and/or video streams
    DE19948974A1 (en) * 1999-10-11 2001-04-12 Nokia Mobile Phones Ltd Method for recognizing and selecting a tone sequence, in particular a piece of music
    US6188010B1 (en) * 1999-10-29 2001-02-13 Sony Corporation Music search by melody input
    US7224995B2 (en) * 1999-11-03 2007-05-29 Digimarc Corporation Data entry method and system
    US7257536B1 (en) * 1999-11-23 2007-08-14 Radiant Systems, Inc. Audio request interaction system
    US6678680B1 (en) * 2000-01-06 2004-01-13 Mark Woo Music search engine
    DE60144222D1 (en) 2000-01-13 2011-04-28 Digimarc Corp AUTHENTICATIVE METADATA AND EMBEDDING OF METADATA IN WATERMARK OF MEDIA SIGNALS
    US7444353B1 (en) 2000-01-31 2008-10-28 Chen Alexander C Apparatus for delivering music and information
    US20020073098A1 (en) * 2000-02-28 2002-06-13 Lei Zhang Methodology and system for searching music over computer network and the internet based on melody and rhythm input
    DE10011297C2 (en) * 2000-03-08 2002-03-07 Ingolf Ruge Procedure for creating and transferring a request to a database
    WO2001069575A1 (en) * 2000-03-13 2001-09-20 Perception Digital Technology (Bvi) Limited Melody retrieval system
    AU2001251347A1 (en) * 2000-04-05 2001-10-23 Aei Music Network, Inc. Expert system for play list generation
    US8121843B2 (en) 2000-05-02 2012-02-21 Digimarc Corporation Fingerprint methods and systems for media signals
    US6970886B1 (en) * 2000-05-25 2005-11-29 Digimarc Corporation Consumer driven methods for associating content indentifiers with related web addresses
    US7853664B1 (en) * 2000-07-31 2010-12-14 Landmark Digital Services Llc Method and system for purchasing pre-recorded music
    DE60140135D1 (en) * 2000-08-23 2009-11-19 Gracenote Inc METHOD FOR IMPROVED PLAYBACK OF INFORMATION, CLIENT SYSTEM AND SERVER SYSTEM
    US8205237B2 (en) 2000-09-14 2012-06-19 Cox Ingemar J Identifying works, using a sub-linear time search, such as an approximate nearest neighbor search, for initiating a work-based action, such as an action on the internet
    FI20002161A (en) 2000-09-29 2002-03-30 Nokia Mobile Phones Ltd Method and system for recognizing a melody
    TW535073B (en) * 2000-10-23 2003-06-01 Ntt Comm Corp Music recognition method, system and computer readable medium storing music recognition program, and, commercial recognition method, system and computer readable medium storing commercial recognition program
    FR2815760B1 (en) * 2000-10-24 2003-01-24 Philippe Ecrement METHOD FOR CONSTITUTING A DIGITAL SIGNAL REPRESENTATIVE OF A SOUND SIGNAL AND METHOD FOR RECOGNIZING A SOUND SIGNAL
    AU2002245191A1 (en) 2000-10-26 2002-07-24 Digimarc Corporation Method and system for internet access
    DE10058811A1 (en) * 2000-11-27 2002-06-13 Philips Corp Intellectual Pty Method for identifying pieces of music e.g. for discotheques, department stores etc., involves determining agreement of melodies and/or lyrics with music pieces known by analysis device
    US20020072982A1 (en) * 2000-12-12 2002-06-13 Shazam Entertainment Ltd. Method and system for interacting with a user in an experiential environment
    US8055899B2 (en) 2000-12-18 2011-11-08 Digimarc Corporation Systems and methods using digital watermarking and identifier extraction to provide promotional opportunities
    US8103877B2 (en) 2000-12-21 2012-01-24 Digimarc Corporation Content identification and electronic tickets, coupons and credits
    WO2002051063A1 (en) 2000-12-21 2002-06-27 Digimarc Corporation Methods, apparatus and programs for generating and utilizing content signatures
    JP2004534274A (en) * 2001-03-23 2004-11-11 インスティチュート・フォー・インフォコム・リサーチ Method and system for displaying music information on a digital display for use in content-based multimedia information retrieval
    US7248715B2 (en) * 2001-04-06 2007-07-24 Digimarc Corporation Digitally watermarking physical media
    DE10117870B4 (en) 2001-04-10 2005-06-09 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Method and apparatus for transferring a music signal into a score-based description and method and apparatus for referencing a music signal in a database
    US8457346B2 (en) 2001-04-24 2013-06-04 Digimarc Corporation Digital watermarking image signals on-chip
    US7046819B2 (en) 2001-04-25 2006-05-16 Digimarc Corporation Encoded reference signal for digital watermarks
    DE10144087B4 (en) * 2001-09-08 2008-10-30 Promediascan Ag Method for detecting and registering copyrighted music sequences in radio and television programs
    ATE363118T1 (en) * 2001-09-28 2007-06-15 Ericsson Telefon Ab L M ELECTRONIC CONNECTION DEVICE WITH KARAOKE FUNCTION
    US20050106546A1 (en) * 2001-09-28 2005-05-19 George Strom Electronic communications device with a karaoke function
    US6724914B2 (en) 2001-10-16 2004-04-20 Digimarc Corporation Progressive watermark decoding on a distributed computing platform
    US6528715B1 (en) * 2001-10-31 2003-03-04 Hewlett-Packard Company Music search by interactive graphical specification with audio feedback
    US6995309B2 (en) * 2001-12-06 2006-02-07 Hewlett-Packard Development Company, L.P. System and method for music identification
    US20030140139A1 (en) * 2002-01-14 2003-07-24 Richard Marejka Self-monitoring and trending service system with a cascaded pipeline with a unique data storage and retrieval structures
    US7321667B2 (en) 2002-01-18 2008-01-22 Digimarc Corporation Data hiding through arrangement of objects
    US7824029B2 (en) 2002-05-10 2010-11-02 L-1 Secure Credentialing, Inc. Identification card printer-assembler for over the counter card issuing
    EP1381024B1 (en) * 2002-07-10 2010-01-13 Palm, Inc. Method for retrieving a tone sequence
    CN1703734A (en) * 2002-10-11 2005-11-30 松下电器产业株式会社 Method and apparatus for determining musical notes from sounds
    US7606790B2 (en) * 2003-03-03 2009-10-20 Digimarc Corporation Integrating and enhancing searching of media content and biometric databases
    US20080017017A1 (en) * 2003-11-21 2008-01-24 Yongwei Zhu Method and Apparatus for Melody Representation and Matching for Music Retrieval
    US20050216512A1 (en) * 2004-03-26 2005-09-29 Rahav Dor Method of accessing a work of art, a product, or other tangible or intangible objects without knowing the title or name thereof using fractional sampling of the work of art or object
    JP5279270B2 (en) 2004-08-06 2013-09-04 ディジマーク コーポレイション Fast signal detection and distributed computing in portable computing devices
    KR20060073100A (en) * 2004-12-24 2006-06-28 삼성전자주식회사 Sound searching terminal of searching sound media's pattern type and the method
    US7788279B2 (en) * 2005-11-10 2010-08-31 Soundhound, Inc. System and method for storing and retrieving non-text-based information
    US20070162761A1 (en) 2005-12-23 2007-07-12 Davis Bruce L Methods and Systems to Help Detect Identity Fraud
    US8738749B2 (en) 2006-08-29 2014-05-27 Digimarc Corporation Content monitoring and host compliance evaluation
    US8707459B2 (en) 2007-01-19 2014-04-22 Digimarc Corporation Determination of originality of content
    US8010511B2 (en) 2006-08-29 2011-08-30 Attributor Corporation Content monitoring and compliance enforcement
    US7378588B1 (en) 2006-09-12 2008-05-27 Chieh Changfan Melody-based music search
    JP2010518428A (en) 2007-02-01 2010-05-27 ミューズアミ, インコーポレイテッド Music transcription
    JP2010518459A (en) * 2007-02-14 2010-05-27 ミューズアミ, インコーポレイテッド Web portal for editing distributed audio files
    US8283546B2 (en) * 2007-03-28 2012-10-09 Van Os Jan L Melody encoding and searching system
    US7840177B2 (en) * 2007-05-23 2010-11-23 Landmark Digital Services, Llc Device for monitoring multiple broadcast signals
    US8494257B2 (en) 2008-02-13 2013-07-23 Museami, Inc. Music score deconstruction
    US8805110B2 (en) 2008-08-19 2014-08-12 Digimarc Corporation Methods and systems for content processing
    US20100205628A1 (en) 2009-02-12 2010-08-12 Davis Bruce L Media processing methods and arrangements
    US9390167B2 (en) 2010-07-29 2016-07-12 Soundhound, Inc. System and methods for continuous audio matching
    US8049093B2 (en) * 2009-12-30 2011-11-01 Motorola Solutions, Inc. Method and apparatus for best matching an audible query to a set of audible targets
    CN102959544B (en) * 2010-05-04 2016-06-08 沙扎姆娱乐有限公司 For the method and system of synchronized multimedia
    US9047371B2 (en) 2010-07-29 2015-06-02 Soundhound, Inc. System and method for matching a query against a broadcast stream
    EP2648181B1 (en) 2010-12-01 2017-07-26 YAMAHA Corporation Musical data retrieval on the basis of rhythm pattern similarity
    US11062615B1 (en) 2011-03-01 2021-07-13 Intelligibility Training LLC Methods and systems for remote language learning in a pandemic-aware world
    US10019995B1 (en) 2011-03-01 2018-07-10 Alice J. Stiebel Methods and systems for language learning based on a series of pitch patterns
    US9035163B1 (en) 2011-05-10 2015-05-19 Soundbound, Inc. System and method for targeting content based on identified audio and multimedia
    CN102522083B (en) * 2011-11-29 2014-03-05 北京百纳威尔科技有限公司 Method for searching hummed song by using mobile terminal and mobile terminal thereof
    US8893167B2 (en) 2012-02-07 2014-11-18 Turner Broadcasting System, Inc. Method and system for automatic content recognition based on customized user preferences
    US10957310B1 (en) 2012-07-23 2021-03-23 Soundhound, Inc. Integrated programming framework for speech and text understanding with meaning parsing
    US9288509B2 (en) 2012-12-28 2016-03-15 Turner Broadcasting System, Inc. Method and system for providing synchronized advertisements and services
    US9507849B2 (en) 2013-11-28 2016-11-29 Soundhound, Inc. Method for combining a query and a communication command in a natural language computer system
    US9292488B2 (en) 2014-02-01 2016-03-22 Soundhound, Inc. Method for embedding voice mail in a spoken utterance using a natural language processing computer system
    US11295730B1 (en) 2014-02-27 2022-04-05 Soundhound, Inc. Using phonetic variants in a local context to improve natural language understanding
    US9564123B1 (en) 2014-05-12 2017-02-07 Soundhound, Inc. Method and system for building an integrated user profile
    JP6245650B2 (en) * 2014-08-26 2017-12-13 株式会社エクシング Music search program
    WO2017028116A1 (en) * 2015-08-16 2017-02-23 胡丹丽 Intelligent desktop speaker and method for controlling intelligent desktop speaker
    US10701438B2 (en) 2016-12-31 2020-06-30 Turner Broadcasting System, Inc. Automatic content recognition and verification in a broadcast chain
    CN109036463B (en) * 2018-09-13 2021-02-12 广州酷狗计算机科技有限公司 Method, device and storage medium for acquiring difficulty information of songs

    Family Cites Families (7)

    * Cited by examiner, † Cited by third party
    Publication number Priority date Publication date Assignee Title
    JP2797644B2 (en) * 1990-05-02 1998-09-17 ブラザー工業株式会社 Karaoke device with vocalization function
    JP2897659B2 (en) * 1994-10-31 1999-05-31 ヤマハ株式会社 Karaoke equipment
    DE69521597T2 (en) * 1995-03-08 2002-07-11 Generalmusic S.P.A., San Giovanni In Marignano Device with microprocessor for the selection and recognition of pieces of music
    US5616876A (en) * 1995-04-19 1997-04-01 Microsoft Corporation System and methods for selecting music on the basis of subjective content
    JP3087602B2 (en) * 1995-05-02 2000-09-11 ヤマハ株式会社 Communication karaoke system
    US5732216A (en) * 1996-10-02 1998-03-24 Internet Angles, Inc. Audio message exchange system
    DE19652225A1 (en) * 1996-12-16 1998-06-25 Harald Rieck Process for automatic identification of melodies

    Also Published As

    Publication number Publication date
    DE69908226T2 (en) 2004-03-25
    DE69908226D1 (en) 2003-07-03
    US6121530A (en) 2000-09-19
    EP0944033A1 (en) 1999-09-22

    Similar Documents

    Publication Publication Date Title
    EP0944033B1 (en) Melody retrieval system and method
    US7487180B2 (en) System and method for recognizing audio pieces via audio fingerprinting
    US6188010B1 (en) Music search by melody input
    US20060224260A1 (en) Scan shuffle for building playlists
    US6476306B2 (en) Method and a system for recognizing a melody
    KR100952186B1 (en) Method of identifying pieces of music
    JP3433818B2 (en) Music search device
    Soulez et al. Improving polyphonic and poly-instrumental music to score alignment
    KR100893797B1 (en) Content searching device and method, and communication system and method
    KR100895009B1 (en) System and method for recommending music
    US20060065105A1 (en) Music search system and music search apparatus
    WO2003054852A3 (en) System and method for music inditification
    EP1244093A2 (en) Sound features extracting apparatus, sound data registering apparatus, sound data retrieving apparatus and methods and programs for implementing the same
    US20120166631A1 (en) Device and method for monitoring, rating and/or tuning to an audio content channel
    JPH11120198A (en) Musical piece retrieval device
    EP1426921A1 (en) Music searching apparatus and method
    US20040060424A1 (en) Method for converting a music signal into a note-based description and for referencing a music signal in a data bank
    WO2004027646A1 (en) Music classification device, music classification method, and program
    Zhu et al. Music key detection for musical audio
    JP2000187671A (en) Music retrieval system with singing voice using network and singing voice input terminal equipment to be used at the time of retrieval
    JPH10319948A (en) Sound source kind discriminating method of musical instrument included in musical playing
    JPH11272274A (en) Method for retrieving piece of music by use of singing voice
    JP2002055695A (en) Music search system
    JPH10161654A (en) Musical classification determining device
    JP2005115164A (en) Musical composition retrieving apparatus

    Legal Events

    Date Code Title Description
    PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

    Free format text: ORIGINAL CODE: 0009012

    AK Designated contracting states

    Kind code of ref document: A1

    Designated state(s): DE FI FR GB IT

    AX Request for extension of the european patent

    Free format text: AL;LT;LV;MK;RO;SI

    17P Request for examination filed

    Effective date: 20000316

    AKX Designation fees paid

    Free format text: DE FI FR GB IT

    GRAH Despatch of communication of intention to grant a patent

    Free format text: ORIGINAL CODE: EPIDOS IGRA

    GRAH Despatch of communication of intention to grant a patent

    Free format text: ORIGINAL CODE: EPIDOS IGRA

    GRAA (expected) grant

    Free format text: ORIGINAL CODE: 0009210

    AK Designated contracting states

    Designated state(s): DE FI FR GB IT

    REG Reference to a national code

    Ref country code: GB

    Ref legal event code: FG4D

    REF Corresponds to:

    Ref document number: 69908226

    Country of ref document: DE

    Date of ref document: 20030703

    Kind code of ref document: P

    ET Fr: translation filed
    PLBE No opposition filed within time limit

    Free format text: ORIGINAL CODE: 0009261

    STAA Information on the status of an ep patent application or granted ep patent

    Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

    26N No opposition filed

    Effective date: 20040302

    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: IT

    Payment date: 20080308

    Year of fee payment: 10

    Ref country code: FI

    Payment date: 20080307

    Year of fee payment: 10

    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: FR

    Payment date: 20080310

    Year of fee payment: 10

    Ref country code: DE

    Payment date: 20080327

    Year of fee payment: 10

    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: GB

    Payment date: 20090309

    Year of fee payment: 11

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: FI

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20090319

    REG Reference to a national code

    Ref country code: FR

    Ref legal event code: ST

    Effective date: 20091130

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: DE

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20091001

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: FR

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20091123

    GBPC Gb: european patent ceased through non-payment of renewal fee

    Effective date: 20100319

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: IT

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20090319

    Ref country code: GB

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20100319