EP1147511A1 - Verfahren zur automatischen erkennung von musikstücken und klangsignalen - Google Patents

Verfahren zur automatischen erkennung von musikstücken und klangsignalen

Info

Publication number
EP1147511A1
EP1147511A1 EP00940675A EP00940675A EP1147511A1 EP 1147511 A1 EP1147511 A1 EP 1147511A1 EP 00940675 A EP00940675 A EP 00940675A EP 00940675 A EP00940675 A EP 00940675A EP 1147511 A1 EP1147511 A1 EP 1147511A1
Authority
EP
European Patent Office
Prior art keywords
signal
vectors
model
unknown
group
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP00940675A
Other languages
English (en)
French (fr)
Inventor
Constantin Papaodysseus
Constantin Triantafillou
George Roussopoulos
Constantin Alexiou
Athanasios Panagopoulos
Dimitrios Fragoulis
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of EP1147511A1 publication Critical patent/EP1147511A1/de
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/02Means for controlling the tone frequencies, e.g. attack or decay; Means for producing special musical effects, e.g. vibratos or glissandos
    • G10H1/06Circuits for establishing the harmonic content of tones, or other arrangements for changing the tone colour
    • G10H1/12Circuits for establishing the harmonic content of tones, or other arrangements for changing the tone colour by filtering complex waveforms
    • G10H1/125Circuits for establishing the harmonic content of tones, or other arrangements for changing the tone colour by filtering complex waveforms using a digital filter
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/121Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods
    • G10H2240/131Library retrieval, i.e. searching a database or selecting a specific musical piece, segment, pattern, rule or parameter set
    • G10H2240/141Library retrieval matching, i.e. any of the steps of matching an inputted segment or phrase with musical database contents, e.g. query by humming, singing or playing; the steps may include, e.g. musical analysis of the input, musical feature extraction, query formulation, or details of the retrieval process
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2250/00Aspects of algorithms or signal processing methods without intrinsic musical character, yet specifically adapted for or used in electrophonic musical processing
    • G10H2250/131Mathematical functions for musical analysis, processing, synthesis or composition
    • G10H2250/215Transforms, i.e. mathematical transforms into domains appropriate for musical signal processing, coding or compression
    • G10H2250/235Fourier transform; Discrete Fourier Transform [DFT]; Fast Fourier Transform [FFT]

Definitions

  • This invention refers to a method of automatic recognition of musical compositions and sound signals and it is used in order to identify musical compositions and sound signals transmitted by radio, TV and/or performed in public places.
  • the unknown musical composition or sound signal is received, in which the same procedure of extracting a corresponding set of characteristics is applied. These characteristics are compared with the corresponding sets of characteristics of the model signals and, by means of a number original criteria, it is decided if one (and which one exactly) of the model signals corresponds to the unknown signal under consideration. This procedure is described in figure 1.
  • the whole frequency band from 0 to 11025 Hz is divided to sub-bands that are almost exponentially distributed.
  • Hz is divided in 60 sub-bands.
  • each model signal is digitised with a random sampling frequency F s preferably greater than or equal to 11025 Hz and a window of 8192 or 16384 or 32768 sample length, slides on the obtained digitised signal.
  • F s random sampling frequency
  • a window of 8192 or 16384 or 32768 sample length slides on the obtained digitised signal.
  • an adaptive Fast Fourier Transform is applied and the Discrete Fourier Transform absolute value is obtained.
  • the frequency domain window is divided in sections according to the aforementioned frequency sub-bands choice (see Table 1) and then, in every such section, all the peaks of the absolute value of the Fourier transform are spotted and the greater one is obtained. The value of this peak is called "section representative".
  • Wf_ 32768 samples is obtained; notice that in any case this window will be of the same length with the sliding window which was used for the model signals.
  • the L greater value representatives are spotted, where the value of L is the same with the one used for the model signals.
  • the window slides for l samples where the value of ⁇ i may vary from 0,55 * F s to 1,9 * F s samples, with most frequently used value the
  • STEP is a parameter expressing the shift step, that usually belongs to the interval [0.005, 0.01], the more frequently used value being 0.0075.
  • the identification procedure described so far is depicted in figure 3.
  • each group of unknown signal representatives is being compared with elements of the set of representatives of each model signal separately.
  • each of the S+l groups of M unknown signal representatives is compared with groups of M model signal representatives by means of the method consisting of the following steps:
  • V ! [60555249474339343330292220171411952 l]
  • step E 2 If, indeed, it is greater than or equal to 0.5 ⁇ *L, we proceed to step E 2 below. If it is smaller than 0.51* , then we consider that the set of the tests performed so far did not result to a successful recognition, so, after considering U j as the next representative- vector of the model signal, we start the comparison procedure again, beginning from the comparison of the vector V j with the new U j .
  • step E 3 If it is greater or equal, we proceed to step E 3 below. If it is smaller, then we consider that the set of tests performed so far did not result to a successful recognition, so, after considering U as the next representative- vector of the model signal, the comparison procedure starts again beginning from the comparison of the vector V j with the new U j .
  • step E M If it is greater or equal, we proceed to step E M below. If it is smaller, then we consider that the set of tests performed so far did not result to a successful recognition, so, after considering U j as the next representative- vector of the model signal, the comparison procedure starts again beginning from the comparison of the vector V j with the new U j .
  • V M the M representative vector of the unknown signal corresponding to the same with V j shift coefficient fj .
  • the comparison procedure starts again beginning from the comparison of the vector V with the new U . If all possible vectors of the model signal are unsuccessfully compared with one group of representatives of the unknown signal corresponding to the specific shift coefficient / , then we repeat the comparison procedure, using the group of representatives of the unknown signal corresponding to the next shift coefficient f i+l . If the comparison of a specific set of model vectors with all (S+l) groups of representatives of the unknown signal is unsuccessful, then we proceed to the comparison of the unknown signal with another set of model vectors.
  • the L greater value representatives are spotted, where the value of L is the same with the one used in the first criterion.
  • the irrevocable group of representatives of the unknown signal is compared to elements of the set of the representatives of the model signal, by means of a method similar to the first criterion consisting of the steps briefly described below:
EP00940675A 1999-07-08 2000-07-07 Verfahren zur automatischen erkennung von musikstücken und klangsignalen Withdrawn EP1147511A1 (de)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
GR99100235 1999-07-08
GR99100235 1999-07-08
PCT/GR2000/000024 WO2001004870A1 (en) 1999-07-08 2000-07-07 Method of automatic recognition of musical compositions and sound signals

Publications (1)

Publication Number Publication Date
EP1147511A1 true EP1147511A1 (de) 2001-10-24

Family

ID=10943871

Family Applications (1)

Application Number Title Priority Date Filing Date
EP00940675A Withdrawn EP1147511A1 (de) 1999-07-08 2000-07-07 Verfahren zur automatischen erkennung von musikstücken und klangsignalen

Country Status (3)

Country Link
EP (1) EP1147511A1 (de)
GR (1) GR1003625B (de)
WO (1) WO2001004870A1 (de)

Families Citing this family (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6990453B2 (en) 2000-07-31 2006-01-24 Landmark Digital Services Llc System and methods for recognizing sound and music signals in high noise and distortion
US7853664B1 (en) 2000-07-31 2010-12-14 Landmark Digital Services Llc Method and system for purchasing pre-recorded music
DE10134471C2 (de) * 2001-02-28 2003-05-22 Fraunhofer Ges Forschung Verfahren und Vorrichtung zum Charakterisieren eines Signals und Verfahren und Vorrichtung zum Erzeugen eines indexierten Signals
TW582022B (en) * 2001-03-14 2004-04-01 Ibm A method and system for the automatic detection of similar or identical segments in audio recordings
DE10117870B4 (de) 2001-04-10 2005-06-09 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Verfahren und Vorrichtung zum Überführen eines Musiksignals in eine Noten-basierte Beschreibung und Verfahren und Vorrichtung zum Referenzieren eines Musiksignals in einer Datenbank
DE60236161D1 (de) 2001-07-20 2010-06-10 Gracenote Inc Automatische identifizierung von klangaufzeichnungen
DE10157454B4 (de) 2001-11-23 2005-07-07 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Verfahren und Vorrichtung zum Erzeugen einer Kennung für ein Audiosignal, Verfahren und Vorrichtung zum Aufbauen einer Instrumentendatenbank und Verfahren und Vorrichtung zum Bestimmen der Art eines Instruments
US6995309B2 (en) * 2001-12-06 2006-02-07 Hewlett-Packard Development Company, L.P. System and method for music identification
DK1504445T3 (da) * 2002-04-25 2008-12-01 Landmark Digital Services Llc Robust og invariant lydmönster-matching
DE10232916B4 (de) * 2002-07-19 2008-08-07 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Vorrichtung und Verfahren zum Charakterisieren eines Informationssignals
GB2391322B (en) * 2002-07-31 2005-12-14 British Broadcasting Corp Signal comparison method and apparatus
JP4933899B2 (ja) 2004-02-19 2012-05-16 ランドマーク、ディジタル、サーヴィセズ、エルエルシー 放送源の識別のための方法および装置
DE102004023436B4 (de) * 2004-05-10 2006-06-14 M2Any Gmbh Vorrichtung und Verfahren zum Analysieren eines Informationssignals
DE102004028694B3 (de) * 2004-06-14 2005-12-22 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Vorrichtung und Verfahren zum Umsetzen eines Informationssignals in eine Spektraldarstellung mit variabler Auflösung
DE102004028693B4 (de) 2004-06-14 2009-12-31 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Vorrichtung und Verfahren zum Bestimmen eines Akkordtyps, der einem Testsignal zugrunde liegt
CN100485399C (zh) 2004-06-24 2009-05-06 兰德马克数字服务有限责任公司 表征两个媒体段的重叠的方法
WO2006086556A2 (en) 2005-02-08 2006-08-17 Landmark Digital Services Llc Automatic identfication of repeated material in audio signals
US8453170B2 (en) 2007-02-27 2013-05-28 Landmark Digital Services Llc System and method for monitoring and recognizing broadcast data
CN103971689B (zh) * 2013-02-04 2016-01-27 腾讯科技(深圳)有限公司 一种音频识别方法及装置
CN104093079B (zh) 2014-05-29 2015-10-07 腾讯科技(深圳)有限公司 基于多媒体节目的交互方法、终端、服务器和系统

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5210820A (en) * 1990-05-02 1993-05-11 Broadcast Data Systems Limited Partnership Signal recognition system and method
US5874686A (en) * 1995-10-31 1999-02-23 Ghias; Asif U. Apparatus and method for searching a melody
US5778335A (en) * 1996-02-26 1998-07-07 The Regents Of The University Of California Method and apparatus for efficient multiband celp wideband speech and music coding and decoding

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of WO0104870A1 *

Also Published As

Publication number Publication date
GR1003625B (el) 2001-08-31
WO2001004870A1 (en) 2001-01-18
GR990100235A (el) 2001-03-30

Similar Documents

Publication Publication Date Title
WO2001004870A1 (en) Method of automatic recognition of musical compositions and sound signals
Delforouzi et al. Adaptive digital audio steganography based on integer wavelet transform
US6453252B1 (en) Process for identifying audio content
JP4418748B2 (ja) ストリームに繰り返し埋め込まれたメディアオブジェクトを識別し、セグメント化するためのシステムおよび方法
JP5728888B2 (ja) 信号処理装置および方法、並びにプログラム
JP2006505821A (ja) 指紋情報付マルチメディアコンテンツ
JP2000105146A (ja) 複合音信号中での音特定方法及び装置
EP1515310A1 (de) System und Verfahren zur hochqualitativen Verlängerung und Verkürzung eines digitalen Audiosignals
US10089994B1 (en) Acoustic fingerprint extraction and matching
US20060041753A1 (en) Fingerprint extraction
CA2537328A1 (en) Method of processing and storing mass spectrometry data
EP1451803A2 (de) System und verfahren zur identifizierung von musik
CN110277087B (zh) 一种广播信号预判预处理方法
Gajic et al. Robust speech recognition using features based on zero crossings with peak amplitudes
CN106716529A (zh) 对数字音频信号中的前回声进行辨别和衰减
KR100527002B1 (ko) 음성 신호의 에너지 분포 특성을 고려한 쉐이핑 장치 및 방법
Yamashita et al. Spectral subtraction iterated with weighting factors
Richly et al. Short-term sound stream characterization for reliable, real-time occurrence monitoring of given sound-prints
GB2294619A (en) Inaudible insertion of information into an audio signal
Wang et al. Audio fingerprint based on spectral flux for audio retrieval
Adjila et al. Silence Detection and Removal Method Based on the Continuous Average Energy of Speech Signal
Tang Evaluation of double sided periodic substitution (DSPS) method for recovering missing speech in packet voice communications
JP3252005B2 (ja) 適応ブロック長変換符号化のブロック長選択装置
WO1998022935A9 (en) Formant extraction using peak-picking and smoothing techniques
Jančovič et al. A probabilistic union model with automatic order selection for noisy speech recognition

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20010718

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20030201