EP1147511A1 - Verfahren zur automatischen erkennung von musikstücken und klangsignalen - Google Patents
Verfahren zur automatischen erkennung von musikstücken und klangsignalenInfo
- Publication number
- EP1147511A1 EP1147511A1 EP00940675A EP00940675A EP1147511A1 EP 1147511 A1 EP1147511 A1 EP 1147511A1 EP 00940675 A EP00940675 A EP 00940675A EP 00940675 A EP00940675 A EP 00940675A EP 1147511 A1 EP1147511 A1 EP 1147511A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- signal
- vectors
- model
- unknown
- group
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/02—Means for controlling the tone frequencies, e.g. attack or decay; Means for producing special musical effects, e.g. vibratos or glissandos
- G10H1/06—Circuits for establishing the harmonic content of tones, or other arrangements for changing the tone colour
- G10H1/12—Circuits for establishing the harmonic content of tones, or other arrangements for changing the tone colour by filtering complex waveforms
- G10H1/125—Circuits for establishing the harmonic content of tones, or other arrangements for changing the tone colour by filtering complex waveforms using a digital filter
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/121—Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods
- G10H2240/131—Library retrieval, i.e. searching a database or selecting a specific musical piece, segment, pattern, rule or parameter set
- G10H2240/141—Library retrieval matching, i.e. any of the steps of matching an inputted segment or phrase with musical database contents, e.g. query by humming, singing or playing; the steps may include, e.g. musical analysis of the input, musical feature extraction, query formulation, or details of the retrieval process
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2250/00—Aspects of algorithms or signal processing methods without intrinsic musical character, yet specifically adapted for or used in electrophonic musical processing
- G10H2250/131—Mathematical functions for musical analysis, processing, synthesis or composition
- G10H2250/215—Transforms, i.e. mathematical transforms into domains appropriate for musical signal processing, coding or compression
- G10H2250/235—Fourier transform; Discrete Fourier Transform [DFT]; Fast Fourier Transform [FFT]
Definitions
- This invention refers to a method of automatic recognition of musical compositions and sound signals and it is used in order to identify musical compositions and sound signals transmitted by radio, TV and/or performed in public places.
- the unknown musical composition or sound signal is received, in which the same procedure of extracting a corresponding set of characteristics is applied. These characteristics are compared with the corresponding sets of characteristics of the model signals and, by means of a number original criteria, it is decided if one (and which one exactly) of the model signals corresponds to the unknown signal under consideration. This procedure is described in figure 1.
- the whole frequency band from 0 to 11025 Hz is divided to sub-bands that are almost exponentially distributed.
- Hz is divided in 60 sub-bands.
- each model signal is digitised with a random sampling frequency F s preferably greater than or equal to 11025 Hz and a window of 8192 or 16384 or 32768 sample length, slides on the obtained digitised signal.
- F s random sampling frequency
- a window of 8192 or 16384 or 32768 sample length slides on the obtained digitised signal.
- an adaptive Fast Fourier Transform is applied and the Discrete Fourier Transform absolute value is obtained.
- the frequency domain window is divided in sections according to the aforementioned frequency sub-bands choice (see Table 1) and then, in every such section, all the peaks of the absolute value of the Fourier transform are spotted and the greater one is obtained. The value of this peak is called "section representative".
- Wf_ 32768 samples is obtained; notice that in any case this window will be of the same length with the sliding window which was used for the model signals.
- the L greater value representatives are spotted, where the value of L is the same with the one used for the model signals.
- the window slides for l samples where the value of ⁇ i may vary from 0,55 * F s to 1,9 * F s samples, with most frequently used value the
- STEP is a parameter expressing the shift step, that usually belongs to the interval [0.005, 0.01], the more frequently used value being 0.0075.
- the identification procedure described so far is depicted in figure 3.
- each group of unknown signal representatives is being compared with elements of the set of representatives of each model signal separately.
- each of the S+l groups of M unknown signal representatives is compared with groups of M model signal representatives by means of the method consisting of the following steps:
- V ! [60555249474339343330292220171411952 l]
- step E 2 If, indeed, it is greater than or equal to 0.5 ⁇ *L, we proceed to step E 2 below. If it is smaller than 0.51* , then we consider that the set of the tests performed so far did not result to a successful recognition, so, after considering U j as the next representative- vector of the model signal, we start the comparison procedure again, beginning from the comparison of the vector V j with the new U j .
- step E 3 If it is greater or equal, we proceed to step E 3 below. If it is smaller, then we consider that the set of tests performed so far did not result to a successful recognition, so, after considering U as the next representative- vector of the model signal, the comparison procedure starts again beginning from the comparison of the vector V j with the new U j .
- step E M If it is greater or equal, we proceed to step E M below. If it is smaller, then we consider that the set of tests performed so far did not result to a successful recognition, so, after considering U j as the next representative- vector of the model signal, the comparison procedure starts again beginning from the comparison of the vector V j with the new U j .
- V M the M representative vector of the unknown signal corresponding to the same with V j shift coefficient fj .
- the comparison procedure starts again beginning from the comparison of the vector V with the new U . If all possible vectors of the model signal are unsuccessfully compared with one group of representatives of the unknown signal corresponding to the specific shift coefficient / , then we repeat the comparison procedure, using the group of representatives of the unknown signal corresponding to the next shift coefficient f i+l . If the comparison of a specific set of model vectors with all (S+l) groups of representatives of the unknown signal is unsuccessful, then we proceed to the comparison of the unknown signal with another set of model vectors.
- the L greater value representatives are spotted, where the value of L is the same with the one used in the first criterion.
- the irrevocable group of representatives of the unknown signal is compared to elements of the set of the representatives of the model signal, by means of a method similar to the first criterion consisting of the steps briefly described below:
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GR99100235 | 1999-07-08 | ||
GR99100235 | 1999-07-08 | ||
PCT/GR2000/000024 WO2001004870A1 (en) | 1999-07-08 | 2000-07-07 | Method of automatic recognition of musical compositions and sound signals |
Publications (1)
Publication Number | Publication Date |
---|---|
EP1147511A1 true EP1147511A1 (de) | 2001-10-24 |
Family
ID=10943871
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP00940675A Withdrawn EP1147511A1 (de) | 1999-07-08 | 2000-07-07 | Verfahren zur automatischen erkennung von musikstücken und klangsignalen |
Country Status (3)
Country | Link |
---|---|
EP (1) | EP1147511A1 (de) |
GR (1) | GR1003625B (de) |
WO (1) | WO2001004870A1 (de) |
Families Citing this family (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6990453B2 (en) | 2000-07-31 | 2006-01-24 | Landmark Digital Services Llc | System and methods for recognizing sound and music signals in high noise and distortion |
US7853664B1 (en) | 2000-07-31 | 2010-12-14 | Landmark Digital Services Llc | Method and system for purchasing pre-recorded music |
DE10134471C2 (de) * | 2001-02-28 | 2003-05-22 | Fraunhofer Ges Forschung | Verfahren und Vorrichtung zum Charakterisieren eines Signals und Verfahren und Vorrichtung zum Erzeugen eines indexierten Signals |
TW582022B (en) * | 2001-03-14 | 2004-04-01 | Ibm | A method and system for the automatic detection of similar or identical segments in audio recordings |
DE10117870B4 (de) | 2001-04-10 | 2005-06-09 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Verfahren und Vorrichtung zum Überführen eines Musiksignals in eine Noten-basierte Beschreibung und Verfahren und Vorrichtung zum Referenzieren eines Musiksignals in einer Datenbank |
DE60236161D1 (de) | 2001-07-20 | 2010-06-10 | Gracenote Inc | Automatische identifizierung von klangaufzeichnungen |
DE10157454B4 (de) | 2001-11-23 | 2005-07-07 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Verfahren und Vorrichtung zum Erzeugen einer Kennung für ein Audiosignal, Verfahren und Vorrichtung zum Aufbauen einer Instrumentendatenbank und Verfahren und Vorrichtung zum Bestimmen der Art eines Instruments |
US6995309B2 (en) * | 2001-12-06 | 2006-02-07 | Hewlett-Packard Development Company, L.P. | System and method for music identification |
DK1504445T3 (da) * | 2002-04-25 | 2008-12-01 | Landmark Digital Services Llc | Robust og invariant lydmönster-matching |
DE10232916B4 (de) * | 2002-07-19 | 2008-08-07 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zum Charakterisieren eines Informationssignals |
GB2391322B (en) * | 2002-07-31 | 2005-12-14 | British Broadcasting Corp | Signal comparison method and apparatus |
JP4933899B2 (ja) | 2004-02-19 | 2012-05-16 | ランドマーク、ディジタル、サーヴィセズ、エルエルシー | 放送源の識別のための方法および装置 |
DE102004023436B4 (de) * | 2004-05-10 | 2006-06-14 | M2Any Gmbh | Vorrichtung und Verfahren zum Analysieren eines Informationssignals |
DE102004028694B3 (de) * | 2004-06-14 | 2005-12-22 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zum Umsetzen eines Informationssignals in eine Spektraldarstellung mit variabler Auflösung |
DE102004028693B4 (de) | 2004-06-14 | 2009-12-31 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zum Bestimmen eines Akkordtyps, der einem Testsignal zugrunde liegt |
CN100485399C (zh) | 2004-06-24 | 2009-05-06 | 兰德马克数字服务有限责任公司 | 表征两个媒体段的重叠的方法 |
WO2006086556A2 (en) | 2005-02-08 | 2006-08-17 | Landmark Digital Services Llc | Automatic identfication of repeated material in audio signals |
US8453170B2 (en) | 2007-02-27 | 2013-05-28 | Landmark Digital Services Llc | System and method for monitoring and recognizing broadcast data |
CN103971689B (zh) * | 2013-02-04 | 2016-01-27 | 腾讯科技(深圳)有限公司 | 一种音频识别方法及装置 |
CN104093079B (zh) | 2014-05-29 | 2015-10-07 | 腾讯科技(深圳)有限公司 | 基于多媒体节目的交互方法、终端、服务器和系统 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5210820A (en) * | 1990-05-02 | 1993-05-11 | Broadcast Data Systems Limited Partnership | Signal recognition system and method |
US5874686A (en) * | 1995-10-31 | 1999-02-23 | Ghias; Asif U. | Apparatus and method for searching a melody |
US5778335A (en) * | 1996-02-26 | 1998-07-07 | The Regents Of The University Of California | Method and apparatus for efficient multiband celp wideband speech and music coding and decoding |
-
1999
- 1999-07-08 GR GR990100235A patent/GR1003625B/el not_active IP Right Cessation
-
2000
- 2000-07-07 WO PCT/GR2000/000024 patent/WO2001004870A1/en not_active Application Discontinuation
- 2000-07-07 EP EP00940675A patent/EP1147511A1/de not_active Withdrawn
Non-Patent Citations (1)
Title |
---|
See references of WO0104870A1 * |
Also Published As
Publication number | Publication date |
---|---|
GR1003625B (el) | 2001-08-31 |
WO2001004870A1 (en) | 2001-01-18 |
GR990100235A (el) | 2001-03-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2001004870A1 (en) | Method of automatic recognition of musical compositions and sound signals | |
Delforouzi et al. | Adaptive digital audio steganography based on integer wavelet transform | |
US6453252B1 (en) | Process for identifying audio content | |
JP4418748B2 (ja) | ストリームに繰り返し埋め込まれたメディアオブジェクトを識別し、セグメント化するためのシステムおよび方法 | |
JP5728888B2 (ja) | 信号処理装置および方法、並びにプログラム | |
JP2006505821A (ja) | 指紋情報付マルチメディアコンテンツ | |
JP2000105146A (ja) | 複合音信号中での音特定方法及び装置 | |
EP1515310A1 (de) | System und Verfahren zur hochqualitativen Verlängerung und Verkürzung eines digitalen Audiosignals | |
US10089994B1 (en) | Acoustic fingerprint extraction and matching | |
US20060041753A1 (en) | Fingerprint extraction | |
CA2537328A1 (en) | Method of processing and storing mass spectrometry data | |
EP1451803A2 (de) | System und verfahren zur identifizierung von musik | |
CN110277087B (zh) | 一种广播信号预判预处理方法 | |
Gajic et al. | Robust speech recognition using features based on zero crossings with peak amplitudes | |
CN106716529A (zh) | 对数字音频信号中的前回声进行辨别和衰减 | |
KR100527002B1 (ko) | 음성 신호의 에너지 분포 특성을 고려한 쉐이핑 장치 및 방법 | |
Yamashita et al. | Spectral subtraction iterated with weighting factors | |
Richly et al. | Short-term sound stream characterization for reliable, real-time occurrence monitoring of given sound-prints | |
GB2294619A (en) | Inaudible insertion of information into an audio signal | |
Wang et al. | Audio fingerprint based on spectral flux for audio retrieval | |
Adjila et al. | Silence Detection and Removal Method Based on the Continuous Average Energy of Speech Signal | |
Tang | Evaluation of double sided periodic substitution (DSPS) method for recovering missing speech in packet voice communications | |
JP3252005B2 (ja) | 適応ブロック長変換符号化のブロック長選択装置 | |
WO1998022935A9 (en) | Formant extraction using peak-picking and smoothing techniques | |
Jančovič et al. | A probabilistic union model with automatic order selection for noisy speech recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20010718 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20030201 |