EP1377924A2 - Method and device for extracting a signal identifier, method and device for creating a database from signal identifiers and method and device for referencing a search time signal - Google Patents
Method and device for extracting a signal identifier, method and device for creating a database from signal identifiers and method and device for referencing a search time signalInfo
- Publication number
- EP1377924A2 EP1377924A2 EP02714186A EP02714186A EP1377924A2 EP 1377924 A2 EP1377924 A2 EP 1377924A2 EP 02714186 A EP02714186 A EP 02714186A EP 02714186 A EP02714186 A EP 02714186A EP 1377924 A2 EP1377924 A2 EP 1377924A2
- Authority
- EP
- European Patent Office
- Prior art keywords
- signal
- time
- database
- search
- identifier
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/121—Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods
- G10H2240/131—Library retrieval, i.e. searching a database or selecting a specific musical piece, segment, pattern, rule or parameter set
- G10H2240/135—Library retrieval index, i.e. using an indexing scheme to efficiently retrieve a music piece
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2250/00—Aspects of algorithms or signal processing methods without intrinsic musical character, yet specifically adapted for or used in electrophonic musical processing
- G10H2250/005—Algorithms for electrophonic musical instruments or musical processing, e.g. for automatic composition or resource allocation
- G10H2250/011—Genetic algorithms, i.e. using computational steps analogous to biological selection, recombination and mutation on an initial population of, e.g. sounds, pieces, melodies or loops to compose or otherwise generate, e.g. evolutionary music or sound synthesis
Definitions
- Method and device for extracting a signal identifier Method and device for generating a database from signal identifiers and method and device for referencing a search time signal
- the present invention relates to the processing of time signals having a harmonic component, and in particular to the generation of a signal identifier for a time signal to describe the time signal by means of a database in which a plurality of signal identifiers for a plurality of time signals are stored to be able to.
- time signals with a harmonic component such as For example, audio data that can be identified and referenced are useful for many users. Particularly in a situation where there is an audio signal whose title and author are unknown, it is often desirable to find out who wrote the song. There is a need for this, for example, if the wish is present, e.g. B. to acquire a CD of the artist in question. If the present audio signal only includes the time signal content, but no name about the artist, the music publisher, etc., then it is not possible to identify the origin of the audio signal or who wrote the song. The only hope was then to hear the audio piece together with reference data regarding the author or the source, where the audio signal can be obtained, in order to then be able to obtain the desired title.
- Audio signals or more generally, time signals that have a harmonic component can be searched by such search engines. are not processed if they do not include any textual search information.
- a realistic inventory of audio files includes several thousand stored audio files up to hundreds of thousands of audio files.
- Music database information can be stored on a central Internet server, and potential searches can be made over the Internet.
- the central music databases on local hard disk systems are also conceivable for users. It is desirable to be able to search through such music databases in order to find reference data about an audio file, of which only the file itself, but no reference data, is known.
- Similar pieces are, for example, the pieces with a similar melody, a similar set of instruments, or simply with similar sounds, such as. B. sound of the sea, twittering of birds, male voices, female voices, etc.
- US Patent No. 5,918,223 discloses a method and apparatus for content based analysis, storage, retrieval and segmentation of audio information. This method is based on extracting several acoustic features from one audio signal. Volume, bass, pitch, brightness and melody-frequency-based cepstral coefficients are measured in a time window of a certain length at periodic intervals. Each measurement data record consists of a sequence of measured feature vectors. Each audio file is specified by the complete set of feature sequences calculated for each feature. Furthermore, the first derivatives are calculated for each sequence of feature vectors. Then statistical values such as mean and standard deviation are calculated. This Set of values is stored in an N vector, that is, a vector with n elements.
- This procedure is applied to a variety of audio files to derive an N vector for each audio file.
- This gradually builds up a database from a large number of N vectors.
- a search N vector is then extracted from an unknown audio file using the same procedure.
- a distance calculation of the predefined N vector and the N vectors stored in the database is then determined.
- the N-vector is output which has the minimum distance to the search N-vector.
- the output N-vector is assigned data about the author, the title, the source of procurement, etc., so that an audio file can be identified with regard to its origin.
- the disadvantage of this method is that several characteristics are calculated and arbitrary heuristics for calculating the parameters are introduced.
- the mean value and standard deviation calculations for all feature vectors for an entire audio file reduce the information given by the time history of the feature vectors to a few feature sizes. This leads to a high loss of information.
- the object of the present invention is to provide a method and a device for extracting a signal identifier from a time signal, which enable a meaningful identification of a time signal without excessive loss of information.
- This object is achieved by a method for extracting a signal identifier from a time signal according to patent claim 1 or by a device for extracting a signal identifier from a time signal according to patent claim 19.
- Another object of the present invention is to provide a method and an apparatus for producing a ner database from signal identifiers and to create a method and an apparatus for referencing a search time signal by means of such a database.
- This object is achieved by a method for generating a database according to claim 13, a device for generating a database according to claim 20, a method for referencing a search time signal according to claim 14 or a device for referencing a search time signal according to claim 21.
- the present invention is based on the knowledge that with time signals that have a harmonic component, the time profile of the time signal can be used to extract a signal identifier of the time signal from the time signal, which on the one hand provides a good fingerprint for the time signal, but which on the other hand, it is manageable with regard to the amount of data to enable an efficient search of a large number of signal identifiers in a database.
- An essential property of time signals with a harmonic component are recurring signal edges in the time signal. B. two successive signal edges with the same or similar length allow the specification of a period and thus a frequency in the time signal with high temporal and frequency resolution, if not only the presence of the signal edges themselves, but also the temporal occurrence of the signal edges in the time signal is taken into account.
- the time signal consists of frequencies which are successive in time.
- the audio signal is thus characterized in such a way that a tone, that is to say a frequency, is present at a certain point in time, and that this tone, ie this frequency, is followed by another tone, ie a different frequency, at a later point in time.
- the description of the time signal by means of a sequence of temporal sampling values is therefore passed to a description of the time signal by coordinate tuples from frequency and time of occurrence of the frequency.
- the signal identifier or in other words the feature vector (MV), which is used to describe the time signal, thus comprises a sequence of signal identification values which, depending on the embodiment, more or less roughly reproduces the time profile of the time signal.
- the time signal is therefore not characterized on the basis of its spectral properties, as in the prior art, but on the basis of the chronological sequence of frequencies in the time signal.
- At least two detected signal edges are thus required to calculate a frequency value from the detected signal edges.
- the selection of these two signal edges from the total detected signal edges, on the basis of which frequency values are calculated, is diverse. First, two successive signal edges of essentially the same length can be used. The frequency value is then the reciprocal of the time interval between these edges. Alternatively, a selection can also be made based on the amplitude of the detected signal edges. In this way, two successive signal edges of the same amplitude can also be taken in order to determine a frequency value. However, it is not always necessary to take two successive signal edges, but e.g. B. always the second, third, fourth, ... signal edge of the same amplitude or length.
- any two signal edges can also be taken in order to obtain the coordinate tuple using statistical methods and on the basis of the superposition laws.
- a flute tone provides two signal edges with a high amplitude, between which there is a wave crest with a lower amplitude.
- a selection of the two detected signal edges is made according to the amplitude.
- the temporal sequence of tones is the most natural way of characterization, since, as can be recognized most easily by music signals, the essence of the audio signal lies in the temporal sequence of tones.
- the most immediate sensation a listener receives from a music signal is the temporal sequence of tones.
- a catchy melody which generally consists of a sequence consists of simple tones, whereby the theme or the simple melody significantly shapes the ability to recognize regardless of rhythm, pitch, any instrument accompaniment, etc.
- the concept according to the invention is based on this knowledge and provides a signal identifier which consists of a chronological sequence of frequencies or, depending on the embodiment, of a chronological sequence of frequencies, i. H. Tones derived from statistical methods.
- An advantage of the present invention is that the signal identification as a time sequence of frequencies represents a fingerprint of high information content for time signals with a harmonic component and to a certain extent constitutes the essential or the core of a time signal.
- the signal identifier extracted according to the invention represents a strong compression of the time signal, but is still based on the temporal course of the time signal and thus on the natural understanding of time signals, e.g. B. pieces of music is adjusted.
- Another advantage of the present invention is that the sequential nature of the signal identification means that the distance calculation referencing algorithms in the prior art can be dispensed with and algorithms known from DNA sequencing can be used in a database to reference the time signal and that, in addition, similarity calculations can also be performed using DNA sequencing algorithms with replace / insert / delete operations.
- a further advantage of the present invention is that the Hough transformation, for which efficient algorithms exist from image processing and image recognition, can be used in a favorable manner to detect the time occurrence of signal edges in the time signal.
- the signal identifier of a time signal extracted according to the invention is independent of whether the search signal identifier is derived from the entire time signal or only from a section of the time signal, since according to the algorithms of DNA sequencing, a time stepwise Comparison of the search signal identifier with a reference signal identifier can be carried out, the portion of the time signal to be identified being, so to speak, automatically identified in the reference time signal where the highest match between the search signal identifier and the reference signal identifier exists due to the temporally sequential comparison ,
- FIG. 1 shows a block diagram of the device according to the invention for extracting a signal identifier from a time signal
- FIG. 2 shows a block diagram of a preferred exemplary embodiment, in which preprocessing of the audio signal is shown
- FIG. 3 shows a block diagram of an exemplary embodiment for the signal identification generation
- FIG. 4 shows a block diagram for a device according to the invention for generating a database and for referencing a search time signal in the database.
- Fig. 5 graphical representation of a section of Mozart KV 581 by frequency-time coordinate tuple.
- the device comprises a device 12 for performing a signal edge detection, a device 14 for determining the distance between two selected detected edges, a device 16 for frequency calculation and a device 18 for signal recognition generation using coordinate tuples output from the device 16 for frequency calculation each have a frequency value and an occurrence time for this frequency value.
- the device 12 for detecting the temporal occurrence of signal edges in the time signal preferably carries out a Hough transformation.
- the Hough transformation is in the U.S. - Patent No. 3,069,654 by Paul V. C. Hough.
- the Hough transformation is used for the detection of complex structures and in particular for the automatic detection of complex lines in photographs or other image representations.
- the Hough transform is thus generally a technique that can be used to extract features with a special shape within an image.
- the Hough transformation is used to extract signal edges with specified time lengths from the time signal.
- a signal edge is initially specified by its length in time.
- a signal edge would be defined by the rising edge of the sine function from 0 to 90 °.
- a signal edge could also be specified by increasing the sine function from -90 ° to + 90 °.
- the temporal length of a signal edge takes into account the sampling frequency with which the samples were generated, corresponds to a certain number of samples.
- the length of a signal edge can thus be easily determined by the specification of the number of samples that the signal edge is to include.
- a signal edge as a signal edge only if it is continuous and has a predominantly monotonic course, that is to say in the In the case of a positive signal edge it has a predominantly monotonously increasing profile.
- negative signal edges ie monotonically falling signal edges, can also be detected.
- Another criterion for the classification of signal edges is that a signal edge is only detected as a signal edge if it covers a certain level range. In order to suppress noise disturbances, it is preferred to specify a minimum level range or amplitude range for a signal edge, wherein monotonically increasing signal edges below this level range are not detected as signal edges.
- a further restriction is made for referencing audio signals to the effect that only signal edges are sought whose specified time length is greater than a minimum limit length and less than a maximum time limit length.
- the signal edge detection unit 12 thus supplies a signal edge and the time of the occurrence of the signal edge. It is irrelevant whether the time of the first sample value of the signal edge, the time of the last sample value of the signal edge or the time of any sample value is taken within the signal edge, as long as signal edges are treated equally.
- the device 14 for determining a time interval between two successive signal edges, the lengths of which are the same apart from a predetermined tolerance value examines the signal edges output by the device 12 and extracts two successive signal edges that are the same or within a specific predetermined tolerance value in the are essentially the same. If a simple sine tone is considered, a period of the sine tone is determined by the time interval between two successive z. B. given positive quarter waves.
- the device 16 for calculating a frequency value from the determined time interval is based on this. The frequency value corresponds to the inverse of the determined time interval.
- a representation of a time signal can be provided with a high temporal and at the same time frequency resolution by specifying the frequencies occurring in the time signal and specifying the times of occurrence corresponding to the frequencies. If the results of the device 16 for frequency calculation are represented graphically, a diagram according to FIG. 5 is obtained.
- FIG. 5 shows a section with a length of about 13 seconds of the clarinet quintet in A major, Larghetto, KV 581 by Wolfgang Amadeus Mozart, as it would appear at the output of the device 16 for frequency calculation.
- This section features a clarinet that plays a melody-leading solo part and an accompanying string quartet.
- the coordinate tuples shown in FIG. 5 result as they could be generated by the device 16 for frequency calculation.
- the device 18 serves to generate from the results of the device 16 a signal identifier that is cheap and suitable for a signal identifier database.
- the signal identifier is generally generated from a plurality of coordinate tuples, each coordinate tuple comprising a frequency value and an occurrence time, so that the signal identifier comprises a sequence of signal identifier values which reproduce the time profile of the time signal.
- the device 18 serves to extract the essential information from the frequency-time diagram of FIG. 5, which could be generated by the device 16, in order to generate a fingerprint of the time signal, which is compact on the one hand and, on the other hand, can differentiate the time signal from other time signals with sufficient accuracy and differentiation.
- An audio file 20 is input to an audio I / O handler as the time signal.
- the audio I / O handler 22 reads the audio file from a hard drive, for example.
- the audio data stream can also be read directly via a sound card.
- the device 22 closes the audio file again and loads the next audio file to be processed or terminates the reading process.
- the device 24 serves, on the one hand, to carry out a sampling rate conversion if necessary, or to achieve a volume modification of the audio signal.
- Audio signals are available on different media in different sampling frequencies. As has already been explained, however, the time of the occurrence of a signal edge in the audio gnal used to describe the audio signal, so that the sampling rate must be known in order to correctly detect the occurrence times of signal edges, and in addition to correctly detect frequency values.
- a sampling rate conversion can be carried out by decimation or interpolation in order to bring the audio signals of different sampling rates to the same sampling rate.
- the device 24 is therefore provided in order to carry out a sampling rate setting.
- the PCM samples are also subjected to an automatic level adjustment, which is also provided in the device 24.
- the average signal power of the audio signal is determined in the device 24 for automatic level adjustment in a look-ahead buffer.
- the audio signal section which lies between two signal power minima is multiplied by a scaling factor which is the product of a weighting factor and the quotient of the full scale and the maximum level within the segment.
- the length of the look-ahead buffer is variable.
- the audio signal which has been preprocessed in this way is then fed into the device 12, which carries out a signal edge detection, as has been described with reference to FIG. 1.
- the Hough transformation is preferably used for this.
- a circuit implementation of the Hough transformation is disclosed in WO 99/26167.
- the amplitude of a signal edge determined by the Hough transformation and the time of detection of a signal edge are then transferred to the device 14 in FIG. 1.
- the device 16 from FIG. 1 and, when a piece of music is processed accordingly, leads to the frequency-time diagram of FIG. 5, in which the frequency-time-coordinate tuples obtained are shown graphically, which by Mozart, Köchel-Directory 581.
- the signal identification generator 18 can be constructed as shown in FIG. 3.
- the device 18 is divided into a device 18a for determining the cluster areas, a device 18b for grouping, a device 18c for averaging over a group, a device 18d for interval setting, a device for quantizing 18e and finally a device 18f to get the signal identifier for the time signal.
- characteristic distribution point clouds which are referred to as clusters or clusters, are worked out in the device 18a for determining the cluster areas. This is done by deleting all isolated frequency-time tuples that exceed a predetermined minimum distance from the nearest spatial neighbor. Such isolated frequency-time tuples are, for example, the points in the top right corner of the 5.
- the pitch contour strip band d consists of clusters of a certain frequency width and length, these clusters being caused by played tones. These tones are indicated in Fig.
- the tone al has a frequency of 440 Hz.
- the tone hl has a frequency of 494 Hz.
- the tone c2 has a frequency of 523 Hz, the tone cis2 has a frequency of 554 Hz, while the tone d2 has a frequency of 587 Hz ,
- the strip width for single tones also depends on a vibrato of the musical instrument producing the single tones.
- the coordinate tuples of the pitch contour strip band are combined or grouped into a processing block to be processed separately in a time window of n samples.
- the block size can be chosen to be equidistant or variable.
- a relatively rough division can be selected, for example a one-second raster, which corresponds to a certain number of samples per block over the present sampling rate, or a smaller division.
- the grid can always be chosen so that a tone falls into the grid.
- a group or a block is then determined by the time interval between two local extreme values of the polynomial.
- This approach delivers particularly for relatively monophonic sections, relatively large groups of samples, such as occur between 6 and 12 seconds, while for relatively polyphonic intervals of the piece of music, in which the coordinate tuples are distributed over a large frequency range, such as e.g. B. at about 2 seconds in Fig. 5 or at 12 seconds of Fig. 5 smaller groups can be determined, which in turn leads to the fact that the signal recognition is carried out on the basis of relatively small groups, so that the information compression is smaller than with a fixed block formation is.
- a weighted average over all coordinate tuples present in a block is determined as required.
- the tuples outside the pitch contour strip band had already been "hidden" beforehand.
- this fading out can also be dispensed with, which leads to all coordinate tuples calculated by the device 16 being averaged by the the device 18c is carried out are taken into account.
- a jump distance for determining the center of the next, i.e. H. temporally following, group of samples determined.
- the value that has been calculated by the device 18c is quantized into non-equidistant raster values.
- the tone frequency scale As has already been explained, being divided according to the frequency range supplied by a conventional piano and deviating from 27.5 Hz (tone A2) extends to 4186 Hz (tone c5) and 88 tone stages includes. If the averaged value at the output of the device 18c lies between two adjacent semitones, it receives the value of the closest reference tone.
- the quantized values can be post-processed by the device 18f, post-processing could consist, for example, of a pitch offset correction, a transposition into another tone scale, etc.
- FIG. 4 schematically shows a device for referencing a search time signal in a database 40, the database 40 having signal identifiers of a plurality of database time signals Track_l to Track_m, which are stored in a library 42, which is preferably separate from the database 40.
- the database In order to be able to reference a time signal on the basis of the database 40, the database must first be filled, which can be achieved in a “learning” mode.
- audio files 41 are gradually fed to a vector generator 43 which has a reference identifier for each audio file and stores it in the database in such a way that it can be recognized to which audio file the signal identifier belongs, for example in library 42.
- the signal identifier MVll, ...., MVln corresponds to the time signal Track_l.
- the signal identifier MV21, .-., MV2n belongs to the time signal Track_2.
- the signal identifier MVml, ..., MVmn belongs to the time signal Track_m.
- the vector generator 43 is designed to generally perform the functions shown in FIG. 1, and is measured a preferred embodiment as shown in FIGS. 2 and 3 implemented. In the “learning” mode, the vector generator 43 gradually processes various audio files (Track_l to Track_m) in order to store signal identifiers for the time signals in the database, ie to fill the database.
- an audio file 41 is to be referenced using the database 40.
- the search time signal 41 is processed by the vector generator 43 in order to generate a search identifier 45.
- the search identifier 45 is then converted into a DNA Sequencer 46 is fed in to be compared with the reference identifiers in database 40.
- DNA sequencer 46 is also arranged to make a statement about the search time signal with regard to the plurality of database time signals from library 42. Der DNA sequencer searches the database 40 for a matching reference identifier with the search identifier 45 and transfers a pointer to the corresponding audio file in the library 42 associated with the reference identifier.
- the DNA sequencer 46 thus compares the search identifier 45 or parts thereof with the reference identifiers in the database. If the specified sequence or a partial sequence thereof is present, the associated time signal is referenced in library 42.
- the DNA sequencer 46 preferably executes a Boyer-Moore algorithm, which is described, for example, in the textbook “Algorithms on Strings, Trees and Sequences", Dan Gusfield, Cambridge University Press, 1997. According to a first alternative, exact agreement is sought Making a statement is therefore to say that the search time signal is identical to a time signal in library 42. Alternatively or in addition, the similarity of two sequences can also be achieved by using replace / insert / delete operations and a pitch offset correction (pitch offset correction) are examined.
- pitch offset correction pitch offset correction
- the database 40 is preferably structured in such a way that it is composed of the concatenation of signal identification sequences, the end of each vector signal identification of a time signal being defined by a separator, so that the search is not continued over time signal file boundaries. If several matches are found, all referenced time signals are given.
- a similarity measure can be introduced by using the Replace / Insert / Delete operations, the time signal being referenced in the library 42 which is most similar to the search time signal 41 on the basis of a predetermined similarity measure. It is further preferred to determine a similarity measure of the search audio signal to a plurality of signals in the library and then to output the n most similar sections in the library 42 in descending order.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Mobile Radio Communication Systems (AREA)
- Radar Systems Or Details Thereof (AREA)
Abstract
Description
Claims
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE10117871A DE10117871C1 (en) | 2001-04-10 | 2001-04-10 | Signal identification extraction method for identification of audio data uses coordinate points provided by frequency values and their occurence points |
DE10117871 | 2001-04-10 | ||
PCT/EP2002/002703 WO2002084539A2 (en) | 2001-04-10 | 2002-03-12 | Method and device for extracting a signal identifier, method and device for creating a corresponding database |
Publications (2)
Publication Number | Publication Date |
---|---|
EP1377924A2 true EP1377924A2 (en) | 2004-01-07 |
EP1377924B1 EP1377924B1 (en) | 2004-09-22 |
Family
ID=7681083
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP02714186A Expired - Lifetime EP1377924B1 (en) | 2001-04-10 | 2002-03-12 | Method and device for extracting a signal identifier, method and device for creating a database from signal identifiers and method and device for referencing a search time signal |
Country Status (9)
Country | Link |
---|---|
US (1) | US20040158437A1 (en) |
EP (1) | EP1377924B1 (en) |
JP (1) | JP3934556B2 (en) |
AT (1) | ATE277381T1 (en) |
AU (1) | AU2002246109A1 (en) |
CA (1) | CA2443202A1 (en) |
DE (2) | DE10117871C1 (en) |
HK (1) | HK1059492A1 (en) |
WO (1) | WO2002084539A2 (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE10232916B4 (en) * | 2002-07-19 | 2008-08-07 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus and method for characterizing an information signal |
PL1684263T3 (en) * | 2005-01-21 | 2011-03-31 | Cugate Ag | Method of generating a footprint for an audio signal |
DE102005030326B4 (en) | 2005-06-29 | 2016-02-25 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus, method and computer program for analyzing an audio signal |
DE102005030327A1 (en) * | 2005-06-29 | 2007-01-04 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus, method and computer program for analyzing an audio signal |
WO2010135623A1 (en) * | 2009-05-21 | 2010-11-25 | Digimarc Corporation | Robust signatures derived from local nonlinear filters |
DE102017213510A1 (en) * | 2017-08-03 | 2019-02-07 | Robert Bosch Gmbh | Method and apparatus for generating a machine learning system, and virtual sensor device |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR772961A (en) * | 1934-05-07 | 1934-11-09 | Method of recording music played on a keyboard instrument, and apparatus based thereon | |
US3069654A (en) * | 1960-03-25 | 1962-12-18 | Paul V C Hough | Method and means for recognizing complex patterns |
US3979557A (en) * | 1974-07-03 | 1976-09-07 | International Telephone And Telegraph Corporation | Speech processor system for pitch period extraction using prediction filters |
US4697209A (en) * | 1984-04-26 | 1987-09-29 | A. C. Nielsen Company | Methods and apparatus for automatically identifying programs viewed or recorded |
DE4324497A1 (en) * | 1992-07-23 | 1994-04-21 | Roman Koller | Remote control unit e.g. for computer network - provides on-off switching, power control, and remote operation acoustically for various electrical installations |
US5918223A (en) * | 1996-07-22 | 1999-06-29 | Muscle Fish | Method and article of manufacture for content-based analysis, storage, retrieval, and segmentation of audio information |
AU6049298A (en) * | 1997-01-31 | 1998-08-25 | T-Netix, Inc. | System and method for detecting a recorded voice |
DE19948974A1 (en) * | 1999-10-11 | 2001-04-12 | Nokia Mobile Phones Ltd | Method for recognizing and selecting a tone sequence, in particular a piece of music |
US6990453B2 (en) * | 2000-07-31 | 2006-01-24 | Landmark Digital Services Llc | System and methods for recognizing sound and music signals in high noise and distortion |
-
2001
- 2001-04-10 DE DE10117871A patent/DE10117871C1/en not_active Expired - Fee Related
-
2002
- 2002-03-12 CA CA002443202A patent/CA2443202A1/en not_active Abandoned
- 2002-03-12 AU AU2002246109A patent/AU2002246109A1/en not_active Abandoned
- 2002-03-12 WO PCT/EP2002/002703 patent/WO2002084539A2/en active IP Right Grant
- 2002-03-12 EP EP02714186A patent/EP1377924B1/en not_active Expired - Lifetime
- 2002-03-12 AT AT02714186T patent/ATE277381T1/en active
- 2002-03-12 US US10/473,801 patent/US20040158437A1/en not_active Abandoned
- 2002-03-12 DE DE50201116T patent/DE50201116D1/en not_active Expired - Lifetime
- 2002-03-12 JP JP2002582410A patent/JP3934556B2/en not_active Expired - Lifetime
-
2004
- 2004-04-02 HK HK04102412A patent/HK1059492A1/en not_active IP Right Cessation
Non-Patent Citations (1)
Title |
---|
See references of WO02084539A2 * |
Also Published As
Publication number | Publication date |
---|---|
HK1059492A1 (en) | 2004-07-02 |
DE10117871C1 (en) | 2002-07-04 |
DE50201116D1 (en) | 2004-10-28 |
WO2002084539A2 (en) | 2002-10-24 |
US20040158437A1 (en) | 2004-08-12 |
JP3934556B2 (en) | 2007-06-20 |
WO2002084539A3 (en) | 2003-10-02 |
ATE277381T1 (en) | 2004-10-15 |
CA2443202A1 (en) | 2002-10-24 |
EP1377924B1 (en) | 2004-09-22 |
JP2004531758A (en) | 2004-10-14 |
AU2002246109A1 (en) | 2002-10-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
DE10232916B4 (en) | Apparatus and method for characterizing an information signal | |
EP1368805B1 (en) | Method and device for characterising a signal and method and device for producing an indexed signal | |
DE10117870B4 (en) | Method and apparatus for transferring a music signal into a score-based description and method and apparatus for referencing a music signal in a database | |
EP1407446B1 (en) | Method and device for characterising a signal and for producing an indexed signal | |
EP2099024B1 (en) | Method for acoustic object-oriented analysis and note object-oriented processing of polyphonic sound recordings | |
EP1405222B9 (en) | Method and device for producing a fingerprint and method and device for identifying an audio signal | |
EP1371055B1 (en) | Device for the analysis of an audio signal with regard to the rhythm information in the audio signal using an auto-correlation function | |
EP1388145B1 (en) | Device and method for analysing an audio signal in view of obtaining rhythm information | |
DE10157454B4 (en) | A method and apparatus for generating an identifier for an audio signal, method and apparatus for building an instrument database, and method and apparatus for determining the type of instrument | |
DE102004028693B4 (en) | Apparatus and method for determining a chord type underlying a test signal | |
EP1377924B1 (en) | Method and device for extracting a signal identifier, method and device for creating a database from signal identifiers and method and device for referencing a search time signal | |
EP1671315B1 (en) | Process and device for characterising an audio signal | |
EP1743324B1 (en) | Device and method for analysing an information signal |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20030930 |
|
AK | Designated contracting states |
Kind code of ref document: A2 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE TR |
|
AX | Request for extension of the european patent |
Extension state: AL LT LV MK RO SI |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
RIN1 | Information on inventor provided before grant (corrected) |
Inventor name: KATAI, ANDRAS Inventor name: KAUFMANN, MATTHIAS Inventor name: UHLE, CHRISTIAN Inventor name: RICHTER, CHRISTIAN Inventor name: HIRSCH, WOLFGANG Inventor name: KLEFENZ, FRANK Inventor name: BRANDENBURG, KARLHEINZ |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
REG | Reference to a national code |
Ref country code: HK Ref legal event code: DE Ref document number: 1059492 Country of ref document: HK |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE TR |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT;WARNING: LAPSES OF ITALIAN PATENTS WITH EFFECTIVE DATE BEFORE 2007 MAY HAVE OCCURRED AT ANY TIME BEFORE 2007. THE CORRECT EFFECTIVE DATE MAY BE DIFFERENT FROM THE ONE RECORDED. Effective date: 20040922 Ref country code: IE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20040922 Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20040922 Ref country code: TR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20040922 Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20040922 |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D Free format text: NOT ENGLISH |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP |
|
GBT | Gb: translation of ep patent filed (gb section 77(6)(a)/1977) |
Effective date: 20040922 |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D Free format text: GERMAN |
|
REF | Corresponds to: |
Ref document number: 50201116 Country of ref document: DE Date of ref document: 20041028 Kind code of ref document: P |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20041222 Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20041222 Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20041222 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20050102 |
|
LTIE | Lt: invalidation of european patent or patent extension |
Effective date: 20040922 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20050312 Ref country code: LU Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20050312 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: BE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20050331 Ref country code: MC Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20050331 |
|
NLV1 | Nl: lapsed or annulled due to failure to fulfill the requirements of art. 29p and 29m of the patents act | ||
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FD4D |
|
ET | Fr: translation filed | ||
REG | Reference to a national code |
Ref country code: HK Ref legal event code: GR Ref document number: 1059492 Country of ref document: HK |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20050623 |
|
BERE | Be: lapsed |
Owner name: FRAUNHOFER-GESELLSCHAFT ZUR FORDERUNG DER ANGEWAND Effective date: 20050331 |
|
BERE | Be: lapsed |
Owner name: *FRAUNHOFER-GESELLSCHAFT ZUR FORDERUNG DER ANGEWAN Effective date: 20050331 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: PT Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20050222 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R082 Ref document number: 50201116 Country of ref document: DE Representative=s name: SCHOPPE, ZIMMERMANN, STOECKELER, ZINKLER & PAR, DE |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R081 Ref document number: 50201116 Country of ref document: DE Owner name: MUFIN GMBH, DE Free format text: FORMER OWNER: FRAUNHOFER-GESELLSCHAFT ZUR FOERDERUNG DER ANGEWANDTEN FORSCHUNG E.V., 80686 MUENCHEN, DE Effective date: 20111109 Ref country code: DE Ref legal event code: R082 Ref document number: 50201116 Country of ref document: DE Representative=s name: SCHOPPE, ZIMMERMANN, STOECKELER, ZINKLER, SCHE, DE Effective date: 20111109 Ref country code: DE Ref legal event code: R082 Ref document number: 50201116 Country of ref document: DE Representative=s name: SCHOPPE, ZIMMERMANN, STOECKELER, ZINKLER & PAR, DE Effective date: 20111109 |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: 732E Free format text: REGISTERED BETWEEN 20120105 AND 20120111 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: CA Effective date: 20120207 Ref country code: FR Ref legal event code: TP Owner name: MUFIN GMBH, DE Effective date: 20120207 |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: PC Ref document number: 277381 Country of ref document: AT Kind code of ref document: T Owner name: MUFIN GMBH, DE Effective date: 20120402 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 15 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 16 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20190322 Year of fee payment: 18 Ref country code: CH Payment date: 20190322 Year of fee payment: 18 Ref country code: GB Payment date: 20190322 Year of fee payment: 18 Ref country code: DE Payment date: 20181211 Year of fee payment: 18 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: AT Payment date: 20190328 Year of fee payment: 18 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R119 Ref document number: 50201116 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PL |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: MM01 Ref document number: 277381 Country of ref document: AT Kind code of ref document: T Effective date: 20200312 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LI Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200331 Ref country code: DE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20201001 Ref country code: AT Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200312 Ref country code: CH Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200331 Ref country code: FR Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200331 |
|
GBPC | Gb: european patent ceased through non-payment of renewal fee |
Effective date: 20200312 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200312 |