EP4097695A1 - Verfahren und vorrichtung zur erkennung von akustischen anomalien - Google Patents
Verfahren und vorrichtung zur erkennung von akustischen anomalienInfo
- Publication number
- EP4097695A1 EP4097695A1 EP21702020.5A EP21702020A EP4097695A1 EP 4097695 A1 EP4097695 A1 EP 4097695A1 EP 21702020 A EP21702020 A EP 21702020A EP 4097695 A1 EP4097695 A1 EP 4097695A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- abcd
- audio segments
- anomaly
- audio
- feature vectors
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 77
- 239000013598 vector Substances 0.000 claims abstract description 80
- 230000007774 longterm Effects 0.000 claims abstract description 15
- 238000001514 detection method Methods 0.000 claims description 20
- 238000004590 computer program Methods 0.000 claims description 15
- 230000002123 temporal effect Effects 0.000 claims description 14
- PCTMTFRHKVHKIS-BMFZQQSSSA-N (1s,3r,4e,6e,8e,10e,12e,14e,16e,18s,19r,20r,21s,25r,27r,30r,31r,33s,35r,37s,38r)-3-[(2r,3s,4s,5s,6r)-4-amino-3,5-dihydroxy-6-methyloxan-2-yl]oxy-19,25,27,30,31,33,35,37-octahydroxy-18,20,21-trimethyl-23-oxo-22,39-dioxabicyclo[33.3.1]nonatriaconta-4,6,8,10 Chemical compound C1C=C2C[C@@H](OS(O)(=O)=O)CC[C@]2(C)[C@@H]2[C@@H]1[C@@H]1CC[C@H]([C@H](C)CCCC(C)C)[C@@]1(C)CC2.O[C@H]1[C@@H](N)[C@H](O)[C@@H](C)O[C@H]1O[C@H]1/C=C/C=C/C=C/C=C/C=C/C=C/C=C/[C@H](C)[C@@H](O)[C@@H](C)[C@H](C)OC(=O)C[C@H](O)C[C@H](O)CC[C@@H](O)[C@H](O)C[C@H](O)C[C@](O)(C[C@H](O)[C@H]2C(O)=O)O[C@H]2C1 PCTMTFRHKVHKIS-BMFZQQSSSA-N 0.000 abstract 7
- 238000012544 monitoring process Methods 0.000 description 8
- 238000001228 spectrum Methods 0.000 description 5
- 230000006978 adaptation Effects 0.000 description 4
- 238000013459 approach Methods 0.000 description 4
- 238000013528 artificial neural network Methods 0.000 description 4
- 238000013145 classification model Methods 0.000 description 4
- 230000004807 localization Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 239000011521 glass Substances 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 230000001052 transient effect Effects 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 1
- 210000003484 anatomy Anatomy 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000003203 everyday effect Effects 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 231100001261 hazardous Toxicity 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- 230000007257 malfunction Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000000513 principal component analysis Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 238000009987 spinning Methods 0.000 description 1
- 238000007619 statistical method Methods 0.000 description 1
- 238000012109 statistical procedure Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/02—Mechanical actuation
- G08B13/04—Mechanical actuation by breaking of glass
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/16—Actuation by interference with mechanical vibrations in air or other fluid
- G08B13/1654—Actuation by interference with mechanical vibrations in air or other fluid using passive vibration detection systems
- G08B13/1672—Actuation by interference with mechanical vibrations in air or other fluid using passive vibration detection systems using sonic detecting means, e.g. a microphone operating in the audio frequency range
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
- G08B21/04—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
- G08B21/0438—Sensor means for detecting
- G08B21/0469—Presence detectors to detect unsafe condition, e.g. infrared sensor, microphone
Definitions
- Embodiments of the present invention relate to a method and a device for detecting acoustic anomalies. Further exemplary embodiments relate to a corresponding computer program. According to exemplary embodiments, a normal situation is recognized and anomalies are recognized in comparison to this normal situation.
- a recording typically encompasses a certain period of time which, when viewed below, is subdivided into one or more time windows. Based on this subdivision and depending on the length of the noise (cf. transient or longer, stationary sound), a noise can extend over one or more audio segments / time windows.
- an anomaly that is to say a sound deviation from the “normal acoustic state”, that is to say the amount of noises regarded as “normal”, must be recognized.
- anomalies are broken glass (burglary detection), a pistol shot (monitoring of public events) or a chainsaw (monitoring of nature reserves).
- the problem is that the sound of the anomaly (not-OK-class) is often not known or cannot be precisely defined or described (e.g. how can a broken machine sound?).
- the second problem is that novel algorithms for sound classification using deep neural networks are very sensitive to changed (and often unknown) acoustic conditions in the operational scenario.
- classification models that are trained with audio data can be achieved with a high-quality microphone, for example were recorded, with the classification of audio data which were recorded by means of a poorer microphone, only poor recognition rates.
- Possible solution approaches are in the area of "Domain Adaptation", ie the adaptation of the models or the audio data to be classified in order to achieve greater robustness in recognition, but in practice it is often logistically difficult and too expensive to make representative audio recordings Record at the later place of use of an audio analysis system and then annotate them with regard to the contained sound events.
- the third problem of audio analysis of environmental noises lies in concerns about data protection, since classification methods can theoretically also be used to recognize and transcribe speech signals (e.g. when recording a conversation near the audio sensor).
- a classification model based on machine learning algorithms can be trained to recognize certain noise classes by means of supervised learning.
- Current studies show that neural networks in particular are very sensitive to changed acoustic conditions and that additional adaptation of classification models to the respective acoustic situation of the application has to be carried out.
- the object of the present invention is to create a concept for the detection of anomalies which optimizes the learning behavior and which enables a reliable and precise detection of anomalies.
- Embodiments of the present invention provide a method for detecting acoustic anomalies.
- the method comprises the steps of obtaining a long-term recording with a plurality of first audio segments assigned to respective first time windows and analyzing the plurality of first audio segments in order to each of the A plurality of the first audio segments have a first feature vector describing the respective first audio segment, such as e.g. B. to obtain a spectrum for the audio segment (time-frequency spectrum) or an audio fingerprint with certain characteristics for the audio segment.
- the result of the analysis of a long-term recording subdivided into a plurality of time windows is a plurality of first (one- or multi-dimensional) feature vectors for the plurality of first audio segments (assigned to the corresponding times / windows of the long-term recording) that correspond to the Represent "normal state".
- the method comprises further steps of hardening a further recording with one or more second audio segments assigned to respective second audio windows and analyzing the one or more second audio segments in order to obtain one or more feature vectors describing the one or more second audio segments.
- the result of the second part of the method is, for example, a large number of second feature vectors (e.g. with corresponding points in time of the further recording).
- the one or more second feature vectors are compared with the plurality of first feature vectors (for example by comparing the identities or similarities or by recognizing a sequence) in order to recognize at least one anomaly.
- the recognition of different forms of anomalies would be conceivable, namely a sound anomaly (i.e. recognition of the first appearance of a previously unheard sound), a temporal anomaly (e.g. changed repetition pattern of a sound that has already been heard) or a spatial anomaly (occurrence of a sound that has already been heard in a previously unknown spatial position).
- Embodiments of the present invention are based on the knowledge that a “normal acoustic state” and “normal noises” can be learned independently solely through a long-term sound analysis (phase 1 of the method comprising the steps of obtaining a long-term recording and analyzing the same). This means that this long-term sound analysis results in an independent or autonomous adaptation of an analysis system to a specific acoustic scene. No annotated training data (recording + semantic class annotation) are required, which saves a great deal of time, effort and costs.
- this acoustic "normal state” or the "normal” noises have been recorded, the current noise environment can be carried out in a subsequent analysis phase (phase 2 with the steps of obtaining a further recording and analyzing it).
- phase 1 involves learning a model using the normal background noise on the basis of a statistical procedure or machine learning, whereby this model then allows (in phase 2) to compare currently recorded background noise with it with regard to their degree of novelty (probability of an anomaly).
- Another advantage of this approach is that the privacy of people who may be in the direct vicinity of the acoustic sensors is protected.
- the multitude of first audio segments in themselves and / or in their order describes this normal situation.
- the multiplicity of the first audio segments represents a kind of reference in itself and / or in their combination.
- the aim of the method is to identify anomalies in comparison to this normal situation.
- the result of the clustering described above is a description of the reference on the basis of first audio segments.
- the second audio segments are then compared individually or in their combination (that is, sequence) with the reference in order to represent the anomaly.
- the anomaly is a deviation of the current acoustic situation described by the second feature vectors from the reference described by the first feature vectors.
- the first feature vectors alone or in their combination represent a reference mapping of the normal state, while the second feature vectors individually or in their combination describe the current acoustic situation, so that in the step 126 the anomaly can be recognized in the form of a deviation of the description of the current acoustic situation (cf. second feature vectors) from the reference (cf. first feature vectors).
- the anomaly is thus defined in that at least one of the second acoustic feature vectors deviates from the sequence of the first acoustic feature vectors. Possible deviations can be: aural anomalies, temporal anomalies and spatial anomalies.
- a large number of first audio segments are recorded by phase 1, which are also referred to below as “normal” or “normal” noises / audio segments. According to exemplary embodiments knowing these “normal” audio segments makes it possible to recognize a so-called aural anomaly. In this case, the substep of identifying a second feature vector, which differs from the analyzed first feature vectors, is then carried out.
- the method comprises the substep of identifying a repetition pattern in the plurality of first time windows. Repeating audio segments are identified and the resulting pattern is determined. According to exemplary embodiments, the identification takes place on the basis of repeated, identical or similar first feature vectors belonging to different first audio segments. In accordance with exemplary embodiments, identical and similar first feature vectors or first audio segments can also be grouped into one or more groups during identification.
- the method includes the recognition of a sequence of first feature vectors belonging to the first audio segments or the recognition of a sequence of groups of identical or similar first feature vectors or first audio segments.
- the basic steps therefore advantageously make it possible to recognize normal noises or to recognize normal audio objects.
- the combination of these normal audio objects in terms of time in a specific sequence or a specific repetition pattern then represents, in total, a normal acoustic state.
- this method then enables the sub-step of comparing the repeat pattern of the first audio segments and / or sequence in the first audio segments with the repeat pattern of the second audio segments and / or sequence in the second audio segments to take place. This comparison enables the detection of a temporal anomaly.
- the method can include the step of determining a respective position for the respective first audio segments. According to an exemplary embodiment, it is also possible to determine the respective position for the respective second audio segments are made. According to an exemplary embodiment, this then enables the detection of a spatial anomaly to be undertaken through the substep of comparing the position assigned to the respective first audio segments with the position assigned to the corresponding respective second audio segment.
- At least two microphones are used for spatial localization, for example, while one microphone is sufficient for the other two types of anomaly.
- each feature vector can each have one dimension or several dimensions for the different audio segments.
- a possible implementation of a feature vector would be a time-frequency spectrum, for example.
- the dimensional space can also be reduced.
- the method includes the step of reducing the dimensions of the feature vector.
- the method can have the step of determining a probability of occurrence of the respective first audio segment and of giving up the probability of occurrence together with the respective first feature vector.
- the method can have the step of determining a probability of occurrence of the respective first audio segment and outputting the probability of occurrence with the respective first feature vector and an associated first time window.
- the method can also run in a computer-implemented manner.
- the method has a computer program with a program code for carrying out the method.
- FIG. 1 For exemplary embodiments, relate to a device with an interface and a processor.
- the interface is used to obtain a long-term recording with a multiplicity of first audio segments assigned to respective first time windows and to obtain a further recording with one or more second audio segments assigned to respective second time windows.
- the processor is designed to handle the plurality of the first audio segments in order to obtain a first feature vector describing the respective first audio segment for each of the plurality of first audio segments.
- the processor is designed to analyze the one or more second audio segments in order to obtain one or more feature vectors describing the one or more second audio segments.
- the processor is designed to match the one or more second feature vectors with the plurality of first feature vectors in order to identify at least one anomaly.
- the device comprises a receiving unit connected to the interface, such as, for. B. a microphone or a microphone array.
- the microphone array advantageously enables a position to be determined, as has already been explained above.
- the device comprises an output interface for outputting the above-explained probability of occurrence.
- FIG. 1 shows a schematic flow diagram to illustrate the method according to a basic exemplary embodiment
- FIG. 3 shows a schematic block diagram to illustrate a device according to a further exemplary embodiment.
- 1 shows a method 100 which is divided into two phases 110 and 120.
- Step 112 includes a long Time recording of the normal acoustic state in the application scenario.
- the analysis device 10 (cf. FIG. 3) is set up in the target environment, so that a long-term recording 113 of the normal state is recorded.
- This long-term recording can, for example, last for 10 minutes, 1 hour or 1 day (generally more than 1 minute, more than 30 minutes, more than 5 hours or more than 24 hours and / or up to 10 hours, up to 1 day, up to 3 days or up to 10 days (including the time window defined by the upper and lower).
- This long-term recording 113 is then subdivided, for example.
- the subdivision can be in equally long time ranges, such as B. 1 second or 0.1 seconds or dynamic time ranges.
- Each time range comprises an audio segment.
- step 114 which is generally referred to as analyzing, these audio segments are examined separately or in combination.
- a so-called feature vector 115 (first feature vectors) is determined for each audio segment during the analysis.
- each feature vector 115 “codes” the sound at a specific point in time.
- Feature vectors 115 can be determined, for example, by an energy spectrum for a specific frequency range or generally a time-frequency spectrum.
- step 114 typical or dominant noises can then optionally also be identified by means of unsupervised learning processes (for example clustering).
- unsupervised learning processes for example clustering
- time segments or audio segments are grouped which here express similar feature vectors 115 and which accordingly have a similar sound.
- No semantic classification of a sound eg “car” or “airplane”
- unsupervised learning takes place on the basis of frequencies of repetitive or similar audio segments.
- an unsupervised learning of the temporal sequence and / or typical repetition patterns of certain noises to take place in step 114.
- the result of the clustering is a compilation of audio segments or noises that are normal or typical for this area. For example, a probability of occurrence can also be assigned to each audio segment. Furthermore, a Repetition patterns or a sequence, that is to say a combination of several audio segments, can be identified that is typical or normal for the current environment. For this purpose, different audio segments can also be assigned a probability to each grouping, each repetition pattern or each sequence.
- phase 120 has the three basic steps 122 and 124 and 126.
- an audio recording 123 is again recorded. This is typically significantly shorter in comparison to the audio recording 113. This audio recording is shorter in comparison to audio recording 113, for example. However, it can also be a continuous audio recording.
- This audio recording 123 is then analyzed in a subsequent step 124. The content of this step is comparable to step 114. In this case, the digital audio recording 123 is again converted into feature vectors. If these second feature vectors 125 are now available, they can be compared with the feature vectors 115.
- step 126 The comparison is made in step 126 with the aim of determining anomalies. Very similar feature vectors and very similar sequences of feature vectors indicate that there is no anomaly. Deviations from previously determined patterns (repeated patterns, typical sequences, etc.) or deviations from the previously determined audio segments identified by other / new feature vectors indicate an anomaly. These are recognized in step 126. In step 126, different types of anomalies can be identified. These are for example:
- a probability can be output for each of the three types of anomaly at time X. This is illustrated by arrows 126z, 126k and 126r (one arrow per anatomy type) in FIG.
- threshold values can be defined in accordance with exemplary embodiments when feature vectors are similar or when groups of feature vectors are similar, so that the result then also has one
- threshold for an anomaly.
- This application of threshold values can also be linked to the output of the probability distribution or appear in combination with it, e. B. to enable more accurate temporal detection of anomalies.
- step 114 in the adjustment phase 110 can also have an unsupervised learning of typical spatial positions and / or movements of specific noises.
- the microphone 18 shown in FIG. 3 two microphones or a microphone array with at least two microphones are present.
- a spatial localization of the current dominant sound sources / audio segments is then also possible in the second phase 120 by means of a multi-channel recording.
- the technology on which this is based can, for example, be beamforming.
- 2a illustrates the temporal anomaly.
- audio segments ABC for both phase 1 and phase 2 are plotted along the time axis t; in phase 1 it was recognized that a normal situation or normal sequence exists such that the audio segments ABC appear in the sequence ABC. A repetition pattern was recognized for one of them, which can be followed by another group ABC after the first group ABC.
- this pattern ABCABC is recognized in phase 2, it can be assumed that no anomaly or at least no temporal anomaly is present. If, however, the pattern ABCAABC shown here is recognized, then there is a temporal anomaly, since a further audio segment A is arranged between the two groups ABC. This audio segment A or abnormal audio segment A is provided with a double frame.
- FIG. 2b A sound anomaly is further illustrated in FIG. 2b.
- the audio segments ABCABC were again recorded along the time axis t (cf. FIG. 2a).
- the acoustic anomaly during detection is shown by the fact that a further audio segment, here audio segment D, appears in phase 2.
- This audio segment D has an increased length, e.g. B. over two time ranges and is therefore illustrated as DD.
- the acoustic anomaly is provided with a double frame in the order of species of the audio segment.
- This sonic anomaly can be, for example, a sound that was never heard during the learning phase.
- a thunder can be present here, which differs in terms of loudness / intensity and in terms of length from the previous elements ABC.
- a local anomaly is illustrated.
- two audio segments A and B were recognized at two different positions, position 1 and position 2.
- both elements A and B were recognized, and it was established through localization that both audio segment A and audio segment B are at position 1.
- the presence of audio segment B at position 1 represents a spatial anomaly.
- the device 10 essentially comprises the input interface 12, such as, for. B. a microphone interface and a processor 14.
- the processor 14 receives the one or more (simultaneously present) audio signals from the microphone 18 or the microphone array 18 'and analyzes them. To this end, it essentially carries out steps 114, 124 and 126 explained in connection with FIG. 1.
- the result to be output (cf. output interface 16) is a set of feature vectors that represent the normal state or, in phase 2, an output of the anomalies recognized, e.g. B. assigned to a specific type and / or assigned to a specific point in time.
- a probability of anomalies or a probability of anomalies at specific times or, in general, a probability of feature vectors at specific times can take place.
- the device 10 or the audio system is designed in accordance with exemplary embodiments (simultaneously) different types of anomalies, e.g. B. to recognize at least two anomalies.
- the following areas of application are conceivable:
- aspects have been described in connection with a device, it goes without saying that these aspects also represent a description of the corresponding method, so that a block or a component of a device can also be used as a corresponding method step or as a feature of a method step understand is. Analogously, aspects that have been described in connection with or as a method step also represent a description of a corresponding block or details or features of a corresponding device.
- Some or all of the method steps can be carried out by a hardware apparatus (or under Using a hardware device) such as a microprocessor, a programmable computer ter or an electronic circuit. In some exemplary embodiments, some or more of the most important process steps can be carried out by such an apparatus.
- exemplary embodiments of the invention can be implemented in hardware or in software.
- the implementation can be implemented using a digital storage medium such as a floppy disk, DVD, Blu-ray disk, CD, ROM, PROM, EPROM, EEPROM or FLASH memory, hard disk or other magnetic or optical Memory are carried out on the electronically readable control signals are stored, which can interact with a programmable computer system or cooperate in such a way that the respective method is carried out.
- the digital storage medium can therefore be computer-readable.
- Some exemplary embodiments according to the invention thus include a data carrier which has electronically readable control signals which are able to interact with a programmable computer system in such a way that one of the methods described herein is carried out.
- exemplary embodiments of the present invention can be implemented as a computer program product with a program code, the program code being effective to carry out one of the methods when the computer program product runs on a computer.
- the program code can, for example, also be stored on a machine-readable carrier.
- exemplary embodiments include the computer program for performing one of the methods described herein, the computer program being stored on a machine-readable carrier.
- an exemplary embodiment of the method according to the invention is thus a computer program which has a program code for performing one of the methods described herein when the computer program runs on a computer.
- a further exemplary embodiment of the method according to the invention is thus a data carrier (or a digital storage medium or a computer-readable medium) on which the computer program for performing one of the methods described herein is recorded.
- the data carrier, the digital storage medium or the computer-readable medium are typically tangible and / or non-transitory or non-transitory.
- a further exemplary embodiment of the method according to the invention is thus a data stream or a sequence of signals which represents or represents the computer program for carrying out one of the methods described herein.
- the data stream or the sequence of signals can, for example, be configured to be transferred via a data communication connection, for example via the Internet.
- Another exemplary embodiment comprises a processing device, for example a computer or a programmable logic component, which is configured or adapted to carry out one of the methods described herein.
- a processing device for example a computer or a programmable logic component, which is configured or adapted to carry out one of the methods described herein.
- Another exemplary embodiment comprises a computer on which the computer program for performing one of the methods described herein is installed.
- a further exemplary embodiment according to the invention comprises a device or a system which is designed to transmit a computer program for carrying out at least one of the methods described herein to a receiver.
- the transmission can take place electronically or optically, for example.
- the receiver can be, for example, a computer, a mobile device, a storage device or a similar device.
- the device or the system can comprise, for example, a file server for transmitting the computer program to the recipient.
- a programmable logic component for example a field-programmable gate array, an FPGA
- a field-programmable gate array can interact with a microprocessor in order to carry out one of the methods described herein.
- the methods in some exemplary embodiments are implemented by a any hardware device performed. This can be hardware that can be used universally, such as a computer processor (CPU), or hardware specific to the method, such as an ASIC, for example.
- the devices described herein can be implemented, for example, using a hardware apparatus, or using a computer, or using a combination of a hardware apparatus and a computer.
- the devices described herein, or any components of the devices described herein, can be implemented at least partially in hardware and / or in software (computer program).
- the methods described herein can be implemented, for example, using a hardware apparatus, or using a computer, or using a combination of a hardware apparatus and a computer.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Gerontology & Geriatric Medicine (AREA)
- Business, Economics & Management (AREA)
- Emergency Management (AREA)
- Measurement Of Mechanical Vibrations Or Ultrasonic Waves (AREA)
- Testing And Monitoring For Control Systems (AREA)
- Emergency Alarm Devices (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102020200946.5A DE102020200946A1 (de) | 2020-01-27 | 2020-01-27 | Verfahren und Vorrichtung zur Erkennung von akustischen Anomalien |
PCT/EP2021/051804 WO2021151915A1 (de) | 2020-01-27 | 2021-01-27 | Verfahren und vorrichtung zur erkennung von akustischen anomalien |
Publications (2)
Publication Number | Publication Date |
---|---|
EP4097695A1 true EP4097695A1 (de) | 2022-12-07 |
EP4097695B1 EP4097695B1 (de) | 2024-02-21 |
Family
ID=74285498
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP21702020.5A Active EP4097695B1 (de) | 2020-01-27 | 2021-01-27 | Verfahren und vorrichtung zur erkennung von akustischen anomalien |
Country Status (4)
Country | Link |
---|---|
US (1) | US20220358952A1 (de) |
EP (1) | EP4097695B1 (de) |
DE (1) | DE102020200946A1 (de) |
WO (1) | WO2021151915A1 (de) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114220457A (zh) * | 2021-10-29 | 2022-03-22 | 成都中科信息技术有限公司 | 双通道通信链路的音频数据处理方法、装置及存储介质 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2944903B1 (fr) * | 2009-04-24 | 2016-08-26 | Thales Sa | Systeme et methode pour detecter des evenements audio anormaux |
DE102012211154B4 (de) * | 2012-06-28 | 2019-02-14 | Robert Bosch Gmbh | Überwachungssystem, Freiflächenüberwachung sowie Verfahren zur Überwachung eines Überwachungsbereichs |
FR2994495B1 (fr) * | 2012-08-10 | 2015-08-21 | Thales Sa | Procede et systeme pour detecter des evenements sonores dans un environnement donne |
DE102014012184B4 (de) * | 2014-08-20 | 2018-03-08 | HST High Soft Tech GmbH | Vorrichtung und Verfahren zur automatischen Erkennung und Klassifizierung von akustischen Signalen in einem Überwachungsbereich |
US10134422B2 (en) * | 2015-12-01 | 2018-11-20 | Qualcomm Incorporated | Determining audio event based on location information |
DE102017010402A1 (de) * | 2017-11-09 | 2019-05-09 | Guido Mennicken | Automatisiertes Verfahren zur Überwachung von Waldgebieten auf Rodungsaktivitäten |
DE102017012007B4 (de) | 2017-12-22 | 2024-01-25 | HST High Soft Tech GmbH | Vorrichtung und Verfahren zur universellen akustischen Prüfung von Objekten |
DE102018211758A1 (de) * | 2018-05-07 | 2019-11-07 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung, verfahren und computerprogramm zur akustischen überwachung eines überwachungsbereichs |
-
2020
- 2020-01-27 DE DE102020200946.5A patent/DE102020200946A1/de active Pending
-
2021
- 2021-01-27 EP EP21702020.5A patent/EP4097695B1/de active Active
- 2021-01-27 WO PCT/EP2021/051804 patent/WO2021151915A1/de active Search and Examination
-
2022
- 2022-07-26 US US17/874,072 patent/US20220358952A1/en active Pending
Also Published As
Publication number | Publication date |
---|---|
WO2021151915A1 (de) | 2021-08-05 |
DE102020200946A1 (de) | 2021-07-29 |
EP4097695B1 (de) | 2024-02-21 |
US20220358952A1 (en) | 2022-11-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3317878A1 (de) | Verfahren und vorrichtung zum erzeugen einer datenbank | |
DE69010193T2 (de) | Überwachung. | |
DE202017102381U1 (de) | Vorrichtung zum Verbessern der Robustheit gegen "Adversarial Examples" | |
DE112020004052T5 (de) | Sequenzmodelle zur audioszenenerkennung | |
WO2005111598A1 (de) | Vorrichtung und verfahren zur beurteilung einer güteklasse eines zu prüfenden objekts | |
EP3291234B1 (de) | Verfahren zum beurteilen einer qualität eines stimmeinsatzes eines sprechenden | |
DE102014012184A1 (de) | Vorrichtung und Verfahren zur automatischen Erkennung und Klassifizierung von akustischen Signalen in einem Überwachungsbereich | |
EP3977430A1 (de) | Verfahren und vorrichtung zur detektion von rauch | |
EP4097695B1 (de) | Verfahren und vorrichtung zur erkennung von akustischen anomalien | |
DE102020209446A1 (de) | Computerimplementiertes Verfahren und Computerprogramm zum maschinellen Lernen einer Robustheit eines akustischen Klassifikators, akustisches Klassifikationssystem für automatisiert betreibbare Fahrsysteme und automatisiert betreibbares Fahrsystem | |
DE102018205561A1 (de) | Vorrichtung zur Klassifizierung von Signalen | |
WO2022013045A1 (de) | Verfahren zum automatischen lippenlesen mittels einer funktionskomponente und zum bereitstellen der funktionskomponente | |
DE60025333T2 (de) | Sprachdetektion mit stochastischer konfidenzmassbewertung des frequenzspektrums | |
EP2483834A1 (de) | Verfahren und vorrichtung zum erkennen einer fehldetektion eines objekts in einem bild | |
DE102019213697A1 (de) | Verfahren zum Erkennen einer Annäherung und/oder Entfernung eines Einsatzfahrzeugs relativ zu einem Fahrzeug | |
WO2022180218A1 (de) | Vorrichtung zur verarbeitung von mindestens einem eingangsdatensatz unter verwendung eines neuronalen netzes sowie verfahren | |
DE102021207849A1 (de) | Verfahren zum Nachtrainieren einer Videoüberwachungsvorrichtung, Computerprogramm, Speichermedium und Videoüberwachungsvorrichtung | |
DE102020200847A1 (de) | Verfahren und Vorrichtung zur Objektidentifikation basierend auf Sensordaten | |
DE102020213289A1 (de) | Bildverarbeitungssystem | |
DE102019207700A1 (de) | Klassifikationsvorrichtung zur Objektdetektion in Umfeldsensordaten und Verfahren | |
DE102023200017B3 (de) | Verfahren zur Fehlererkennung in Montage- und Instandhaltungsprozessen | |
DE112013004687T5 (de) | System und Verfahren zum Verarbeiten von Ereignissen in einer Umgebung | |
DE102016213807A1 (de) | Konzept zum Überwachen eines Parkplatzes für Kraftfahrzeuge | |
EP1393302B1 (de) | Verfahren und vorrichtung zur automatischen differenzierung und/oder detektion von akustischen signalen | |
DE102020202603A1 (de) | Vorrichtung und Verfahren zum Erkennen eines charakteristischen Signals im Umfeld eines Fahrzeugs |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: UNKNOWN |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20220724 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
DAV | Request for validation of the european patent (deleted) | ||
DAX | Request for extension of the european patent (deleted) | ||
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: GRANT OF PATENT IS INTENDED |
|
INTG | Intention to grant announced |
Effective date: 20230913 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
P01 | Opt-out of the competence of the unified patent court (upc) registered |
Effective date: 20231212 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE PATENT HAS BEEN GRANTED |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R081 Ref document number: 502021002760 Country of ref document: DE Owner name: FRAUNHOFER-GESELLSCHAFT ZUR FOERDERUNG DER ANG, DE Free format text: FORMER OWNER: ANMELDERANGABEN UNKLAR / UNVOLLSTAENDIG, 80297 MUENCHEN, DE |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D Free format text: NOT ENGLISH |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 502021002760 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D Free format text: LANGUAGE OF EP DOCUMENT: GERMAN |
|
REG | Reference to a national code |
Ref country code: LT Ref legal event code: MG9D |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: MP Effective date: 20240221 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240621 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240522 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: HR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 Ref country code: RS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240521 Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: RS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240521 Ref country code: NO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240521 Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240621 Ref country code: HR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240522 Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: PL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240621 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240621 Ref country code: PL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 Ref country code: LV Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SM Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240221 |