CN108028048B - Method and apparatus for correlating noise and for analysis - Google Patents

Method and apparatus for correlating noise and for analysis Download PDF

Info

Publication number
CN108028048B
CN108028048B CN201680048720.2A CN201680048720A CN108028048B CN 108028048 B CN108028048 B CN 108028048B CN 201680048720 A CN201680048720 A CN 201680048720A CN 108028048 B CN108028048 B CN 108028048B
Authority
CN
China
Prior art keywords
location
ambient noise
noise
determining
parameters
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201680048720.2A
Other languages
Chinese (zh)
Other versions
CN108028048A (en
Inventor
托马斯·施波雷尔
托比亚斯·克劳斯
尤迪特·利贝特劳
萨拉·开普林格
汉纳·卢卡舍维奇
迪特马尔·开普林格
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fraunhofer Gesellschaft zur Forderung der Angewandten Forschung eV
Original Assignee
Fraunhofer Gesellschaft zur Forderung der Angewandten Forschung eV
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fraunhofer Gesellschaft zur Forderung der Angewandten Forschung eV filed Critical Fraunhofer Gesellschaft zur Forderung der Angewandten Forschung eV
Publication of CN108028048A publication Critical patent/CN108028048A/en
Application granted granted Critical
Publication of CN108028048B publication Critical patent/CN108028048B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/61Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/632Query formulation
    • G06F16/634Query by example, e.g. query by humming
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/018Audio watermarking, i.e. embedding inaudible data in the audio signal
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0216Noise filtering characterised by the method used for estimating noise
    • G10L21/0232Processing in the frequency domain
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching

Abstract

Embodiments of the present invention provide a method for correlating noise of at least one signal class (e.g., interference noise) of a plurality of signal classes (e.g., interference noise and non-interference noise). The method comprises the following steps: "receiving the ambient noise" and "establishing whether the ambient noise or a set of parameters derived from the ambient noise satisfies a predefined rule describing a signal class of a plurality of signal classes". Starting from this, the following steps are carried out: "logging meets a predefined rule", "recording the ambient noise received for the migration time window", "deriving a parameter set from the ambient noise of the migration time window and storing the parameter set", or "issuing an activation signal for another device to associate the noise".

Description

Method and apparatus for correlating noise and for analysis
Technical Field
Embodiments of the present invention relate to a method and apparatus for associating/relating noise to/from at least one signal class of a plurality of signal classes, and to a method and apparatus for analysing noise of at least one signal class of a plurality of signal classes.
Background
For example, noise may be subdivided into signal classes, such as interference noise and non-interference noise. For example, a subdivision of more interfering noise and less interfering noise is also conceivable.
It is not always easy to classify interference noise. It is important to know that there are different factors that affect whether the noise is perceived as interference noise. Birds are not subjectively perceived as disturbing noise even when their chirps are large (objective measurable parameter: sound pressure level) and significantly different from other environmental noises (objective measurable parameter: dynamics). However, a significantly quieter aircraft pass will be perceived by more test personnel as disturbing noise than the bird just mentioned.
The result here is that, for example, using current methods, in order to allow a predictive noise assessment, the assessment must be left to the test person when it comes to interfering noise investigation environments, such as the health area of a hotel, hotel or workplace.
For example, a purely automated evaluation with respect to absolute loudness or volume or increase with respect to level is used as a first cue, but not sufficient for a final evaluation. Accordingly, improved methods are needed.
Disclosure of Invention
Main aspects of the invention
It is an object of another aspect to provide a concept that allows for identifying and analyzing signal classes of noise, such as interference noise and subjectively perceived interference noise.
This object is achieved by the independent claims.
Embodiments of the present invention provide a method for correlating noise of at least one signal class (e.g., interference noise) of a plurality of signal classes (e.g., interference noise and non-interference noise). The method comprises the steps of "receiving ambient noise" and "establishing whether the ambient noise or a set of parameters derived from the ambient noise satisfies a predefined rule describing a signal class of a plurality of signal classes". Thereby, the following steps are performed: "logging in that a predefined rule has been met", "recording the ambient noise received for the migration time window", "deriving a set of parameters from the ambient noise of the migration time window and storing the set of parameters", or "issuing an activation signal for another device to associate noise of at least one signal class".
Embodiments of this aspect are based on the following findings: from the database as determined for the corresponding other aspect (see below), it is possible to identify the presence of subjectively perceived disturbing noise, or to associate noise with a category, e.g. by comparing the current noise environment with noise from the database or parameters obtained from or stored in the database, such as audio fingerprints. The method can be performed in an automated manner and the stored database alone can be used to allow the assessment of the predicted noise conditions (bird singing versus air conditioning) without any subjective assessment by humans.
Identifying a rule match may illustratively be done by comparing the ambient noise to previously buffered ambient noise, or by comparing a currently derived parametric data set (audio fingerprint) to a previously determined parametric data set, or by deriving a psycho-acoustic parameter and comparing it to its predetermined threshold.
Another embodiment relates to an apparatus for correlating noise of at least one of a plurality of signal classes. The device comprises a microphone for continuously listening to the current ambient noise, a processor for comparing the current ambient noise with data stored in a database (recording the interference noise or parameters describing the interference noise) and an interface for outputting information once the interference noise has been identified in the current environment. Here, the data (e.g. a previously determined recording) and the previously determined audio fingerprint or the previously established threshold value of the psychoacoustic parameter may be stored internally or, according to a further embodiment, read externally using, for example, a database that has been determined according to further aspects.
From these identified objective interference noise or signal classes, the information can be further processed individually or in combination with an indication of time, an indication of location or classification of the interference noise into one of the classes (corresponding interference group: slight interference, high interference). According to a preferred embodiment, this information is output to an external database.
Since in this embodiment only an evaluation of the location or the one location is provided, or according to further embodiments it is also conceivable to extend the evaluation to several locations, such as several locations in a space or an outdoor environment, i.e. several neighboring locations (e.g. distributed over a city). Thus, a further embodiment provides a method wherein the steps of "recording", "comparing" and "outputting" are received for two adjacent positions. When there is information of two adjacent locations, for example, a relation between the recordings of the first location and the second location may be determined in order to determine a movement, a spatial extension or a direction of the subjectively perceived disturbing noise.
According to another embodiment, similar to identifying the interference noise, it is also conceivable to identify a different sequence (e.g. control instructions) which is used to output a corresponding control signal. Here, the recording associated with the control signal may be a voice command, or as previously described, a sound signal classified as interference noise. For example, the control signal is output by a device that itself performs the method, for example to start recording, or an external device, such as another device disposed at a different location that is switched to the recording mode by the control signal.
According to further embodiments, the device outlined above may further comprise a communication interface for communicating with a database to read previously determined interference noise or parameters, or for outputting information about interference noise. According to a further embodiment, it is also possible for a device to communicate with another device using the communication interface, so that interference noise can be obtained and/or analyzed for two adjacent locations.
Embodiments of sub-aspects provide a method for analyzing noise of a signal class. The method comprises the step of continuously recording the current ambient noise at a first location and a second location. Recording here in turn means recording the ambient noise directly or deriving it from a set of parameters related to the ambient noise, such as an audio fingerprint or psycho-acoustic parameters. In addition, for each recording, a comparison with a previously obtained recording of the subjectively perceived interference noise or with parameters describing the interference noise is performed in order to identify the interference noise for each position (first position and second position). The relationship between the recordings can be determined from the two recordings (first recording and second recording) comprising one interference noise at different positions, so that the generated interference noise can be analyzed more accurately, for example with respect to its position, extension or movement.
Embodiments of this aspect are based on the following findings: using two recorded relations of the same interference noise at two different locations it is possible to obtain extended information about the interference noise itself. Here, the interference noise in the respective environment (i.e. at the first and second locations) is first identified and, when identified, correlated with each other. Advantageously, it is possible here to obtain information about the movement of the interference noise or the spreading of the interference noise or the propagation direction of the interference noise. In addition, it is also possible to distinguish between local interference noise (i.e. at only one location) and global events (i.e. events occurring at several locations). Using this method it is possible to identify the propagation of characteristic noise events and their movements.
According to an embodiment, the step of determining the relationship between the first record and the second record is done by analyzing a level difference between the first record and the second record. Alternatively or in addition, it is also possible to establish a time offset in the step of determining the relationship, i.e. a delay time or run time offset between events in two recordings established at two different locations. In addition, the two recordings can also be evaluated with respect to differences in frequency and hall effect. Using all these analysis parameters it is possible to determine the distance between the noise source and the recording location, since the sound generally decreases with increasing distance and/or there is a frequency shift such that the upper frequencies are cancelled out.
According to a further embodiment, the method comprises analyzing the audio event or the respective source with respect to the distance between the first location and the second location, analyzing with respect to the movement of the subjective disturbance noise source and/or comparing with respect to the number of subjective disturbance noise sources. For example, the three analyses are based on evaluating the relationship between the first record and the second record, i.e. comparing the above-mentioned factors.
It is to be mentioned in this respect that the migration time window is preferably used for continuous recording. Further, as in the above-described aspect, it is also conceivable to read the noise to be compared from the outside.
It is noted here that the method can of course be extended to the third position.
In an embodiment according to this aspect, when the interfering signal has been determined at the first location, recording may be started at the second location to allow temporal analysis of the propagating interfering signal.
Another embodiment relates to a system for analyzing a signal of a signal class. The system comprises two units with one microphone, each for continuously recording the current ambient noise. The two units may be positioned at different locations, such as adjacent locations. Here, "recording" again means directly recording the ambient noise and deriving the ambient noise from the parameters (e.g. audio fingerprint). Furthermore, the system comprises at least one processor, which may be integrated in the first unit or the second unit and which is configured to identify noise by comparing the first recording and the second recording of the first and second unit with previously obtained audio fingerprints of signals of at least one recording/signal class or parameters of signals describing signal classes. Additionally, the processor is configured to establish a relationship between the first record and the second record.
According to an embodiment, the two units may be connected to each other via a communication interface, such as a radio interface.
According to a further embodiment, a computer program for performing one of the methods described above is provided.
Further developments are defined in the dependent claims.
Additional aspects
Embodiments of the present invention provide a method for generating a database from buffered records. The method comprises the steps of "receiving ambient noise", which illustratively comprises interference noise, and "buffering ambient noise for a transition time window", such as 30 or 60 seconds, for example or preferably more than 5 seconds. Alternatively, it is also conceivable that the method comprises the steps of "deriving a parameter set relating to the ambient noise" and "buffering a parameter set of the migration time window". Buffered ambient noise or buffered parameter sets are commonly referred to as recordings. In addition, the method comprises the step of "obtaining a signal" identifying a signal class (such as interference noise) of a plurality of signal classes (interference noise and non-interference noise) in the ambient noise. The third basic step is to "store the buffered record in response to a signal" in a memory such as an internal or external memory. These steps of obtaining and storing are repeated to build up a database comprising a plurality of buffered records of the same signal class.
Embodiments of the present invention are based on the following findings: using a device that continuously records and stores the relevant positions in the environment, a database may be built up in which the recordings or characteristics of the recordings (such as audio fingerprints or psycho-acoustic parameters) are stored, enabling such sound sequences to be identified at a later point in time. Here, the concept assumes, for example, that the step of "identifying subjective interference noise or a class of noise" is performed by a person using a button or key or another input interface to identify or mark the interference noise or signal class. This signal is used as an indicator to cut off the sequence or extract features from the current continuous recording and store them in a memory with a database to be formed. It is thus possible to easily build a library of interference noise or classifiers for unambiguously correlating sound description parameters, which then allows for a prediction of the subjective noise perception.
According to an embodiment, subjective disturbing noise may be described by parameters including individual parameters like volume, dynamic, range, dynamic increase, spectrum, monotone or repeated characters (e.g. audio fingerprints), or by psychoacoustic parameters like sharpness, pulse characteristics, roughness, pitch, varying intensity or volume. Thus, according to a further embodiment, the method comprises the step of determining a buffered recorded audio fingerprint or determining psycho-acoustic parameters. Usually, it is sufficient to store the recordings or audio fingerprints in a database, while the psycho-acoustic parameters represent additional information. An advantage of audio fingerprinting is that the recording is stored in an anonymous way.
In a single step of obtaining a signal from a user interface, such as a key press, another alternative or additional signal may also be obtained that subjectively evaluates the currently identified control noise. This subjective evaluation involves assigning the audio signal to a signal class (e.g., less interference or more interference). Here, the subjective assessment is stored in association with the corresponding part or parameter.
According to further embodiments, a timestamp may also be stored in conjunction with the portion or parameter. According to further embodiments, it is also conceivable to store the current position information by means of a GPS receiver, for example. It is also feasible to store the data to be buffered in a data-reduced manner in order not to make the database too large.
It is noted here that according to an embodiment, the memory or database is directly contained in the respective device performing the method, or may be provided externally according to another embodiment.
Further embodiments relate to a corresponding apparatus. The device comprises a microphone for continuous recording, a buffer for buffering, an interface for receiving the signal and a further memory for storing recordings belonging to a signal class (audio file, audio fingerprint or psychoacoustic parameters) related to the identified disturbing noise. According to further embodiments, the device may comprise an input interface (e.g. a key) with which the presence of subjective disturbance noise may be confirmed or noise may be generally assigned to a signal class. For example, the input device may be expanded by classifying one of several signal classes (i.e., by evaluation). According to still further embodiments, the device may further comprise a communication interface through which an external memory (external database) is connected.
According to a further embodiment, a computer program for performing one of the described methods is provided.
Drawings
Embodiments of the invention will be described in detail below with reference to the attached drawing figures, wherein:
FIG. 1a is a flow chart depicting a method of "building a database" according to aspect 1 in a basic variant;
FIG. 1b is a flow chart for depicting an expansion method according to aspect 1;
fig. 1c to 1f show variants of the apparatus of aspect 1;
fig. 2a is a flow chart of a method for depicting a corresponding basic variant of aspect 2 "identifying noise of a signal class";
FIG. 2b is a flow chart of an expanded embodiment of aspect 2;
FIG. 2c is a schematic block diagram of the apparatus of aspect 2;
FIG. 3a is a flow chart depicting a method of aspect 3 "analyzing noise of various signal classes";
fig. 3b is a schematic block diagram of the apparatus of aspect 3.
Detailed Description
Before discussing embodiments of the present aspect in more detail below, it is pointed out that elements and structures of the same effect have the same reference numerals, so that the description thereof is mutually applicable or interchangeable.
Fig. 1a shows a method 100 for building a database, comprising the steps of "receiving and recording 110 and signal reception using a microphone 11" 120. When the signal 120 has been received (see where decision 125), the record of step 110 is stored in a database, shown using step 130. This step 130 essentially represents the end of the method 100 (see end point 135).
With respect to the step of "recording 110" it is noted that there may typically be coding sub-steps at the time of recording. The encoding may also be implemented such that a so-called audio fingerprint is obtained, i.e. a derivation of the characteristic parameters for recording. The audio fingerprint is strongly compressed and thus anonymized when compared to the recording, wherein the audio fingerprint still allows the use of the audio fingerprint to identify comparable noise, i.e. noise of the same class. In general, an audio fingerprint may be described such that it is a representation of an audio signal that represents all of the essential features of the audio signal, enabling subsequent classification. Audio fingerprints are typically insufficient to allow decoding to form a true audio signal and thus privacy is protected. Similarly, or in parallel to the encoding, there may be a sub-step of deriving parameters describing the recording, such as psychoacoustic parameters.
The recording process 110 may also be described as a circular buffer, as recordings will typically be repeatedly overwritten and therefore only buffered for a predetermined period, such as for example 120 seconds, 60 seconds or 30 seconds, or typically more than 5 seconds. The circular buffer also provides the advantage of meeting privacy requirements. When the signal 120 is obtained, this time window of the ambient noise of the previous cycle is stored or eventually stored in another memory (e.g., a database) so that it is available later, using step 130. To efficiently build a database, the method 100 is repeatedly performed for several signals of one or different signal classes.
The method 100 is used to build a database in which subjective interference noise received (i.e. recorded) by the microphone 11 is identified. When the user has identified disturbing noise in the environment, the identification is done using a step performed by the user, which uses the keys 12 (or generally the user input interface 12) to exemplarily perform the "signal 120 output" step. Since the microphone 110 listens for ambient noise and buffers it in step 110, these interference noises are also recorded so that the buffered recording or a part thereof can be stored in a permanent memory for building a database (see step 130). In case the user does not recognize the disturbing noise, the method will be repeated, using the arrow from the subjective evaluation (decision element 125) to the starting point 101 to illustrate the procedure.
The advantage of this approach is that in this way a sufficiently wide database can be built, which comprises a plurality of records or parameters (such as audio fingerprints) associated with subjectively perceived disturbing noise.
It is noted here that this procedure may cause a dependency of the time point of the signal on the time window. Illustratively, the start of a time window at the time of the signal is a fixed distance, such as 30 seconds or 60 seconds, before the time of the signal, which may cause a correlation. In addition, the end of the time window may also depend on the time of the signal, such that the time of the signal coincides with the end of the time window, or there is a time distance of 5 seconds (ends before the time of the signal), for example. Typically, the correlation is chosen such that the recording time window will always precede the time of the signal, which may also be within the time window.
FIG. 1b shows an extension method 100' that also allows for building a database, but with extended information. The method 100' is generally based on the method 100 and is limited in its course by a start 101 and an end 135. Thus, the method 100' further comprises the following basic steps: the recording 110', receiving 120' is related to subjective noise assessment or generally to the assignment of the received signal to a signal of one of a plurality of signal classes (e.g. non-interfering noise, slightly interfering noise and highly interfering noise), as well as the storage of a buffer record 130 as with a database. In addition, steps 130 and 120' are connected through decision point 125.
In this embodiment, the recording step 110' is subdivided into two sub-steps, namely 110a ' and 110b '. For example, step 110a involves calculating psycho-acoustic parameters such as coarseness, clarity, volume, pitch, and pulse characteristics and/or variation intensity. Step 110b is generalized to determine the audio fingerprint describing the recording so that the audio fingerprint identification characteristic feature can be used again later.
There may be different input means to perform the step 120' of subjective noise assessment. These are "evaluation using keys or buttons on the device performing the method 100 '(reference numeral 12 a')", correlating subjective noise evaluations using questionnaires (reference numeral 12b '), or evaluating using smart devices (reference numeral 12 c'). To perform the step of subjective noise assessment 120', the three assessment variants 12a', 12b 'and 12c' may be used individually or in combination. Once the presence assessment (see decision point 125), the psycho-acoustic parameters (reference numeral 110a '), and/or the audio fingerprint (reference numeral 110b') are stored in memory, which is shown in step 130.
According to further embodiments, time and/or location information may be added in addition to the pure parameters or the fingerprint or part of the audio recording. These are also stored in step 130 and originate from a further step 132, which correspondingly comprises: a current location is determined and/or a current time is determined.
When the database has been built and has a corresponding size (see step 130), it may be evaluated by correlation or statistical evaluation as shown in step 132.
A typical application of the above-described methods 100 and 100' is, for example, a situation where the device is located in a hotel room and is monitoring the current ambient noise. When a guest of a hotel wants to be quiet and restful in the room of his hotel, but the disturbing noise prevents him from doing so, he or she can flag these disturbing noises. This may have the result that the sound of the room may not be too loud, but there may be some noise, such as air conditioning, which prevents the guest from falling asleep. Using the device, he or she may perform a subjective evaluation, i.e. classification into signal classes such as "interference", "very interference" or "high interference". The evaluation characterizes the noise situation evaluated using different parameters. Finally, the audio fingerprint, the psychoacoustic parameters, or in general the record associated with one of the signal classes is stored in a database.
Three variations of the device are discussed below with reference to fig. 1c, 1d and 1 e.
Fig. 1c shows a first device variant, namely a device 20, which is connected via an interface or radio interface to an actual signal processing unit (not shown) and is basically configured to emit a signal for identifying interfering signals or specific signal classes. Here, the device 22 in the present embodiment includes two keys 24a and 24b that can perform subjective evaluation. These keys 24a and 24b are associated with different signal classes.
The device 20 may illustratively be a smart device (e.g., tablet, smart watch, smart phone) that includes virtual keys 24a and 24b integrated into an application. The application may also illustratively contain a questionnaire by which other information of a general quality may be collected from the user (e.g., hotel guest).
When the button 24a or 24b is operated, a method of buffering ambient noise or deriving parameters and then actually storing is performed in the actual data collection apparatus. For example, the external device may be a server having microphones at respective monitoring locations.
Fig. 1d shows another variant, in which an internal microphone 26 for receiving ambient noise is integrated in the device 20' comprising the buttons 24a and 24 b. Additionally or alternatively, the external microphone 26e may be connected to the device 20' via an interface.
Fig. 1e shows another variant of the device 20 "which no longer comprises buttons as input means, but which only comprises an internal microphone 26 or an optional or alternative external microphone 26e, and can be controlled using this voice command which can be associated with ambient noise of the signal class.
With reference to the devices 20' and 20 ", it should be noted in this connection that also several external microphones can be connected. It is also conceivable here to record structure-borne sound in addition to normal airborne sound (meaning that the respective device comprises a structure-borne sound receiver).
With reference to the embodiment of fig. 1c and 1d, it is noted that the different buttons 24a and 24b may also be extended by other buttons. To distinguish the buttons, color coding may be provided: red-interference, yellow-indifference, green-very pleasant ambient noise (the latter being exemplarily applicable when the bird sings are clearly visible but perceived as ideal noise).
With reference to fig. 1c to 1d, it should be mentioned that the devices 20, 20' and 20 "may additionally be integrated as software applications and additionally integrated in devices like smartphones, tablets or smartwatches. These software applications can allow the following functions:
-extending the detection of said noise quality by means of questionnaire techniques or different subjective acquisition techniques;
using sensor systems present in other devices (microphone, GPS, tilt sensor, biofeedback functions);
wireless or, if applicable, mechanical connection to the device for data communication developed herein;
-using the software developed herein to fully control the devices developed herein.
Fig. 1f shows the components of the device 20 "'. The device 20 "' comprises a microphone 26, optional calibration means 26k for calibrating the microphone, as well as a processing unit 42 and a memory 44.
The processing means 42 comprise a pre-processing 46 for encoding an audio file or for deriving an audio fingerprint, and a unit 48 for determining psycho-acoustic parameters. Both the metadata of the pre-processing 46 and the psycho-acoustic parameters of the unit 48 are written to the memory 44. In addition, the audio signal may be more accurately stored or stored in the memory 44 by the unit 49 controlled by a button, for example.
The calibration device 26k is used to provide defined sensitivity values for all sensors. Here, measurement or recording of the switch, frequency response, or compression, for example, is performed in advance.
From the stored audio samples, metadata (audio fingerprints of psycho-acoustic parameters) and the markers obtained by means of one of the input means from fig. 1c to 1d, an actual data analysis by means of the data analyzer 50 and an association with the respective signal class can then be performed.
It is noted here that the device is typically a mobile device and may therefore be supplied with electrical power typically using a battery or accumulator. Alternatively, a conventional power supply is also possible. For storing the record the device may further comprise a storage medium, such as a portable storage medium (e.g. an SD card) or a connection to a server. The connection to the server is done through a wired or glass fibre interface or even a radio interface. At the protocol level, there are different ways to do this, which will not be discussed in detail here.
To improve the evaluability, the device may further comprise means for accurate synchronization with other devices, such as, for example, a time code or a world clock. In addition, it is also conceivable to couple the device to a location determination unit, such as a GPS receiver, or to integrate it together, in order to determine which interfering noise has been determined at which location or is perceived as interfering.
It is noted here that, corresponding to other embodiments, the method 100 or 100' may also comprise a pre-calibration (reference calibration device 26 k). This means that, corresponding to an embodiment, the method 100 or 100' discussed above comprises a step of calibration.
With respect to aspect 1, it is noted that, corresponding to the embodiments, it is also conceivable that all these devices perform a data-inductive recording of the measurement data in order to inductively data. Data summarization may also be advantageous for long-term measurements. Depending on the degree of compression or error, it can be ensured that privacy is preserved, since the monitored data can always be compressed, so that essentially only parameters, such as psychoacoustic parameters (roughness, sharpness, pitch, etc.) or audio fingerprints, are recorded. It is again noted here that the exact decision whether to use a recording or audio fingerprint or merely psycho-acoustic parameters is essentially influenced by legal framework conditions for data and consumer protection.
As already discussed above, so-called "audio fingerprints" are used, wherein there are different variants thereof which will be discussed in more detail below. Many methods are known with which features or fingerprints can be extracted from an audio signal. U.S. Pat. No. 5,918,223 discloses a method for content-based analysis, storage, retrieval and segmentation of audio information. Analysis of the audio data generates a set of values called feature vectors, which can be used to classify and rank the similarity between individual audio segments. The volume, pitch, brightness of the tone, bandwidth and so-called Mel-frequency cepstral coefficients (MFCCs) of an audio piece are used as features to characterize or classify the audio piece. The value of each block or frame is stored and then first order derived with respect to time. To describe the variation over time, a statistic, such as a mean or standard deviation, of each of these features, including its first derivative, is calculated therefrom. The set of statistics forms a feature vector. The feature vector is thus a fingerprint of the audio piece and may be stored in a database.
Similar concepts for indexing and characterizing multimedia segments are disclosed by Yao Wang et al in the expert publication "multimedia content analysis" (IEEE journal of signal processing, 11 months 2000, pages 12 to 36). To ensure a valid association of an audio signal with a particular class, a number of features and classifiers have been developed. A time range feature or a frequency range feature is suggested as a feature for classifying the content of the multimedia segment. These include volume, pitch, which is the fundamental frequency of the audio signal shape, spectral characteristics (such as energy content of the frequency band relative to the total energy content), cut-off frequencies in the spectral process, etc. In addition to the short-term feature involving per block size of the so-called audio signal samples, a long-term quantity is suggested which is related to a longer period of the audio piece. Additional exemplary features are formed by forming time differences for each feature. The features obtained in the blocks are also rarely so directly passed on for classification, since they exhibit too high a data rate. One conventional form of further processing is to compute short-term statistics. In these statistics, for example, calculation of the mean, variance, and time correlation coefficient is performed. This reduces the data rate and on the other hand leads to an improved recognition of the audio signal.
WO 02/065782 describes a method of forming fingerprints to form a multimedia signal. The method involves extracting one or several features from the audio signal. The audio signal here is divided into segments, and processing of blocks and frequency bands is performed in each segment. By way of example, band-by-band computation of energy, pitch and standard deviation of the power density spectrum is mentioned.
From DE 10134471 and DE 10109648 an apparatus and a method for classifying an audio signal are known, in which a fingerprint is obtained by a measure of the pitch of the audio signal. Fingerprints here allow for a robust, content-based classification of audio signals. The documents mentioned here disclose several possibilities for generating measures of pitch on an audio signal. In this case, converting the segments of the audio signal into a spectral range is the basis for calculating the pitch. The pitch for one band or all bands may then be computed in parallel. However, a disadvantage of such a system is that as the distortion of the audio signal increases, the fingerprint no longer has sufficient expressiveness and it is no longer possible to identify the audio signal with satisfactory reliability. However, distortion occurs in many cases, particularly when audio signals are transmitted using a system with low transmission quality. This is particularly true in the case of mobile systems or strong data compression today. Such systems, like mobile phones, are mainly implemented for the bi-directional transmission of speech signals and often only music signals are transmitted with a very low quality. There are also additional factors that may have a negative impact on the quality of the transmitted signal, such as low quality microphones, channel interference and transcoding effects. For an apparatus for recognizing and classifying signals, the deterioration of signal quality may cause a serious deterioration of recognition performance. Inspection has revealed that, particularly when using a device or method according to DE 10134471 and DE 10109648, variations in the system while maintaining the recognition criterion of pitch (spectral flatness measure) do not result in a further significant improvement in recognition performance.
When it is assumed that a sufficient database comprising noise, such as interference noise of different signal classes, has been established, it is thus possible to search for a specific interference noise in any environment and then to see whether such interference noise has been identified. This method is illustrated in fig. 2 a.
Fig. 2a shows a method 200 comprising a step 210 of matching ambient noise received via the microphone 11 (step of reference reception 205) to a record from the database 15. Once a match has been found, as shown at decision 215, a signal is output, such as for logging or for excluding further action. The method is repeated as long as no match has been found, which is illustrated using an arrow to the starting point 201.
Corresponding to an embodiment, the respective audio fingerprint, rather than the recording, of the current ambient noise may be compared to the audio fingerprints previously stored in the database 15. The method here comprises determining an audio fingerprint of the current ambient noise and comparing it with audio fingerprints stored in the database 15.
Even when it is assumed in the method 200 that the ambient noise or audio fingerprint is identified/associated in advance by matching it with the ambient noise/audio fingerprint stored in the database 15, in general, the ambient noise may be monitored with respect to a rule. In the case of comparing ambient noise/audio fingerprints, the rule would mean "partial match".
Another such rule may be, for example, simply a volume value to be exceeded or a threshold value related to a psychoacoustic parameter to be exceeded. According to an embodiment, psychoacoustic parameters of the current ambient noise are derived, which are compared by predefined rules with predefined respective thresholds in order to identify the occurrence of such events.
According to an extended embodiment, the method may not only purely identify such disturbing noise, but may also correlate (i.e. classify) the noise as e.g. speech, motor noise, music, church chimes or shots.
One potential application scenario for performing this method, illustratively on a smart phone or a device specifically designed for this purpose, is where the device is located in a hotel room and is monitoring ambient noise. Here, data from the database 15 is used to evaluate the ambient noise and to store how many noise events that may be perceived as interference have occurred over time and which of them may be perceived as interference. This may be, for example, calculating the disturbing air conditioning noise over the course of a day. As an alternative to logging, an audio recording of this noise or a pre-buffered storage of the ambient noise may be performed (see above). The basic idea is that the hotel operator can use this method to predict and evaluate noise perception.
Even when the above embodiments have assumed that the noise of different signal classes is identified and associated separately from each other, it is mentioned here that according to further embodiments it is for example also possible to identify and associate several noises of different signal classes that may overlap.
Fig. 2b shows an expanded method 200' including additional steps between the step or point of decision 215 and the end 216.
These are counted events or used element variables 221, via step 220, thus obtaining the number of events 222. Optionally, audio recording may begin with an event that has been identified, as shown using step 230.
Fig. 2c shows another embodiment of the device 40. It comprises a processor 41 as a central unit which performs the actual steps of analysis/matching. First, it uses the internal microphone 26, wherein access to the external microphones 26e1 and 26e2 is also conceivable. For example, data for matching is stored in the internal memory 44.
Optionally, the processor is configured to determine and match audio fingerprints and/or psychoacoustic parameters in order to obtain corresponding rule matches.
To allow this functionality, another peripheral unit is optionally provided, such as an internal clock 55, a battery 56b or a power supply 56, which may also be typically implemented using a cable 56 k. Optionally, the processor also has access to a further sensor element 57, a control unit 58 (such as a record activation button) or a timer 59. Here, according to a further embodiment, the processor 41 may be further configured to perform an objective noise assessment in order to establish the association in connection with a subjective assessment (identifying subjective tone events).
According to an embodiment, the CPU may classify/categorize each identified noise of the signal classes in different evaluation matrices depending on the respective noise class according to the previously obtained subjective evaluation of the pleasure.
According to further embodiments, an external data storage 60 (such as an external hard disk or server) may also be provided for storing or loading the database. The connection may be a wired connection or a wireless connection. In wireless communication, corresponding to further embodiments, a communication interface 62 will be provided, such as a wireless interface 62w or a wired interface 62k enabling external access.
According to another aspect, a system is provided that consists essentially of two of the previously described devices 40 in combination with each other such that the two devices mutually activate upon receiving corresponding noise (i.e., signal class) in one of the devices. The system is used to analyze or evaluate the noise of the individual noise classes in more detail. Where the method discussed below in figure 3 is performed.
Fig. 3a shows a method 300 comprising steps of noise analysis corresponding to the method 200 or 200' performed at a first location and a second location. This means that step 210 exists twice (refer to 210a and 210 b).
The determined recordings or parameters, such as audio fingerprints at the two locations (generated by steps 210a and 210b), are then compared in a further step 220.
According to an embodiment, the two steps 210 at two adjacent locations may be interdependent as illustrated using the optional step "audio recording on adjacent devices 211". Alternatively, another action may be performed at an adjacent device. The reason for this is that, for example, when a first device performing method 210a identifies noise and activates a second device performing method 210b, the same noise may be identified at different locations. It is finally mentioned here that, starting from the place of decision 215, there is another arrow pointing to the starting point 301, which basically implies the following fact: the method of noise analysis 210a will be performed until a corresponding match has been found.
Since the locations are typically spatially adjacent, it is possible to estimate the propagation of noise, velocity, or a larger noise source in this manner.
Illustratively, when comparing its own analysis to the analysis on different devices at the same time, when the same event has been identified at several devices, it can be determined whether this is a global event such as lightning (refer to reference numeral 232 after the domain of decision 321) or a local event (refer to reference numeral 324 after the domain of decision 321). For global event 323, typically the difference in level between "near" and "far" devices is negligible (level 1/r, r variation relative to r is small). For local events 324, the levels are very different (level 1/r, r varies greatly with respect to r). For example, the local event may be a call for help, an explosion, an open concert. For local events, other analyses may then be performed, i.e., analyses with respect to other parameters 325. From the shift in time or the shift in frequency, the amount of local events, propagation, or timelines may be determined. Determining that the global event 323 or the local event 324 (e.g., its analysis 325) is substantially the end 329 of the method.
For example, one possible application scenario is to distribute several devices in a city center. All devices are connected to each other by a data connection, such as a wired, wireless, ethernet or LAN connection. A connection using a server is also possible. All devices analyze the noise situation (psycho-acoustic parameters, audio fingerprints). One of the devices identifies characteristic events, such as signal classes previously classified in a database. The audio recording is triggered on the spot. At the same time, the device triggers a behavior, such as an action on one of the neighboring nodes. By comparing the two nodes, a global event and a local event can be distinguished, as already discussed above.
For embodiments where two different classes of noise are present at the same time, it is possible, according to another embodiment, to monitor the noise even when the noise does not behave identically (move) or is associated with local events on the one hand and global events on the other hand. For example, a motor vehicle moving from location a to C via B may be monitored for noise, whereas a siren (e.g., siren) sounds at location B over time. Another example is bird calls and traffic noise and fire truck-bird calls can be loud but pleasant, traffic noise quiet enough so as not to cause interference, but rather to monitor the fire truck over several devices and increase "attention" by signal category.
Method 300 is substantially performed by a system comprising two of devices 40 (fig. 2 c).
However, it is also possible that there is little change, as shown in fig. 3b, since an additional interface for connecting the two devices is provided.
Fig. 3b shows a device 70 comprising a microphone 26 and an optional calibration unit 26k on the input side. For example, the audio stream received by the microphone is preprocessed by the preprocessing 46 in order to derive an audio fingerprint (reference numeral 46a) or psycho-acoustic parameters (reference numeral 48). Meanwhile, an event or category may be identified (refer to reference numeral 50). By identifying an event/category, an automatic audio recording may be triggered on the one hand (reference numeral 50a1), or a control instruction may be issued, such as for activating another node (reference numeral 50a2 or another device 70'). The means for outputting control instructions 50a2 may illustratively activate a memory that subsequently receives and records data from the means for generating audio fingerprint 46a or the means for deriving psycho-acoustic parameters 48. The audio signal may also be stored in the memory 44, wherein the recording may also be enabled or prevented here by means of the button 49 a. In this embodiment, the CPU 41 may also be connected to a timer 59.
A device 70', which performs substantially the same function, is disposed at another adjacent position, except for the device 70. The device 70 'further comprises a memory 44 which has stored the audio result for the time segment when the device 70' has been activated by means of the activation means 50a2 or on the basis of noise which is identified and belongs to a class. The data analyzer 72 analyzes the recordings or audio fingerprints or psychoacoustic parameters from the memory 44 of the devices 70 and 70' in a next step (e.g. with respect to expansion). Here, however, it is advantageous for the data analyzer 72 to be connected to both memories of the other device, wherein it is mentioned here that the data analyzer 72 may be arranged in one of the devices 70 and 70' or externally with respect to both of them.
Corresponding to another embodiment, a button (e.g., button 24a ') may be integrated into device 70 such that device 70 also performs the functions of device 20, 20', or 20' ".
Optional element 50a' allows automatic triggering of recording after recognition of a classification. Alternatively, it is also conceivable here to start recording automatically when no noise is found in any signal class that has been obtained.
In other words, method 303 may describe that the functionality of method 200 is substantially covered, i.e. that noise such as speech, motor noise, music, kitchen tiles, shot sounds, etc. is recognized and classified, and that the functionality has been extended by analysis according to multiple microphones at different locations.
It is also noted herein that certain categories of automatic recording, such as explosions and shots, may also suggest terrorism, for example. Here, it may be useful for all neighboring nodes 70/70' to switch directly to recording.
In addition, it is also possible to automatically (e.g., temporally limited) record when a particular noise threshold is exceeded for a period of time. The recordings can also be extended to neighboring nodes in order to thus perform an accurate localization of the signal source by these longer recordings when combining multiple nodes (leading to investigation of interference sources, separating noise sources).
The potential application areas for the three cases are as follows:
-tourism, hotel, health sector, bike path, hiking path;
work protection (office work, machine factory, cabin workplace);
city planning (soundscape, noise mapping);
public safety (monitoring production facilities).
Combinations of methods 100/100', 200/200', and 300 or the functions of apparatuses 20/20'/20 "/20 '", 40, and 70/70' are also contemplated. Such examples are a combination of devices and methods for subjective assessment and recording and machine assessment of the device.
It is noted here that elements already discussed in connection with the different aspects may of course also be applied to the second aspect. Illustratively, the teachings relating to audio fingerprints or psychoacoustic parameters apply to all three aspects, wherein the teachings are only discussed in more detail in connection with the first aspect.
Although some aspects have been described in the context of a device, it is clear that these aspects also represent a description of the corresponding method, so that a block or element of a device also corresponds to a respective method step or feature of a method step. Similarly, aspects described in the context of a method step also represent a description of a corresponding block or item or feature of a corresponding apparatus. For example, some or all of the method steps may be performed by (or using) a hardware device, such as a microprocessor, programmable computer, or electronic circuitry. In some embodiments, some or several of the most important method steps may be performed by such an apparatus.
The inventive encoded signals, such as audio signals or video signals or transport stream signals, may be stored on a digital storage medium or may be transmitted over a transmission medium, such as a wireless transmission medium or a wired transmission medium, such as the internet.
For example, the inventive encoded audio signal may be stored on a digital storage medium or may be transmitted over a transmission medium such as a wireless transmission medium or a wired transmission medium (e.g., the internet).
Embodiments of the invention may be implemented in hardware or software, depending on certain implementation requirements. This embodiment may be performed using a digital storage medium, such as a floppy disk, a DVD, a blu-ray disk, a CD, a ROM, a PROM, an EPROM, an EEPROM or FLASH memory, a hard disk drive or another magnetic or optical memory having electronically readable control signals stored thereon, which cooperate or are capable of cooperating with a programmable computer system such that the respective method is performed. Accordingly, the digital storage medium may be computer-readable.
Some embodiments according to the invention comprise a data carrier comprising electronically readable control signals capable of cooperating with a programmable computer system such that one of the methods described herein is performed.
Generally, embodiments of the invention may be implemented as a computer program product having a program code operable to perform one of the methods when the computer program product runs on a computer.
The program code can be stored, for example, on a machine-readable carrier.
Other embodiments include a computer program for performing one of the methods described herein, wherein the computer program is stored on a machine-readable carrier.
In other words, an embodiment of the inventive methods is thus a computer program comprising program code for performing one of the methods described herein, when the computer program runs on a computer.
Thus, another embodiment of the inventive method is a data carrier (or a digital storage medium or a computer readable medium) comprising a computer program recorded thereon for performing one of the methods described herein.
Thus, another embodiment of the inventive method is a data stream or a signal sequence representing a computer program for performing one of the methods described herein. For example, a data stream or signal sequence may be configured to be transmitted over a data communication connection (e.g., over the internet).
Another embodiment includes a processing apparatus, such as a computer or programmable logic device, configured or adapted to perform one of the methods described herein.
Another embodiment comprises a computer having installed thereon a computer program for performing one of the methods described herein.
Another embodiment according to the present invention comprises a device or system configured to transmit a computer program for performing one of the methods described herein to a receiver. The transfer may be performed electronically or optically. For example, the receiver may be a computer, mobile device, storage device, or the like. For example, the apparatus or system may comprise a file server for transmitting the computer program to the receiver.
In some embodiments, a programmable logic device (e.g., a field programmable gate array, FPGA) may be used to perform some or all of the functions of the methods described herein. In some embodiments, a field programmable gate array may cooperate with a microprocessor to perform one of the methods described herein. Generally, in some embodiments, the method is performed by any hardware device. This may be universally applicable hardware, such as a Computer Processor (CPU), or hardware dedicated to the method, such as an ASIC.
The above-described embodiments are merely illustrative of the principles of the present invention. It is to be understood that modifications and variations of the configurations and details described herein will be apparent to others skilled in the art. Therefore, the invention is intended to be limited only by the scope of the appended patent claims, and not by the specific details given herein by way of illustration and description of the embodiments.

Claims (25)

1. A method (300) for analyzing noise of at least one signal class of a plurality of signal classes, comprising the steps of:
receiving (205) ambient noise at a first location;
establishing (210) whether the ambient noise of the first location or a set of parameters derived from the ambient noise of the first location satisfies a predefined rule describing a signal class of the plurality of signal classes;
receiving ambient noise at a second location;
establishing whether the ambient noise of the second location or a set of parameters derived from the ambient noise of the second location satisfies a predefined rule describing a signal class of the plurality of signal classes; and
determining (320) a relationship between the ambient noise at the first location and the ambient noise at the second location, or a relationship between a set of parameters derived from the ambient noise at the first location and a set of parameters derived from the ambient noise at the second location;
wherein determining (320) the relationship comprises: determining a difference in a level of ambient noise at the first location and the second location, wherein the ambient noise at the first location and the second location is produced by the same interference noise; and/or
Wherein determining (320) the relationship comprises: determining a time offset and/or a run-time offset of ambient noise at the first location and the second location; and/or
Wherein determining (320) the relationship comprises: determining a frequency difference and/or determining a hall effect in ambient noise at the first and second locations; and/or
Wherein the determining (320) comprises: an analysis (325) is performed with respect to a distance between the first location and the second location, a movement of a noise source is analyzed according to a relationship and/or a size of the noise source is analyzed.
2. The method of claim 1, wherein the establishing (210) is performed by comparing the ambient noise to a previously buffered recording, and the predefined rule defines that the ambient noise matches at least part of the previously buffered recording.
3. The method of claim 1, wherein the establishing (210) is performed by comparing a derived parameter set with a previously derived parameter set, and the predefined rule defines at least a partial match of the derived parameter set with the previously derived parameter set.
4. The method of claim 2, wherein the derived parameter set and the previously derived parameter set comprise audio fingerprints.
5. Method according to claim 2, wherein the previously buffered recordings and/or the previously derived parameter sets are received by an external database (15).
6. The method according to claim 1, wherein the derived set of parameters comprises psychoacoustic parameters, and wherein the establishing (210) is performed by evaluating the psychoacoustic parameters of the ambient noise, and the predefined rule comprises a threshold value of the psychoacoustic parameters.
7. The method of claim 6, wherein the psychoacoustic parameters include volume, clarity, pitch, roughness, pulse characteristics, and/or variation intensity.
8. The method of claim 1, wherein the information about the frequency of the noise of the signal class comprises the signal class when stored (220); or
Wherein the information comprises a time indication when the noise of the signal class is identified and/or a location indication where the noise of the signal class is identified.
9. The method according to claim 1, wherein the previously obtained record or parameter is read from an external database (15).
10. The method according to claim 1, wherein the steps of receiving (205), establishing (210) are repeated for adjacent locations.
11. The method of claim 10, further comprising: a relationship between ambient noise at a first location and ambient noise at a second location is determined (320), or a relationship between a set of parameters derived from ambient noise at the first location and a set of parameters derived from ambient noise at the second location is determined.
12. The method of claim 1, wherein the predefined rule describes ambient noise or a set of parameters for a control signal such that an activation signal is issued when the predefined rule has been met.
13. The method (300) according to any one of claims 1-12, wherein the received ambient noise at the first location and/or the second location is recorded for a migration time window, or wherein the set of parameters is derived from the ambient noise at the first location and/or the second location for the migration time window.
14. The method (300) according to any one of claims 1-12, wherein the establishing (210) is performed by comparing ambient noise to a previously buffered recording, and the predefined rule defines that ambient noise matches at least part of the previously buffered recording.
15. The method (300) according to any one of claims 1-12, wherein establishing (210) is performed by comparing a derived parameter set with a previously derived parameter set, and the predefined rule defines at least a partial match of the derived parameter set and the previously derived parameter set.
16. The method (300) of claim 15, wherein the derived set of parameters and the previously derived set of parameters comprise an audio fingerprint.
17. The method (300) according to any of claims 1-12, wherein the previously buffered records and/or the previously derived parameter sets are received from an external database (15).
18. A digital storage medium on which a computer program comprising program code is stored for performing the method according to any of claims 1-17 when the program is run on a computer.
19. An apparatus (40) for correlating noise of at least one signal class of a plurality of signal classes, comprising:
a microphone (11) for receiving ambient noise;
a processor (50) for establishing whether the ambient noise or a set of parameters derived from the ambient noise satisfies a predefined rule describing a signal class of the plurality of signal classes;
an interface for logging that the predefined rule has been met, or recording the environmental noise received for a migration time window, or deriving a set of parameters from the environmental noise of the migration time window and storing the set of parameters of the migration time window, or issuing an activation signal for another device to identify noise;
wherein the processor (50) is configured to determine (320) a relationship between the ambient noise at the first location and the ambient noise at the second location, or a relationship between a set of parameters derived from the ambient noise at the first location and a set of parameters derived from the ambient noise at the second location;
wherein the processor (50) is configured to determine (320) a relationship comprising: determining a difference in a level of ambient noise at the first location and the second location, wherein the ambient noise at the first location and the second location is produced by the same interference noise; and/or
Wherein the processor (50) is configured to determine (320) a relationship comprising: determining a time offset and/or a run-time offset of ambient noise at the first location and the second location; and/or
Wherein the processor (50) is configured to determine (320) a relationship comprising: determining a frequency difference and/or determining a hall effect in ambient noise at the first and second locations; and/or
Wherein the processor (50) is configured to analyze (325) a movement of a noise source and/or a size of a noise source according to a relation with respect to a distance between the first location and the second location.
20. Device (40) according to claim 19, wherein the device (40) comprises a communication interface (62) by means of which protocols can be output and/or by means of which the predefined rules can be read in and/or by means of which communication with another device is possible.
21. The device (40) of claim 20, wherein the device (40) is configured to be networked with another device (40') at a neighboring location in order to determine a relationship between the ambient noise of a first location and the noisy ambient sound of a second location, or a relationship between a set of parameters derived from the ambient noise of the first location and a set of parameters derived from the ambient noise of the second location.
22. A system (77) for analyzing noise of at least one signal class of a plurality of signal classes, comprising:
a first unit (70) comprising a first microphone (11) for receiving ambient noise at a first location;
a second unit (70') comprising a second microphone (11) for receiving ambient noise at a second location; and
a processor (72) for establishing whether the ambient noise of the first location or a set of parameters derived from the ambient noise of the first location fulfils a predefined rule describing a signal class of the plurality of signal classes, and for establishing whether the ambient noise of the second location or a set of parameters derived from the ambient noise of the second location fulfils a predefined rule describing a signal class of the plurality of signal classes;
wherein the processor (72) is configured to determine a relationship between the ambient noise at the first location and the ambient noise at the second location, or a relationship between a set of parameters derived from the ambient noise at the first location and a set of parameters derived from the ambient noise at the second location;
wherein determining (320) the relationship comprises: determining a difference in a level of ambient noise at the first location and the second location, wherein the ambient noise at the first location and the second location is produced by the same interference noise; and/or
Wherein determining (320) the relationship comprises: determining a time offset and/or a run-time offset of ambient noise at the first location and the second location; and/or
Wherein determining (320) the relationship comprises: determining a frequency difference and/or determining a hall effect in ambient noise at the first and second locations; and/or
Wherein the determining (320) comprises: an analysis (325) is performed with respect to a distance between the first location and the second location, a movement of a noise source is analyzed according to a relationship and/or a size of the noise source is analyzed.
23. The system (77) according to claim 22, wherein the first and second units (70, 70') are connected via a communication interface and/or a radio interface.
24. A method (200, 200') for correlating noise of at least one signal class of a plurality of signal classes, comprising the steps of:
receiving (205) ambient noise;
establishing (210) whether the ambient noise or a set of parameters derived from the ambient noise satisfies a predefined rule describing a signal class of the plurality of signal classes;
logging (220) that the predefined rule has been met, or recording the received ambient noise for a migration time window, or deriving the set of parameters from the ambient noise of the migration time window and storing the set of parameters of the migration time window, or issuing an activation signal for another device to identify noise;
wherein the steps of receiving (205), establishing (210) are repeated for adjacent locations;
wherein the method further comprises: determining a relationship (320) between ambient noise at a first location and ambient noise at a second location, or a relationship between a set of parameters derived from ambient noise at the first location and a set of parameters derived from ambient noise at the second location;
wherein determining (320) the relationship comprises: determining a difference in a level of ambient noise at the first location and the second location, wherein the ambient noise at the first location and the second location is produced by the same interference noise; and/or
Wherein determining (320) the relationship comprises: determining a time offset and/or a run-time offset of ambient noise at the first location and the second location; and/or
Wherein determining (320) the relationship comprises: determining a frequency difference and/or determining a hall effect in ambient noise at the first and second locations; and/or
Wherein the determining (320) comprises: an analysis (325) is performed with respect to a distance between the first location and the second location, a movement of a noise source is analyzed according to a relationship and/or a size of the noise source is analyzed.
25. A digital storage medium on which a computer program comprising program code is stored for performing the method (200, 200') according to claim 24 when the program is run on a computer.
CN201680048720.2A 2015-06-30 2016-06-30 Method and apparatus for correlating noise and for analysis Active CN108028048B (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
EP15174634.4 2015-06-30
EP15174634 2015-06-30
PCT/EP2016/065397 WO2017001611A1 (en) 2015-06-30 2016-06-30 Method and device for the allocation of sounds and for analysis

Publications (2)

Publication Number Publication Date
CN108028048A CN108028048A (en) 2018-05-11
CN108028048B true CN108028048B (en) 2022-06-21

Family

ID=56368939

Family Applications (2)

Application Number Title Priority Date Filing Date
CN201680048686.9A Active CN108028047B (en) 2015-06-30 2016-06-30 Method and apparatus for generating database
CN201680048720.2A Active CN108028048B (en) 2015-06-30 2016-06-30 Method and apparatus for correlating noise and for analysis

Family Applications Before (1)

Application Number Title Priority Date Filing Date
CN201680048686.9A Active CN108028047B (en) 2015-06-30 2016-06-30 Method and apparatus for generating database

Country Status (7)

Country Link
US (2) US11003709B2 (en)
EP (2) EP3317879B1 (en)
JP (2) JP6602406B2 (en)
KR (2) KR102087832B1 (en)
CN (2) CN108028047B (en)
CA (2) CA2990891A1 (en)
WO (2) WO2017001611A1 (en)

Families Citing this family (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20190000913A (en) 2012-07-20 2019-01-03 페르노-와싱턴, 인코포레이티드. Automated systems for powered cots
KR102087832B1 (en) 2015-06-30 2020-04-21 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에. 베. Method and device for generating a database
US10402696B2 (en) * 2016-01-04 2019-09-03 Texas Instruments Incorporated Scene obstruction detection using high pass filters
US10095470B2 (en) 2016-02-22 2018-10-09 Sonos, Inc. Audio response playback
US9820039B2 (en) 2016-02-22 2017-11-14 Sonos, Inc. Default playback devices
US10264030B2 (en) 2016-02-22 2019-04-16 Sonos, Inc. Networked microphone device control
US9947316B2 (en) 2016-02-22 2018-04-17 Sonos, Inc. Voice control of a media playback system
US9965247B2 (en) 2016-02-22 2018-05-08 Sonos, Inc. Voice controlled media playback system based on user profile
US10115400B2 (en) 2016-08-05 2018-10-30 Sonos, Inc. Multiple voice services
US9942678B1 (en) 2016-09-27 2018-04-10 Sonos, Inc. Audio playback settings for voice interaction
US10181323B2 (en) 2016-10-19 2019-01-15 Sonos, Inc. Arbitration-based voice recognition
US10475449B2 (en) 2017-08-07 2019-11-12 Sonos, Inc. Wake-word detection suppression
US10048930B1 (en) 2017-09-08 2018-08-14 Sonos, Inc. Dynamic computation of system response volume
US10446165B2 (en) 2017-09-27 2019-10-15 Sonos, Inc. Robust short-time fourier transform acoustic echo cancellation during audio playback
US10621981B2 (en) 2017-09-28 2020-04-14 Sonos, Inc. Tone interference cancellation
US10482868B2 (en) 2017-09-28 2019-11-19 Sonos, Inc. Multi-channel acoustic echo cancellation
US10466962B2 (en) 2017-09-29 2019-11-05 Sonos, Inc. Media playback system with voice assistance
CN107731220B (en) * 2017-10-18 2019-01-22 北京达佳互联信息技术有限公司 Audio identification methods, device and server
US11343614B2 (en) 2018-01-31 2022-05-24 Sonos, Inc. Device designation of playback and network microphone device arrangements
US11175880B2 (en) 2018-05-10 2021-11-16 Sonos, Inc. Systems and methods for voice-assisted media content selection
US10959029B2 (en) 2018-05-25 2021-03-23 Sonos, Inc. Determining and adapting to changes in microphone performance of playback devices
US10587430B1 (en) 2018-09-14 2020-03-10 Sonos, Inc. Networked devices, systems, and methods for associating playback devices based on sound codes
US11521598B2 (en) * 2018-09-18 2022-12-06 Apple Inc. Systems and methods for classifying sounds
US11024331B2 (en) 2018-09-21 2021-06-01 Sonos, Inc. Voice detection optimization using sound metadata
US11100923B2 (en) 2018-09-28 2021-08-24 Sonos, Inc. Systems and methods for selective wake word detection using neural network models
US11899519B2 (en) 2018-10-23 2024-02-13 Sonos, Inc. Multiple stage network microphone device with reduced power consumption and processing load
EP3654249A1 (en) 2018-11-15 2020-05-20 Snips Dilated convolutions and gating for efficient keyword spotting
US11183183B2 (en) 2018-12-07 2021-11-23 Sonos, Inc. Systems and methods of operating media playback systems having multiple voice assistant services
US11132989B2 (en) 2018-12-13 2021-09-28 Sonos, Inc. Networked microphone devices, systems, and methods of localized arbitration
US10602268B1 (en) * 2018-12-20 2020-03-24 Sonos, Inc. Optimization of network microphone devices using noise classification
US10867604B2 (en) 2019-02-08 2020-12-15 Sonos, Inc. Devices, systems, and methods for distributed voice processing
DE102019001966B4 (en) * 2019-03-21 2023-05-25 Dräger Safety AG & Co. KGaA Apparatus, system and method for audio signal processing
US11120794B2 (en) 2019-05-03 2021-09-14 Sonos, Inc. Voice assistant persistence across multiple network microphone devices
US11200894B2 (en) 2019-06-12 2021-12-14 Sonos, Inc. Network microphone device with command keyword eventing
TWI716029B (en) * 2019-07-09 2021-01-11 佑華微電子股份有限公司 Method for detecting random sound segment
US10871943B1 (en) 2019-07-31 2020-12-22 Sonos, Inc. Noise classification for event detection
US11189286B2 (en) 2019-10-22 2021-11-30 Sonos, Inc. VAS toggle based on device orientation
US11200900B2 (en) 2019-12-20 2021-12-14 Sonos, Inc. Offline voice control
US11562740B2 (en) 2020-01-07 2023-01-24 Sonos, Inc. Voice verification for media playback
US11308962B2 (en) 2020-05-20 2022-04-19 Sonos, Inc. Input detection windowing
US11482224B2 (en) 2020-05-20 2022-10-25 Sonos, Inc. Command keywords with input detection windowing
EP4309172A1 (en) * 2021-03-18 2024-01-24 Telefonaktiebolaget LM Ericsson (publ) Predict and trigger a future response to a predicted background noise based on a sequence of sounds
US11533577B2 (en) 2021-05-20 2022-12-20 Apple Inc. Method and system for detecting sound event liveness using a microphone array
CN113643716B (en) * 2021-07-07 2023-09-26 珠海格力电器股份有限公司 Motor noise control method and device, motor and electrical equipment
CN117436709B (en) * 2023-12-20 2024-03-19 四川宽窄智慧物流有限责任公司 Cross-region order data overall warning method

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003009273A1 (en) * 2001-07-16 2003-01-30 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e. V. Method and device for characterising a signal and for producing an indexed signal
CN101129086A (en) * 2005-02-23 2008-02-20 弗劳恩霍夫应用研究促进协会 Apparatus and method for controlling a wave field synthesis rendering device
CN101203917A (en) * 2005-06-22 2008-06-18 弗劳恩霍夫应用研究促进协会 Device and method for determining a point in a film comprising film data applied in chronological order
CN101236250A (en) * 2007-01-30 2008-08-06 富士通株式会社 Sound determination method and sound determination apparatus
KR20090122142A (en) * 2008-05-23 2009-11-26 엘지전자 주식회사 A method and apparatus for processing an audio signal
CN101978424A (en) * 2008-03-20 2011-02-16 弗劳恩霍夫应用研究促进协会 Device and method for acoustic indication
CN102918591A (en) * 2010-04-14 2013-02-06 谷歌公司 Geotagged environmental audio for enhanced speech recognition accuracy
CN103795850A (en) * 2012-10-30 2014-05-14 三星电子株式会社 Electronic device and method for recognizing voice
CN104067339A (en) * 2012-02-10 2014-09-24 三菱电机株式会社 Noise suppression device
US10373611B2 (en) * 2014-01-03 2019-08-06 Gracenote, Inc. Modification of electronic system operation based on acoustic ambience classification

Family Cites Families (56)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5367332A (en) * 1992-10-09 1994-11-22 Apple Computer, Inc. Digital camera with time bracketing feature
US5918223A (en) 1996-07-22 1999-06-29 Muscle Fish Method and article of manufacture for content-based analysis, storage, retrieval, and segmentation of audio information
JP3645053B2 (en) * 1996-12-13 2005-05-11 大昌エンジニアリング株式会社 Real sound monitoring system
JPH10282985A (en) * 1997-04-07 1998-10-23 Omron Corp Hearing support device
FR2762467B1 (en) 1997-04-16 1999-07-02 France Telecom MULTI-CHANNEL ACOUSTIC ECHO CANCELING METHOD AND MULTI-CHANNEL ACOUSTIC ECHO CANCELER
US6381569B1 (en) * 1998-02-04 2002-04-30 Qualcomm Incorporated Noise-compensated speech recognition templates
US6260010B1 (en) * 1998-08-24 2001-07-10 Conexant Systems, Inc. Speech encoder using gain normalization that combines open and closed loop gains
JP3283024B2 (en) * 1999-12-07 2002-05-20 京都府 Noise measurement management system and recording medium for the system
US6990453B2 (en) 2000-07-31 2006-01-24 Landmark Digital Services Llc System and methods for recognizing sound and music signals in high noise and distortion
US7055107B1 (en) 2000-09-22 2006-05-30 Wireless Valley Communications, Inc. Method and system for automated selection of optimal communication network equipment model, position, and configuration
ATE405101T1 (en) 2001-02-12 2008-08-15 Gracenote Inc METHOD FOR GENERATING AN IDENTIFICATION HASH FROM THE CONTENTS OF A MULTIMEDIA FILE
DE10109648C2 (en) 2001-02-28 2003-01-30 Fraunhofer Ges Forschung Method and device for characterizing a signal and method and device for generating an indexed signal
US6996531B2 (en) 2001-03-30 2006-02-07 Comverse Ltd. Automated database assistance using a telephone for a speech based or text based multimedia communication mode
US6683938B1 (en) * 2001-08-30 2004-01-27 At&T Corp. Method and system for transmitting background audio during a telephone call
US6959276B2 (en) * 2001-09-27 2005-10-25 Microsoft Corporation Including the category of environmental noise when processing speech signals
US7158931B2 (en) * 2002-01-28 2007-01-02 Phonak Ag Method for identifying a momentary acoustic scene, use of the method and hearing device
JP3995040B2 (en) * 2002-04-19 2007-10-24 リオン株式会社 Sound recording method and apparatus using the same in noise vibration measurement
FR2842014B1 (en) * 2002-07-08 2006-05-05 Lyon Ecole Centrale METHOD AND APPARATUS FOR AFFECTING A SOUND CLASS TO A SOUND SIGNAL
JP4352790B2 (en) * 2002-10-31 2009-10-28 セイコーエプソン株式会社 Acoustic model creation method, speech recognition device, and vehicle having speech recognition device
JP2005234074A (en) * 2004-02-18 2005-09-02 Sony Corp Apparatus and method for information processing, recording medium, and program
DE102004036154B3 (en) * 2004-07-26 2005-12-22 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for robust classification of audio signals and method for setting up and operating an audio signal database and computer program
CN1727911A (en) * 2004-07-26 2006-02-01 松下电器产业株式会社 Acoustic control positioning system and method thereof
JP2006189367A (en) * 2005-01-07 2006-07-20 Chugoku Electric Power Co Inc:The Noise measuring instrument
US7464029B2 (en) 2005-07-22 2008-12-09 Qualcomm Incorporated Robust separation of speech signals in a noisy environment
US20070183604A1 (en) * 2006-02-09 2007-08-09 St-Infonox Response to anomalous acoustic environments
US20100257974A1 (en) 2006-06-21 2010-10-14 Carrens Vernon M Mechanical machine designed to utilize unbalanced torque to enhance angular momentum
US20080024469A1 (en) * 2006-07-31 2008-01-31 Niranjan Damera-Venkata Generating sub-frames for projection based on map values generated from at least one training image
JP5105912B2 (en) * 2007-03-13 2012-12-26 アルパイン株式会社 Speech intelligibility improving apparatus and noise level estimation method thereof
US8762143B2 (en) * 2007-05-29 2014-06-24 At&T Intellectual Property Ii, L.P. Method and apparatus for identifying acoustic background environments based on time and speed to enhance automatic speech recognition
WO2009046359A2 (en) * 2007-10-03 2009-04-09 University Of Southern California Detection and classification of running vehicles based on acoustic signatures
RU2536679C2 (en) 2008-07-11 2014-12-27 Фраунхофер-Гезелльшафт цур Фёрдерунг дер ангевандтен Time-deformation activation signal transmitter, audio signal encoder, method of converting time-deformation activation signal, audio signal encoding method and computer programmes
FR2943875A1 (en) * 2009-03-31 2010-10-01 France Telecom METHOD AND DEVICE FOR CLASSIFYING BACKGROUND NOISE CONTAINED IN AN AUDIO SIGNAL.
EP2328363B1 (en) * 2009-09-11 2016-05-18 Starkey Laboratories, Inc. Sound classification system for hearing aids
GB0919672D0 (en) * 2009-11-10 2009-12-23 Skype Ltd Noise suppression
EP2362620A1 (en) * 2010-02-23 2011-08-31 Vodafone Holding GmbH Method of editing a noise-database and computer device
JP5496077B2 (en) * 2010-02-25 2014-05-21 三菱重工業株式会社 Sound ray analyzer
JP5467015B2 (en) * 2010-08-24 2014-04-09 公益財団法人鉄道総合技術研究所 Evaluation method of noise in railway vehicles
US8812014B2 (en) * 2010-08-30 2014-08-19 Qualcomm Incorporated Audio-based environment awareness
CN103354937B (en) * 2011-02-10 2015-07-29 杜比实验室特许公司 Comprise the aftertreatment of the medium filtering of noise suppression gain
US9142220B2 (en) * 2011-03-25 2015-09-22 The Intellisis Corporation Systems and methods for reconstructing an audio signal from transformed audio information
US8731474B2 (en) * 2011-05-25 2014-05-20 Shared Spectrum Company Method and system for man-made noise rejection detector
CN102354499B (en) * 2011-07-25 2017-12-19 中兴通讯股份有限公司 The method and apparatus for reducing noise
US20130040694A1 (en) 2011-08-10 2013-02-14 Babak Forutanpour Removal of user identified noise
US9528852B2 (en) * 2012-03-02 2016-12-27 Nokia Technologies Oy Method and apparatus for generating an audio summary of a location
FR2994495B1 (en) * 2012-08-10 2015-08-21 Thales Sa METHOD AND SYSTEM FOR DETECTING SOUND EVENTS IN A GIVEN ENVIRONMENT
US9275625B2 (en) * 2013-03-06 2016-03-01 Qualcomm Incorporated Content based noise suppression
US9384754B2 (en) * 2013-03-12 2016-07-05 Comcast Cable Communications, Llc Removal of audio noise
US9390170B2 (en) * 2013-03-15 2016-07-12 Shazam Investments Ltd. Methods and systems for arranging and searching a database of media content recordings
EP3078026B1 (en) * 2013-12-06 2022-11-16 Tata Consultancy Services Limited System and method to provide classification of noise data of human crowd
US20150179181A1 (en) * 2013-12-20 2015-06-25 Microsoft Corporation Adapting audio based upon detected environmental accoustics
CN103745728B (en) * 2014-01-08 2017-04-12 叶兰玉 Method and device for intelligent active noise reduction for house
CN106031138B (en) * 2014-02-20 2019-11-29 哈曼国际工业有限公司 Environment senses smart machine
US9837102B2 (en) * 2014-07-02 2017-12-05 Microsoft Technology Licensing, Llc User environment aware acoustic noise reduction
CN104517607A (en) * 2014-12-16 2015-04-15 佛山市顺德区美的电热电器制造有限公司 Speed-controlled appliance and method of filtering noise therein
US9728204B2 (en) * 2015-01-26 2017-08-08 Shenzhen Grandsun Electric Co., Ltd. Method and device for drawing a noise map
KR102087832B1 (en) 2015-06-30 2020-04-21 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에. 베. Method and device for generating a database

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003009273A1 (en) * 2001-07-16 2003-01-30 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e. V. Method and device for characterising a signal and for producing an indexed signal
CN101129086A (en) * 2005-02-23 2008-02-20 弗劳恩霍夫应用研究促进协会 Apparatus and method for controlling a wave field synthesis rendering device
CN101203917A (en) * 2005-06-22 2008-06-18 弗劳恩霍夫应用研究促进协会 Device and method for determining a point in a film comprising film data applied in chronological order
CN101236250A (en) * 2007-01-30 2008-08-06 富士通株式会社 Sound determination method and sound determination apparatus
CN101978424A (en) * 2008-03-20 2011-02-16 弗劳恩霍夫应用研究促进协会 Device and method for acoustic indication
KR20090122142A (en) * 2008-05-23 2009-11-26 엘지전자 주식회사 A method and apparatus for processing an audio signal
CN102918591A (en) * 2010-04-14 2013-02-06 谷歌公司 Geotagged environmental audio for enhanced speech recognition accuracy
CN104067339A (en) * 2012-02-10 2014-09-24 三菱电机株式会社 Noise suppression device
CN103795850A (en) * 2012-10-30 2014-05-14 三星电子株式会社 Electronic device and method for recognizing voice
US10373611B2 (en) * 2014-01-03 2019-08-06 Gracenote, Inc. Modification of electronic system operation based on acoustic ambience classification

Also Published As

Publication number Publication date
JP6654209B2 (en) 2020-02-26
KR102087832B1 (en) 2020-04-21
JP6602406B2 (en) 2019-11-06
KR20180022967A (en) 2018-03-06
US11003709B2 (en) 2021-05-11
EP3317879A1 (en) 2018-05-09
CA2990888A1 (en) 2017-01-05
WO2017001611A1 (en) 2017-01-05
KR20180025921A (en) 2018-03-09
KR102137537B1 (en) 2020-07-27
US20180121540A1 (en) 2018-05-03
JP2018528453A (en) 2018-09-27
US11880407B2 (en) 2024-01-23
CA2990891A1 (en) 2017-01-05
CN108028047B (en) 2022-08-30
US20180122398A1 (en) 2018-05-03
JP2018525664A (en) 2018-09-06
EP3317878A1 (en) 2018-05-09
EP3317878B1 (en) 2020-03-25
CN108028048A (en) 2018-05-11
EP3317879B1 (en) 2020-02-19
CN108028047A (en) 2018-05-11
WO2017001607A1 (en) 2017-01-05

Similar Documents

Publication Publication Date Title
CN108028048B (en) Method and apparatus for correlating noise and for analysis
EP3591633B1 (en) Surveillance system and surveillance method using multi-dimensional sensor data
US20210149939A1 (en) Responding to remote media classification queries using classifier models and context parameters
Valero et al. Gammatone cepstral coefficients: Biologically inspired features for non-speech audio classification
Jia et al. SoundLoc: Accurate room-level indoor localization using acoustic signatures
Hollosi et al. Enhancing wireless sensor networks with acoustic sensing technology: use cases, applications & experiments
US20230060936A1 (en) Method for identifying an audio signal
CN109997186B (en) Apparatus and method for classifying acoustic environments
US20230074279A1 (en) Methods, non-transitory computer readable media, and systems of transcription using multiple recording devices
KR101660306B1 (en) Method and apparatus for generating life log in portable termianl
CN111936056B (en) Detecting a subject with a respiratory disorder
Sharma et al. Smartphone audio based distress detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant