WO2012108218A1 - Système de détection d'événements - Google Patents

Système de détection d'événements Download PDF

Info

Publication number
WO2012108218A1
WO2012108218A1 PCT/JP2012/050122 JP2012050122W WO2012108218A1 WO 2012108218 A1 WO2012108218 A1 WO 2012108218A1 JP 2012050122 W JP2012050122 W JP 2012050122W WO 2012108218 A1 WO2012108218 A1 WO 2012108218A1
Authority
WO
WIPO (PCT)
Prior art keywords
sound
event detection
unit
detection system
pattern
Prior art date
Application number
PCT/JP2012/050122
Other languages
English (en)
Japanese (ja)
Inventor
永哉 若山
Original Assignee
日本電気株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電気株式会社 filed Critical 日本電気株式会社
Priority to JP2012556806A priority Critical patent/JP5626372B2/ja
Publication of WO2012108218A1 publication Critical patent/WO2012108218A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/16Actuation by interference with mechanical vibrations in air or other fluid
    • G08B13/1654Actuation by interference with mechanical vibrations in air or other fluid using passive vibration detection systems
    • G08B13/1672Actuation by interference with mechanical vibrations in air or other fluid using passive vibration detection systems using sonic detecting means, e.g. a microphone operating in the audio frequency range

Definitions

  • the present invention relates to an event detection system that detects an event that has occurred.
  • Patent Document 1 a technique for notifying that an abnormal event has been detected when an abnormal sound is detected is known (for example, see Patent Document 1).
  • a technique for detecting an event by sound such as the technique described in Patent Document 1, has a high merit in terms of privacy and cost.
  • a general-purpose microphone or a piezoelectric element can be used for the sensing device. Therefore, depending on the technology for detecting an event by sound, the cost can be significantly reduced as compared with a technology that uses a camera or a dedicated device as a sensing device.
  • the amount of sound data per second is about several tens of kilobytes.
  • a technique for reducing the weight of sound processing has already been established, as represented by Fast Fourier Transform. Such a reduction in processing cost also contributes to a reduction in processor cost.
  • a sensing device such as a microphone can be installed relatively easily as compared with a camera installed at a high place or a dedicated device installed by an expert. For this reason, a financial burden and a physical burden can be reduced.
  • the known technology for detecting an event by sound has a problem in sound identification accuracy. More specifically, depending on a known technique, an intruder may make adjustments such as reducing the generated sound, thereby overlooking an event that should be detected.
  • Patent Document 2 One method for solving such a problem is disclosed in Patent Document 2.
  • the device disclosed in Patent Document 2 is a sound pressure sensor that detects a change in air flow accompanying opening / closing of a door as a low frequency sound instead of a door opening / closing sound, or a pneumatic sensor that detects a change in air pressure as vibration. Is used to detect the opening and closing of the door.
  • the event detection system of the present invention detects an event that has occurred.
  • the event detection system includes a feature amount extraction unit that extracts a time series feature amount vector from signal information of acquired sound, a time series feature amount vector extracted by the feature amount extraction unit, and model data for each sound type Based on the sound identification unit that identifies the type of the acquired sound, the type of the acquired sound, and the sound generation pattern caused by the human action included before and after the acquired sound And an artificiality determination unit that determines the artificiality of the event that generated the acquired sound, and an output unit that performs output according to the determination result determined by the artificiality determination unit.
  • event detection system of the present invention human artifacts are determined by sound. Therefore, some event detection systems of the present invention are advantageous in terms of privacy and cost.
  • FIG. 2 It is a figure which shows an example of the utilization environment of the event detection apparatus which concerns on the 1st Embodiment of this invention. It is a figure which shows an example of the block configuration of the event detection apparatus shown in FIG. It is a figure which shows an example of the data stored in the artifact determination database shown in FIG. 2 in a table format. It is a figure which shows an example of the operation
  • FIG. 14 It is a figure which shows an example of the data stored in the artifact determination database shown in FIG. 14 in a table format. It is a figure which shows an example of the utilization environment of another example of the event detection apparatus of the 4th Embodiment of this invention. It is a figure which shows an example of the block configuration of the event detection apparatus shown in FIG. It is a figure which shows an example of the utilization environment of the event detection apparatus which concerns on the 5th Embodiment of this invention. It is a figure which shows an example of the data stored in the artifact determination database of the event detection apparatus shown in FIG. 18 in a table format. It is a figure which shows an example of the operation
  • FIG. 1 shows an example of the usage environment of the event detection apparatus 100 according to the first embodiment of the present invention.
  • the event detection device 100 detects an event that has occurred. More specifically, the event detection apparatus 100 is installed in the vicinity of the door 901. The event detection apparatus 100 collects opening / closing sounds of the door 901. In addition, the event detection apparatus 100 collects footstep sounds of human beings near the door 901. Further, the event detection device 100 may be an example of an “event detection system” in the embodiment of the present invention.
  • FIG. 2 shows an example of a block configuration of the event detection apparatus 100.
  • the event detection apparatus 100 includes a sound input element 110, a feature amount extraction unit 120, a sound identification unit 130, an artifact determination unit 140, an output unit 150, a sound identification model database 191 and an artifact determination database 192. The function and operation of each component will be described below.
  • Sound input element 110 converts sound into signal information. Then, the sound input element 110 sends the signal information of the sound to the feature amount extraction unit 120.
  • the sound signal information includes a continuous electric signal indicating a time-series change in sound pressure, and discrete sound data obtained by sampling and quantizing the electric signal.
  • Specific examples of the sound input element 110 include a microphone and a piezoelectric element.
  • the feature amount extraction unit 120 When receiving the sound signal information from the sound input element 110, the feature amount extraction unit 120 extracts a feature amount determined to be necessary for sound identification from the sound signal information, and extracts a time-series feature amount vector. Is sent to the sound identification unit 130 as data indicating Specific examples of the feature amount include sound pressure average, sound pressure maximum value, sound pressure dispersion, number of zero crossings, peak frequency, frequency spectrum, pitch, MFCC (Mel-Frequency Ceptstrual Coefficients), MP (Matching Persit), These primary differences, secondary differences, and the like. However, the present invention is not limited to this.
  • the sound identification unit 130 When the sound identification unit 130 receives data indicating the time-series feature amount vector from the feature amount extraction unit 120, the time-series feature amount vector indicated by the data and the data stored in the sound identification model database 191. Is analyzed and the type of the sound is specified. Then, the sound identification unit 130 sends data indicating the type of the sound to the artificial determination unit 140.
  • Specific identification methods include hidden Markov models (HMM: Hidden Markov Model), GMM (Gaussian Mixture Model), SVM (Support Vector Machine), likelihood determination based on statistical models such as Bayesian estimation, DP (Dynamic matching) And similarity determination based on distance measurement between data such as DTW (Dynamic Time Warping).
  • HMM Hidden Markov Model
  • GMM Gaussian Mixture Model
  • SVM Serial Vector Machine
  • likelihood determination based on statistical models such as Bayesian estimation
  • DP Dynamic matching
  • similarity determination based on distance measurement between data such as DTW (Dynamic Time War
  • the artifact determination unit 140 Upon receiving the data indicating the type of sound from the sound identification unit 130, the artifact determination unit 140 records the type of sound indicated by the data in time series. Next, the human artifact determination unit 140 determines the human artifact by comparing and analyzing the time series occurrence pattern of the recorded sound type with the identification sound and the characteristic sound pattern stored in the human artifact determination database 192. . Then, the artifact determination unit 140 sends data indicating the determination result to the output unit 150.
  • the artifact determination unit 140 includes a buffer for recording the type of sound generated during a predetermined time in order to determine the artifact using the time series generation pattern of the type of sound.
  • the output unit 150 When the output unit 150 receives data indicating the determination result from the artifact determination unit 140, the output unit 150 executes output according to the determination result indicated by the data. Specific examples include storage in a database, display, sound, identification sound presentation to the user through light, transmission of a mail to a specific mail address, transmission to a specific telephone number, transmission of a packet to an external device, etc. It is done.
  • the sound identification model database 191 stores data serving as a model for comparing and analyzing sound feature vectors and specifying the sound type.
  • model data include feature parameters, learned model parameters of a statistical model such as HMM, and the like.
  • HMM statistical model
  • the present invention is not limited to this.
  • model data is prepared for each sound type to be identified, both for the feature amount and the learned model parameter.
  • identification sounds, characteristic sound patterns, and anthropogenicity are stored in association with each other.
  • the characteristic sound pattern indicates a time-sequential generation pattern of characteristic sounds for determining artifacts. Artificiality indicates the presence or absence of civilization in the generation of the identification sound when the occurrence of the characteristic sound pattern is detected.
  • FIG. 3 shows an example of data stored in the artificial determination database 192 in a table format.
  • information “door open / close sound” is stored.
  • information “footstep: front and back” is stored.
  • information “present” is stored.
  • the present invention is not limited to this.
  • a combination of a plurality of patterns of data may be registered in the artifact determination database 192.
  • the characteristic sound pattern the case where binary determination is performed on the condition of the presence or absence of detection of the characteristic sound has been described, but the present invention is not limited thereto.
  • Likelihood determination based on a statistical model such as HMM, GMM, SVM, Bayesian estimation, or the like may be performed.
  • FIG. 4 shows an example of an operation flow of the event detection apparatus 100.
  • the sound input element 110 included in the event detection apparatus 100 converts the air vibration of the detected sound into sound signal information (S101).
  • the feature quantity extraction unit 120 extracts a feature quantity from the signal information of the converted sound (S102).
  • the sound identification unit 130 compares and analyzes the extracted feature amount and the model data registered in the sound identification model database 191 to identify the sound type (S103).
  • the human nature determination unit 140 records the identified sound types in time series, and generates the time series generation patterns of the sound types and the identification sounds and feature sound patterns registered in the human nature determination database 192.
  • a comparative analysis is performed to determine the artifact (S104).
  • the output unit 150 performs output based on the identification sound and the presence or absence of artifacts (S105).
  • the event detection apparatus 100 can determine the artifact of the identification sound while considering the privacy and suppressing the cost. That is, when the user opens / closes the door, footsteps are generated at a timing near the door opening / closing sound. On the other hand, when the door is opened / closed by wind or animal movement, footsteps are not generated at the timing near the door opening / closing sound. Therefore, it is possible to detect the door opening / closing performed artificially by using the event detection device 100 to recognize the difference in the generation pattern of the characteristic sound by using the door opening / closing sound as the identification sound and the footstep as the characteristic sound. . That is, it is possible to identify an artificial event without forcing the user to perform a special action.
  • FIG. 5 shows an example of the usage environment of the event detection apparatus 200 according to the second embodiment of the present invention.
  • the event detection apparatus 200 can identify the person who has made a sound in addition to human artifacts.
  • the event detection device 200 is installed at the entrance of the house and collects sounds generated at and near the entrance.
  • a sound indicating the operation is emitted.
  • footsteps are emitted from users 902 in and around the room.
  • the door lock 903 is opened and closed using the key 904, an opening / closing lock sound is generated.
  • a bell 905 is associated with the key 904, and a bell sound is emitted from the bell 905 when the key 904 is used. That is, the event detection apparatus 200 can detect “door open / close sound”, “human footsteps”, “open / close lock sound”, and “bell sound”, respectively.
  • the event detection device 200 may be an example of an “event detection system” in the embodiment of the present invention.
  • the user 902 When the user 902 opens the door 901, the user 902 approaches the door 901 before opening the door 901, and moves away from the door 901 after opening the door 901. For this reason, the footstep sound of the user 902 is emitted before and / or after the door 901 is opened and closed. If the door 901 is locked, the key 904 has a bell 905. For this reason, a bell sound is emitted before and / or after the locking sound when the door 901 is opened and closed.
  • FIG. 6 shows an example of a block configuration of the event detection apparatus 200.
  • constituent elements of the event detection apparatus 200 constituent elements having the same names and having the same reference numerals as the constituent elements of the event detection apparatus 100 show similar functions and operations.
  • the difference between the event detection device 200 and the event detection device 100 will be described.
  • a person who is associated with the identification sound and the characteristic sound pattern is further associated and registered.
  • the human artifact determination unit 240 determines a person who has emitted the identification sound based on the characteristic sound generation pattern registered in the human artifact determination database 292.
  • FIG. 7 shows an example of data stored in the artificial determination database 292 in a table format.
  • the artificiality determination information when “door opening / closing sound” and “opening / closing lock sound” are identified is registered. If “footsteps” are detected before and after the “door opening / closing sound” is identified, it is determined that the door opening and closing, which is the identification sound, is artificial, but the characteristic sound determines the person. It is stated that it can not be done.
  • “bell sound” is detected before and after the “opening and closing lock sound” is identified, the opening and closing lock that is the identification sound is artificial, and the opening and closing lock was further performed. It is described that it can be determined that he is Mr. A. Here, it is assumed that Mr.
  • A is the only person having the key 904. If the “bell sound” is not detected before and after the “open / close lock sound” is identified, the operation of the open / close lock itself is artificial, but the key 904 is used for the operation. I have not been told. Therefore, it can be determined that the suspicious person has performed the lock. From the above, if the identification sound is “open / closed lock sound” and “bell sound” is not detected as the characteristic sound pattern, the artifact is “present”, and the person who emitted the identification sound is “suspicious person” "Is determined.
  • FIG. 8 shows an example of the operation flow of the event detection apparatus 200. Differences between the operation flow of the event detection device 200 and the event detection device 100 will be described.
  • the human nature determination unit 240 records the identified sound types in time series, the time series generation patterns of the sound types, and the identification sounds and feature quantities registered in the human nature determination database 292. The pattern is compared and analyzed to further determine the person who has generated the identification sound in addition to the artifact (S201).
  • the event detection apparatus 200 can further identify the person who has emitted the identification sound. That is, when a bell sound is heard at the time of opening and closing, it indicates that the opening and closing lock has been performed using the key 904 associated with the bell 905. Can be guessed. On the other hand, if there is no bell sound at the time of opening / closing locking, it indicates that there is a high possibility that the opening / closing locking has been performed by some method other than the key 904. Can be guessed.
  • the identification sound as the opening / closing lock sound and the characteristic sound as the bell sound and recognizing the difference in the generation pattern of the characteristic sound by the event detection device 200, the person who performed the opening / closing lock can be identified. That is, it is possible to identify the person who has generated the event without forcing the user to perform a special action.
  • FIG. 9 shows an example of the usage environment of the event detection apparatus 300 according to the third embodiment of the present invention.
  • the event detection apparatus 300 can make a determination according to the occurrence pattern of the characteristic environment change in addition to the characteristic sound.
  • Characteristic environmental change is a human action or a specific person among environmental changes that can be estimated from changes in environmental feature quantities (specifically, light quantity, temperature, air volume, moisture content, radio wave intensity, etc.) It is intended to be caused by the operation of Specifically, the “lighting of the fluorescent lamp” in the case where the occurrence of a specific light amount change at the time of lighting of the fluorescent lamp is detected is highly likely to occur based on a human operation. For this reason, “lighting of the fluorescent lamp” corresponds to a characteristic environmental change. In the present embodiment, a specific description will be given with particular attention paid to the amount of light as the environmental feature amount.
  • the event detection device 300 may be an example of an “event detection system” in the embodiment of the present invention.
  • the event detection device 300 is installed in a room in the house and can collect sound and environmental changes generated in and around the room. In this embodiment, when the door 901 is opened and closed, a sound indicating the operation is emitted. Further, when the switch of the illumination 906 is operated, a switch sound and a light amount change peculiar to the illumination are generated. In other words, the event detection apparatus 300 can detect “a change in the amount of light due to illumination” in addition to the “door open / close sound” and the “switch sound”.
  • FIG. 10 shows an example of a block configuration of the event detection apparatus 300.
  • constituent elements of the event detection apparatus 300 constituent elements having the same names and having the same reference numerals as those of the event detection apparatus 100 show similar functions and operations.
  • the event detection apparatus 300 further includes a light amount detection element 360, a characteristic environment change identification model database 393, and a characteristic environment change identification unit 370, as compared with the event detection apparatus 100.
  • the light amount detection element 360 detects the amount of light around the event detection device 300 and converts it into a format for information processing.
  • Specific examples of the light amount detection element 360 include an illuminance sensor, a light amount sensor, and a photodiode.
  • the above-described “format for information processing” is referred to as an environmental feature amount.
  • the environmental feature quantity is extracted as time-series data in the same manner as the sound feature quantity vector.
  • the environmental feature amount may be prepared as a combination of a plurality of data.
  • the light quantity detection element 360 is composed of a plurality of elements and each element can detect the light quantity in a place different in real space
  • the light quantity obtained by each element may be treated as an independent environmental feature quantity. Good.
  • the amount of light is used as the environmental feature amount.
  • the characteristic environment change identification model database 393 stores data serving as a model for comparing and analyzing environmental feature amounts and specifying the presence and type of characteristic environment changes.
  • model data include learned model parameters of statistical models such as feature quantities and HMMs. For both feature quantities and learned model parameters, model data is prepared for each environmental change type to be identified.
  • the characteristic environment change identification unit 370 compares and analyzes the environmental feature quantity and the characteristic environment change identification model database 393, and specifies the presence and type of the characteristic environment change. Also, the characteristic environment change identification unit 370 notifies the human nature determination unit 340 of the specified environment change type. Specific identification methods include likelihood determination based on statistical models such as HMM, GMM, SVM, and Bayesian estimation, and similarity determination based on distance measurement between data such as DP matching and DTW, as with the sound identification unit 130. It is done. However, the present invention is not limited to these.
  • the anthropogenicity determination database 392 in the event detection device 300 further holds a characteristic environment change and its occurrence pattern in addition to the characteristic sound.
  • the artifact determination unit 340 performs a comparative analysis of the occurrence pattern of the characteristic environment change in addition to the identification sound and the identification sound, the characteristic sound pattern, and the occurrence pattern of the characteristic environment change registered in the artifact determination database 392. And determine human artifacts.
  • FIG. 11 shows an example of data stored in the artificial determination database 392 in a table format.
  • the door opening / closing that is the identification sound is detected. Describes that it is determined to be artificial.
  • FIG. 12 shows an example of the operation flow of the event detection apparatus 300.
  • the event detection apparatus 300 performs the following steps S301 to S303 in addition to steps S101 to S103 for performing sound identification similar to the operation flow of the event detection apparatus 100.
  • the light amount detection element 360 detects the amount of light around the event detection device 300 and converts it into an environmental feature amount (S301).
  • the characteristic environment change identification unit 370 determines whether or not there is a characteristic environment change and the type of environment change by comparing and analyzing the obtained environment feature quantity and the characteristic environment change identification model database 393 (S302).
  • the human artifact determination unit 340 records the identified sound type in the environment change type time series, the time series occurrence pattern of the sound type and the environment change type, and the identification sound registered in the civilization determination database 392. Then, the characteristic sound and the pattern of the characteristic environment change are compared and analyzed to determine the human artifact (S303). Finally, the output unit 150 performs output based on the identification sound and the presence or absence of artifacts (S105).
  • the event detection device 300 can determine the anthropogenicity of the event based on environmental changes in addition to sound. Taking door opening and closing as an example, footsteps may not be detected accurately depending on the installation location. Even if the switch sound when turning on the illumination can be detected, the causal relationship between the switch sound and the door opening / closing sound cannot be understood unless the type of the switch sound is known. For this reason, the switch sound cannot be used for the determination of the artifact of the door opening / closing sound.
  • the causal relationship between the opening / closing of the door and the lighting of the lighting is used to make an artificial opening / closing of the door Can be determined. That is, even if the sound alone cannot be used to specify the civilization or the person who emitted the sound, it is possible to specify the civilization or the person using the person-specific or person-specific environmental changes.
  • an electronic sound having periodicity in frequency or rhythm may be included.
  • Specific examples of the electronic sound include a buzzer sound, an electronic melody, an operation response sound of an electronic device, and a notification sound when a call is received.
  • Many electronic sounds are often emitted by explicit operation of electronic equipment.
  • the electronic sound exhibits the same operation, the characteristics of the emitted sound are almost the same every time it is generated. Therefore, it is possible to determine an artifact or a person with high accuracy.
  • FIG. 13 shows an example of a block configuration of an event detection apparatus 400 according to the fourth embodiment of the present invention.
  • constituent elements of the event detection apparatus 400 constituent elements having the same names as those of the constituent elements of the event detection apparatus 100 show similar functions and operations.
  • the event detection device 400 is a device capable of registering the characteristic sound acquired in the installation environment by learning.
  • the event detection apparatus 400 may be an example of an “event detection system” in the embodiment of the present invention.
  • the event detection device 400 further includes an artificial pattern learning unit 480 as compared with the event detection device 100.
  • the artificial pattern learning unit 480 selects a feature sound candidate having (or can be determined not to have) artificiality from the generation patterns of the sound types identified by the sound identifying unit 130. Furthermore, the artificial pattern learning unit 480 associates the candidate with the determination model when the candidate is detected, and registers the candidate in the artificial determination database 192 as a characteristic sound.
  • the operation of the artificial pattern learning unit 480 will be described more specifically by taking a bell sound as an example. It is difficult to determine whether or not the sound of a bell has an artifact by the sound of the bell itself. This is because the bell may be attached to the neck of an animal such as a cat, or may be attached to a human tool such as a key or a bag. That is, the presence or absence of artifacts in the characteristic sound changes depending on the installation environment.
  • the artificial pattern learning unit 480 registers the bell sound in the artificial determination database 192 as having low artificiality.
  • the artificial pattern learning unit 480 registers the sound of the bell in the artificial determination database 192 as having high artificiality.
  • the event detection apparatus 400 can select and register a sound that can be a characteristic sound from the sound types acquired in the installation environment.
  • selection criteria for feature sound candidates include a high occurrence frequency before and after the identification sound and the characteristic sound that can be determined by human artifacts, and the occurrence of sound before and after the identification sound determined to be artificial. It is done. Therefore, the characteristic sound peculiar to the installation environment can be determined.
  • FIG. 14 shows an example of a block configuration of the event detection apparatus 500.
  • constituent elements of the event detecting apparatus 500 constituent elements having the same names as those of the constituent elements of the event detecting apparatus 100 or the event detecting apparatus 400 have similar functions and operations.
  • the event detection device 500 may be an example of an “event detection system” in the embodiment of the present invention.
  • the event detection device 500 further includes an input unit 590 as compared with the event detection device 400.
  • the input unit 590 has a function of receiving input from a user and controlling the artificial pattern learning unit 480 based on the input. Further, the artificial pattern learning unit 480 determines selection of feature sound candidates and registration in the artificiality determination database 592 based on control from the input unit 590.
  • the event detection apparatus 500 can perform a learning operation at a timing desired by the user based on an input from the user. For this reason, learning according to the convenience of the user can be performed.
  • the learning operation can be controlled by designating before and after the occurrence of the characteristic sound that the user wants to learn.
  • the identification sound and the characteristic sound as a result of learning include sounds that are contrary to the user's intention (for example, noise specific to the installation environment), the learning can be canceled by the user's operation. It will be possible. Therefore, learning that reflects the user's intention is possible.
  • an output operation when a characteristic sound pattern is detected is specified and associated and registered.
  • the artificial pattern learning unit 480 further associates the designated output operation designated via the input unit 590 and registers it in the artificial nature determination database 592.
  • the output unit 150 further performs a specified output operation registered in the anthropogenicity determination database 592 as an operation based on the determination result.
  • FIG. 15 shows an example of data stored in the artificial determination database 592 in a table format. In the example illustrated in FIG. 15, it is determined that when footsteps are generated before and after the door opening / closing sound, it is determined that the operation has human artifacts, and the time and the identification sound are recorded as outputs.
  • the event detection apparatus 500 can register the characteristic sound and the output at the time of determination in the anthropogenicity determination database 592 by a user operation. Therefore, the output operation can be changed according to the importance of the identification sound or the characteristic sound. For example, it can be set to record in the storage device in the case of unimportant sound or frequent sound, to notify the user in the case of important sound, or to generate a trigger when operating other devices. .
  • the artificial pattern learning unit 480 further selects and learns a sound generation pattern specific to the environment of the installation destination, and the artificiality determination unit 140 further determines by weighting the unique sound generation pattern of the installation destination. To do.
  • the event detection apparatus 500 can determine by weighting the characteristic sound learned in the installation environment.
  • the sound unique to the installation site is likely to reflect the habits and intentions of the user at the installation site. Therefore, it is possible to determine the artifacts with higher accuracy.
  • FIG. 16 shows an example of the usage environment of the event detection apparatus 600.
  • Event detectors 600a, b, c,... Installed in a plurality of homes (hereinafter collectively referred to as event detectors 600) are connected via a network 990 and acquired by each event detector 600. Learning is automatically performed based on the sound type occurrence pattern.
  • the event detection device 600 may be an example of an “event detection system” in the embodiment of the present invention.
  • FIG. 17 shows an example of a block configuration of the event detection apparatus 600.
  • constituent elements of the event detecting apparatus 600 constituent elements having the same names as those of the constituent elements of the event detecting apparatus 100 or the event detecting apparatus 400 show similar functions and operations.
  • the event detection device 600 further includes a communication unit 699 as compared with the event detection device 400.
  • the communication unit 699 is connected to the event detection apparatus 600 other than its own apparatus via the network 101, and transmits / receives information related to the characteristic sound.
  • the event detection apparatus 600 can compare the sound generation patterns acquired by the own apparatus and other than the own apparatus, and identify the sound generation patterns that can be generated at the installation destination. Specifically, there is a method of registering as a characteristic sound pattern when a certain sound generation pattern is hardly registered other than the device itself. The sound unique to the installation site is likely to reflect the habits and intentions of the user at the installation site. Therefore, it is possible to determine the artifacts with higher accuracy.
  • the artificialness determination unit 140 further operates the artificial pattern learning unit 480 when it is determined that the amount of feature sound registration is insufficient for the determination by the artificiality determination unit 140 itself. Further, the output unit 150 further presents to the user that the determination of artifacts cannot be performed and that the artifact pattern learning unit 480 is operating.
  • the event detection device 600 can prompt the user to input sound so that further characteristic sounds can be learned when it is difficult to determine human artifacts. Therefore, it becomes possible to determine with higher accuracy.
  • the characteristic sound and the determination of the artificiality using the characteristic sound have been described.
  • the characteristic environment change and the person specific determination are also described. The same applies.
  • an electronic sound having periodicity in frequency or rhythm may be included.
  • Specific examples of the electronic sound include a buzzer sound, an electronic melody, an operation response sound of an electronic device, and a notification sound when a call is received.
  • Many electronic sounds are often emitted by explicit operation of electronic equipment. If the electronic sound exhibits the same operation, the characteristics of the emitted sound are substantially the same every time it is generated. Therefore, it is possible to determine an artifact or a person with high accuracy.
  • FIG. 18 shows an example of the usage environment of the event detection apparatus 700 according to the fifth embodiment of the present invention.
  • the event detection device 700 identifies a person who gives an instruction using electronic sound detection and activates a device / application specified in advance.
  • the event detection device 700 may be an example of an “event detection system” in the embodiment of the present invention.
  • the event detection device 700, the telephone 907, and the personal computer 908 are installed in the home.
  • the event detection device 700 and the personal computer 908 are connected via a network.
  • the connection between the event detection device 700 and the personal computer 908 may be a method using USB (Universal Serial Bus), Ethernet (registered trademark), wireless LAN (Local Area Network), Bluetooth (registered trademark), or the like.
  • the event detection device 700 is installed in a place where a ring tone from the telephone 907 can be heard.
  • the telephone 907 is set to emit a ringtone B when receiving an incoming call from the mobile phone 909.
  • FIG. 19 shows an example of data stored in the anthropogenicity determination database 792 of the event detection apparatus 700 in a table format.
  • the identification sound is silent, that is, when the house is absent, when the ringtone B registered as the characteristic sound is detected, the identification is made that the caller is Mr. B, and the PC 908 is used as an output via the network. Is registered to start the VPN (Virtual Private Network) connection.
  • VPN Virtual Private Network
  • FIG. 20 shows an example of an operation sequence of the telephone set 907, the personal computer 908, the mobile phone 909, and the event detection apparatus 700.
  • Mr. B is a resident at the installation site, and Mr. B carries a mobile phone 909 when going out.
  • the mobile phone 909 makes a call to the telephone 907 (S401).
  • the telephone set 907 receives a call, it rings the ringtone B associated with the mobile phone 909 as the caller (S402).
  • the event detection device 700 detects the ring tone B when no identification sound is input, that is, in a silent state.
  • the event detection apparatus 700 is registered based on the information registered in the artificial determination database 792 that the source of the information is Mr. B, and that the personal computer 908 is activated as an output to perform the VPN connection. (S123). Based on the detection result, the event detection apparatus 700 transmits a packet including the command to the personal computer 908 (S124). The personal computer 908 starts activation and VPN connection in accordance with the received command.
  • NAT Network Address Translation
  • NAPT Network Address Port Translation
  • the telephone ringing tone serving as an event trigger is set to be generated when an incoming call is received from a specific mobile phone 909, and unauthorized operations from a third party can be prevented. Further, since the distribution and setting of the telephone ringtone is a function realized by the existing telephone 907, the introduction of new equipment can be minimized. As described above, it is possible to solve the problems of user knowledge, cost, and network security, which are problems in remote control of home devices.
  • a program for realizing all or part of the functions of the event detection apparatus according to the embodiment of the present invention is recorded on a computer-readable recording medium, and the program recorded on the recording medium is read into a computer system and executed. By doing so, you may process each part.
  • the “computer system” here includes an OS and hardware such as peripheral devices.
  • a computer-readable recording medium refers to a portable medium such as a magneto-optical disk, a ROM, or a nonvolatile semiconductor memory, or a storage device such as a hard disk built in a computer system.
  • Computer-readable recording medium means a program that dynamically holds a program for a short time, such as a communication line when transmitting a program via a network such as the Internet or a communication line such as a telephone line.
  • a volatile memory in a computer system that serves as a server or a client in this case includes a program that holds a program for a certain period of time.
  • the above program may realize part of the functions described above, and may further realize the functions described above in combination with a program already recorded in the computer system.
  • sound collection / feature extraction / identification, artifact determination, and output are all performed by the event detection device. It may be realized separately.
  • the sound collection to identification can be performed with one device, and the determination and output of human artifacts can be performed with another device, and the identification sound and the characteristic sound or similar information can be notified via the network. Good.
  • the artifact determination unit further includes a generation pattern of an electronic sound having periodicity in sound or frequency or rhythm caused by a human action, or an amount of learning of environmental change.
  • the output unit further utilizes the fact that the artificial pattern cannot be determined and the artificial pattern learning function is operating. May be presented to the person.
  • the present invention can detect an event using a sound generation pattern, it can be applied to any environment where a microphone can be installed.
  • a microphone can be installed.
  • the explanation was made with the crime prevention use in general homes in mind, it is naturally applicable to other uses as well. For example, it is possible to detect an abnormal state represented by a caregiver's habit or falling by applying to a home having a care facility and a care receiver.
  • application to hospitals, public facilities, and commercial facilities can be expected.
  • Event detection apparatus 110 Sound input element 120 Feature-value extraction part 130 Sound identification part 140 Artificiality judgment part 150 Output part 191 Sound identification model database 192 Artificiality judgment database 200
  • Event detection apparatus 240 Artificiality judgment part 292 Artificiality judgment database 300
  • Event detection device 340 Artificiality determination unit 360 Light quantity detection element 370 Characteristic environment change identification unit 392 Artificiality change database 393 Characteristic environment change identification model database 400
  • Event detection device 480 Artificial pattern learning unit 500
  • Event detection device 590 Input unit 592 Artificial Sex determination database 600 Event detection device 699 Communication unit 700
  • Event detection device 792 Artificiality determination database 901 Door 902 User 903 Door lock 904 Key 905 Bell 906 Lighting 907 Telephone 908 Personal computer 909 Mobile phone 990 network

Abstract

L'invention concerne un système de détection d'événements qui détecte des événements qui se sont produits. Le système de détection d'événements comprend une unité d'extraction de valeur de caractéristique afin d'extraire une série temporelle de vecteurs de valeurs de caractéristiques à partir d'informations de signaux d'un son acquis, une unité d'identification de son qui identifie le type de son acquis à l'aide d'une analyse comparative de la série temporelle de vecteurs de valeurs de caractéristiques extraite par l'unité d'extraction de valeur de caractéristique et de données de modèle de chaque type de son, une unité de détermination d'anthropogénicité afin de déterminer l'anthropogénicité de l'événement qui a généré le son acquis en fonction du type de son acquis et du motif d'occurrence des sons dus à une action humaine contenu avant ou après le son acquis, et une unité de sortie afin de réaliser une sortie en fonction des résultats de détermination de l'unité de détermination d'anthropogénicité.
PCT/JP2012/050122 2011-02-09 2012-01-05 Système de détection d'événements WO2012108218A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2012556806A JP5626372B2 (ja) 2011-02-09 2012-01-05 事象検知システム

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2011025795 2011-02-09
JP2011-025795 2011-02-09

Publications (1)

Publication Number Publication Date
WO2012108218A1 true WO2012108218A1 (fr) 2012-08-16

Family

ID=46638438

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2012/050122 WO2012108218A1 (fr) 2011-02-09 2012-01-05 Système de détection d'événements

Country Status (2)

Country Link
JP (1) JP5626372B2 (fr)
WO (1) WO2012108218A1 (fr)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014190724A (ja) * 2013-03-26 2014-10-06 Kddi Corp 空間状態判定装置
WO2015118318A1 (fr) * 2014-02-04 2015-08-13 Eddy Labs Ltd Procédés et systèmes de détection d'événements dans un environnement fermé
JP2016126479A (ja) * 2014-12-26 2016-07-11 富士通株式会社 特徴音抽出方法、特徴音抽出装置、コンピュータプログラム、配信システム
JP2016207199A (ja) * 2015-04-24 2016-12-08 公立大学法人広島市立大学 転倒検知装置及び転倒判定方法
KR20170124448A (ko) * 2016-05-02 2017-11-10 김태준 보안 시스템 작동 방법
JP2019023916A (ja) * 2018-10-10 2019-02-14 ヤフー株式会社 推定装置、推定方法および推定プログラム
TWI712994B (zh) * 2019-12-09 2020-12-11 聯騏技研有限公司 具消防警示的門禁系統及消防警示方法

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000268265A (ja) * 1999-03-15 2000-09-29 Chuo Joho Kaihatsu Kk マイクを使用した監視装置
JP2004265372A (ja) * 2003-01-09 2004-09-24 Masayuki Nemoto 防犯装置
JP2005328410A (ja) * 2004-05-17 2005-11-24 Sony Corp 音響装置および音響装置を用いた監視方法
JP2009075983A (ja) * 2007-09-21 2009-04-09 Kyocera Corp 携帯端末および異常通知方法
JP2010071773A (ja) * 2008-09-18 2010-04-02 Transvirtual Inc 開閉部材の開閉動作認識装置、開閉部材の開閉動作認識方法
JP2010232888A (ja) * 2009-03-26 2010-10-14 Ikegami Tsushinki Co Ltd 監視装置

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000268265A (ja) * 1999-03-15 2000-09-29 Chuo Joho Kaihatsu Kk マイクを使用した監視装置
JP2004265372A (ja) * 2003-01-09 2004-09-24 Masayuki Nemoto 防犯装置
JP2005328410A (ja) * 2004-05-17 2005-11-24 Sony Corp 音響装置および音響装置を用いた監視方法
JP2009075983A (ja) * 2007-09-21 2009-04-09 Kyocera Corp 携帯端末および異常通知方法
JP2010071773A (ja) * 2008-09-18 2010-04-02 Transvirtual Inc 開閉部材の開閉動作認識装置、開閉部材の開閉動作認識方法
JP2010232888A (ja) * 2009-03-26 2010-10-14 Ikegami Tsushinki Co Ltd 監視装置

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014190724A (ja) * 2013-03-26 2014-10-06 Kddi Corp 空間状態判定装置
WO2015118318A1 (fr) * 2014-02-04 2015-08-13 Eddy Labs Ltd Procédés et systèmes de détection d'événements dans un environnement fermé
JP2016126479A (ja) * 2014-12-26 2016-07-11 富士通株式会社 特徴音抽出方法、特徴音抽出装置、コンピュータプログラム、配信システム
JP2016207199A (ja) * 2015-04-24 2016-12-08 公立大学法人広島市立大学 転倒検知装置及び転倒判定方法
KR20170124448A (ko) * 2016-05-02 2017-11-10 김태준 보안 시스템 작동 방법
KR101899244B1 (ko) * 2016-05-02 2018-09-17 김태준 보안 시스템 작동 방법
JP2019023916A (ja) * 2018-10-10 2019-02-14 ヤフー株式会社 推定装置、推定方法および推定プログラム
TWI712994B (zh) * 2019-12-09 2020-12-11 聯騏技研有限公司 具消防警示的門禁系統及消防警示方法

Also Published As

Publication number Publication date
JPWO2012108218A1 (ja) 2014-07-03
JP5626372B2 (ja) 2014-11-19

Similar Documents

Publication Publication Date Title
JP5626372B2 (ja) 事象検知システム
US10176705B1 (en) Audio monitoring and sound identification process for remote alarms
Ntalampiras et al. Probabilistic novelty detection for acoustic surveillance under real-world conditions
WO2018018906A1 (fr) Procédé et système de surveillance de commande d'accès vocal et de surveillance d'environnement silencieux
CN112074900B (zh) 用于自然语言处理的音频分析
US20190103005A1 (en) Multi-resolution audio activity tracker based on acoustic scene recognition
Vacher et al. Complete sound and speech recognition system for health smart homes: application to the recognition of activities of daily living
US10706848B1 (en) Anomaly detection for voice controlled devices
CN108604176A (zh) 信息处理装置、信息处理方法及程序
US11162280B2 (en) Door lock/camera apparatus, system and method
US20230186904A1 (en) Systems and methods for classifying sounds
US20210201889A1 (en) System and Method for Determining Occupancy
Song et al. Deep room recognition using inaudible echos
KR20190024254A (ko) 도어벨 장치
KR101995443B1 (ko) 화자 검증 방법 및 음성인식 시스템
KR20130097490A (ko) 음향 정보 기반 상황 인식 장치 및 방법
WO2012121856A1 (fr) Procédé et système de reconnaissance de sons
US11521626B2 (en) Device, system and method for identifying a scene based on an ordered sequence of sounds captured in an environment
US20200143823A1 (en) Methods and devices for obtaining an event designation based on audio data
CN112700765A (zh) 辅助技术
JP2013225248A (ja) 音識別システム、音識別装置、音識別方法およびプログラム
US11620997B2 (en) Information processing device and information processing method
KR101520446B1 (ko) 구타 및 가혹행위 방지를 위한 감시 시스템
US11437019B1 (en) System and method for source authentication in voice-controlled automation
Malik et al. Speaker identification based home automation system for aging populations through speech recognition

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12745137

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2012556806

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12745137

Country of ref document: EP

Kind code of ref document: A1