SG178563A1 - Method and system for event detection - Google Patents

Method and system for event detection Download PDF

Info

Publication number
SG178563A1
SG178563A1 SG2012013140A SG2012013140A SG178563A1 SG 178563 A1 SG178563 A1 SG 178563A1 SG 2012013140 A SG2012013140 A SG 2012013140A SG 2012013140 A SG2012013140 A SG 2012013140A SG 178563 A1 SG178563 A1 SG 178563A1
Authority
SG
Singapore
Prior art keywords
classifier
event
unit
units
evidence
Prior art date
Application number
SG2012013140A
Inventor
Karianto Leman
Huy Dat Tran
Mun Kai Derrick Loh
Haizhou Li
Melvin Wong
Feng Gao
Xin Yan
Original Assignee
Agency Science Tech & Res
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Agency Science Tech & Res filed Critical Agency Science Tech & Res
Priority to SG2012013140A priority Critical patent/SG178563A1/en
Publication of SG178563A1 publication Critical patent/SG178563A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/254Fusion techniques of classification results, e.g. of results related to same input data
    • G06F18/256Fusion techniques of classification results, e.g. of results related to same input data of results relating to different input data, e.g. multimodal recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B17/00Fire alarms; Alarms responsive to explosion
    • G08B17/12Actuation by presence of radiation or particles, e.g. of infrared radiation or of ions
    • G08B17/125Actuation by presence of radiation or particles, e.g. of infrared radiation or of ions by using a video camera to detect fire or smoke

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Testing Or Calibration Of Command Recording Devices (AREA)
  • Alarm Systems (AREA)

Abstract

A method and system for event detection. The system comprising: a plurality of sensors; and respective classifier units coupled to the sensors for processing respective sensor signals from the sensors for the event detection; wherein at least one of the classifier units is adapted for detecting an event based on the sensor signal processed in said at least one classifier unit as a main evidence and for issuing sub-evidence queries to the other classifier units for facilitating the event detection.

Description

1 - oo "METHOD AND SYSTEM FOR EVENT DETECTION
FIELD OF INVENTION ~~ oo Tn The invention broadly relates to a method and system for event detection "BACKGROUND
Le The use of closed-circuit television (CCTV) ‘cameras in lifts (elevators) for the prevention of vandalism, crime and other undesirable acts has gained growing popularity in many countries as they can provide a video recording facility for post-incident ~ investigation. This has resulted in more lifts being retrofitted with such cameras.
Ca : Currently, in addition to cctv cameras, passenger ifts have also been retrofitted ~~ with other types of sensors (e.g. chemical sensors for urine detection) in order to capture vandalism, crime and other undesirable acts. However, some shortcomings include: i. . Unsatisfactory performance of the chemical sensors. : ii. The ability of culprits to get around the system once they get acquainted with the sensors. © © iii. More complex events cannot be automatically detected (e.g.: robbery and : violence against the elderly, children, and women). In addition, the recorded video may not be able to clearly identify the perpetrators. iv. Inability to prevent acts that lead to the damaging of lifts. Acts that cause : damage to lifts typically go undetected. Subsequent investigation by 25 . reviewing the CCTV video is usually tedious and inconclusive. ~~
EN “~v. In crimes against soft targets (e.g. elderly, children, and women), there - oo ." are generally no means of taking immediate ‘action. Currently, only post- - incident investigation can be done ‘upon the filing of police's report. } "© Thereafter a lengthy information and evidence gathering process may follow, coupled with atedious process to apprehend the culprit. ‘Advancements in computational science, coupled with cheaper and more “powerful computing platforms, enables information from cctv cameras to be used to detect anti-social behaviour such as urinating in the lift. Algorithms can be designed to o detect this event in real-time. Currently, a conductive loop is deployed for such detection;
BE however, performance is far from acceptable while maintenance and installation are
E fv relatively cumbersome. SUE . So : :
CnLeRIIia nll LL So RET ) sis : iy - More recently, there are works involving the combination of video and audio for hE the purposes of background-foreground separation, localization of humans, and ~ _-event/behaviour detection. However, these works have separate audio and video : processing threads that function. independently. The respective classifier results are combined to reach a final decision. In such methods using the aggregation of classifier na results; corrupted signals from the other sensors can sway the output and lead to an oo erroneous result. Useful. information can be lost as some: types of sensors are more . a strongly related to the characteristics of certain events than other types of sensors. oo ‘A need therefore exists to provide a method and system for event detection that .. seeksto address at least one of the abovementioned problems.
SUMMARY oo
According to the first aspect of the present invention, there is provided a system for event detection, comprising: a plurality of sensors; and respective classifier units coupled to the sensors for processing respective sensor signals from the sensors for the event detection; wherein at least one of the classifier units is adapted for detecting an 25. event based on the sensor signal processed in said at least.one classifier unit as a main - evidence and for issuing sub-evidence queries to the other classifier units for facilitating
So the event detection. oo - ol | g : y : ‘The at least one classifier unit may comprise a categorisation unit for categorising the sensor signal into event categories. -
So 3 oo : | oo The categorisation unit may be adapted to use one or more parameters from the oC ) sub-evidence queries to the other classifier units in categorising the sensor signal into 7: the event categories. Lo : : : : The at least one classifier unit may comprise-a dominant signature verification ~~ : 3 "unit for verifying a dominant signature in the sensor signal for the event detection. " So od Nags ) * The dominant signature verification unit may be adapted to use one or more ~~ parameters from the sub-evidence queries to the other classifier units in verifying the dominant signature in the sensor signal. oo
BE The at least one classifier ‘unit may comprise a- machine learning unit for classification processing for the event detection. © : SE Co 45 + -The-machine-learning unit-may-be-adapted-to-use-one-or more-parameters-from the sub-evidence queries to the other classifier units in the classification processing.
According to a second aspect of the present invention, there is provided a method of event detection comprising the steps of: obtaining respective sensor signals from a plurality of sensors; and processing the respective sensor signals for the event detection using respective classifier units coupled to the plurality of sensors, wherein an event is detected based on the sensor signal processed in one classifier unit as a main evidence; and issuing sub-evidence queries to the other classifier units for facilitating the
I event detection using said one classifier unit. oo
SESE Co CL oo
The method may comprise categorising the sensor signal into event categories © using a categorisation unit of said one classifier unit. ~
The method may comprise using one or more parameters from the sub-evidence queries to the other classifier units in categorising the sensor signal into the event categories. = EE oo oo ;
Cee a EE . oo | The method may comprise verifying a dominant signature in the sensor signal for ~~ the event detection using a dominant signature verification unit of said one classifier unit ~ 5. . Sn The method may comprise using one or more parameters from the sub-evidence queries to the other classifier units in verifying the dominant signature in the sensor © osignal oo
Le - The method -may comprise ‘performing classification ‘processing for the event 10° detection using a machine leaming unit of said one classifier unit. So gs : CL The method may comprise using one or more parameters from the sub-evidence ~~ queries to the other classifier units in the classification processing. So oo
According to a third aspect of the present invention, there is provided a data storage medium having stored thereon computer program code means for instructing a ~~ computer system to execute a method of event detection, and as defined in the second aspect. -
Co Co SR " BRIEF DESCRIPTION OF THE DRAWINGS
Embodiments of the invention will be better understood and readily apparent . to one of ordinary. skill in the art from the following written description, by way of example only, and in conjunction with the drawings, in which: © Figure.1 is a schematic drawing illustrating the configuration of an audio and video awareness system for lift monitoring, according to an example embodiment of the ~ presentinvention. Co ~ _. Figure 2A is a flow chart illustrating the system architecture and processing steps of an audio ‘and video awareness system for lift monitoring, according to an example "embodiment of the present invention. ro. n ~ - :
Figure 2B is a flow chart illustrating the system architecture and processing steps of an
REN audio and video awareness system for lift monitoring, according to ‘another example = ~~ embodiment of the presentinvention.. © Co : Figure 2C'is a flow chart illustrating the system architecture and processing steps of an - - "audio and’ video awareness system for lift monitoring, ‘according to a further example = ~~ embodiment of the present invention. EE | SR ~~» Figure 3is a flow chart illustrating a method for event detection, according to an 10. example embodiment of the present invention. LL 2 : Figure 4 is a schematic of a computer system for implementing the method and | 3 “system for event detection. SEE CLI oo :
DETAILED DESCRIPTION Co Co
According to embodiments of the present invention, there is provided an audio and video awareness system for lift monitoring that comprises an artificial intelligence (Al) system that fuses the computations of signals from a plurality of sensor types. In the . following description, the system is described in relation to a lift (elevator). However, it will be appreciated by a person. skilled in the art that the audio and video awareness system described herein may be used in any suitable environment or location. oo 25° | Some portions of the description which follows are explicitly or implicitly presented in- terms of algorithms and functional or symbolic representations of . operations on data within a computer memory. These algorithmic descriptions and : functional or symbolic representations are the means used by those skilled in the data processing arts to convey most effectively the substance of their work to others skilled in theart. An algorithm is here, and generally, conceived to be a self-consistent sequence of steps leading to a desired result. The steps are those requiring physical manipulations - of physical quantities, such as electrical, magnetic or optical signals capable of being. ~ stored, transferred, combined, compared, and otherwise manipulated. oo -
Ss : - Le “ - : - Unless specifically stated otherwise, and as apparent from the following, it willbe
Se appreciated that throughout the present specification, discussions utilizing terms suchas ~~ scanning’, “calculating”, ‘determining’, ‘replacing’, ‘generating’, ‘initializing’, ~~ “outputting”, or the like, refer to the action and processes of a computer system, or - : ‘similar electronic device, that manipulates and transforms data represented as physical - 3 ‘quantities within the computer system into other data similarly represented as physical ok quantities’ within the computer system or other information. storage, transmission or “display devices. EE fo IE Ln an oe = The present specification also discloses apparatus for performing the operations I = - ; of the methods. Such apparatus may be specially constructed for the required purposes, | oo 7 or may comprise a genéral purpose computer or other device selectively activated or - "reconfigured by a. computer program stored in the computer. The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general purpose machines may be used with programs in . accordance with the teachings herein. Alternatively, the construction of more specialized apparatus to perform the required method steps may be appropriate. The structure of a conventional general purpose computer will appear from the description below.
Co So in addition, the present specification also implicitly discloses a computer - program, in that it would be apparent to the person skilled in the art that the individual co . steps of the method described herein may be put into effect by computer code. The
Lo computer program is not intended to be limited to any particular programming language and implementation thereof. - It will be appreciated that a variety of programming x languages and coding thereof may be used to implement the teachings of the disclosure contained herein. Moreover, the computer program is not intended to be limited to any : particular control flow. There are many other variants of the computer program, which ~~ can use different control flows without departing from the spirit or scope of the invention. 0 Ce .. n : Furthermore, one or more of the steps of the ‘computer program may be : : performed in‘parallel rather than sequentially. Such a computer program may be stored 3 on any computer’ readable medium. The computer readable medium may . include :
: } storage devices such as magnetic or optical disks, memory chips, or other storage - . “devices suitable for interfacing with a general purpose computer. The computer readable ; - medium’ nay “also include a hard-wired medium such as exemplified in the Internet oo system; or witeless medium such as exemplified in the GSM mobile telephone system. ~~
The computer program when loaded and executed on'such a general-purpose computer oo
CL effectively results in an apparatus that implements the steps. of the preferred method.
LL The invention may also be implemented as hardware modules. More particular, . } (in the hardware sense, a module is a functional hardware unit designed for use with other components or modules. . For example, a module ‘may be implemented using oo discrete electronic components, or it can form a portion of an entire electronic. circuit Cu oo
B on such as'an Application Specific Integrated Circuit (ASIC). Numerous other possibilities 3 : i exist. Those skilled in the art will appreciate that the. system can also be implemented - "as a combination of hardware and software modules. | Lo Co | oo oo oo oo Figure 1 is a schematic drawing illustrating the configuration of an audio and ~~ video awareness system for lift monitoring, designated generally as reference numeral ~~ 100, according to an example embodiment of the present invention.
The system. 100 comprises 2 lifts 102/112, an analytical engine 122 oo (implemented on a computer), a communication module 124 and a lift control module 126. Each of the 2 lifts 102/112 comprise 3 types of sensors: a closed-circuit television (CCTV) camera 108/118, two contact microphones 106a/b / 116a/lb embedded on the : lower parts of the lift walls, an audio microphone 104/114 embedded on the lift ceiling, 25° anda laminated lift floor 110/120 with a relatively high acoustic signal-to-noise ratio. The audio microphones 104/114 provide acoustic audio signals, the contact microphones ~106a/b / 116alb provide vibration signals and the (CCTV) cameras 108/118 provide video signals. The lift control module 126 is connected and in communication with the analytical engine 122 and/or communication module 124 to facilitate control of the lifts : operation. It will be appreciated bya person skilled in the art that although the system 100 comprises 2 lifts 102/112, any suitable number of lifts can be monitored. In other words, sensors from multiple lifts can be fed to the analytical engine 122. 3 oo Co ‘The cctv cameras 108/118 are preferably analogue CCTV cameras of PAL "specification. The analogue signals from the: CCTV cameras 108/118, contact +" microphones 106a/b: 116a/b and audio microphones 104/114 are sent to the analytical a : engine 122 located, for example, in a lift motor room a distance away, for example, through: co-axial cables 128a — d. Appropriate amplifiers (not shown) can be applied
Co % ) along the way to_ maintain signal quality. The video signals are digitized by hardware ~~ encoders (which may be located within the analytical engine 122) into- digital streams ~~ - Lo such as H.264. The audio and contact signals are digitized by audio processors (which may also. be located within the analytical engine 122) for digitization into audio streams eg. in the WAV format. These digital video and audio streams are synchronized using oo the analytical engine's 122 time stamp at the time of their digitization. oo z Cl The analogue signals from the 3 types of sensors are processed at the analytical ~~ “engine 122, wherein information from one of the types of sensors used in a master classifier and information from the other sensors is called by the master classifier from slave classifiers associated with the remaining types of sensors. For example, the - system may be primarily driven. by audio analytics in that the microphone signal is the . main determining factor in deducing the occurrence of a particular event. Upon a suspicious detection from an audio source, computational results from video feeds can be used to verify the validity of the detection from audio signal. This entails an ‘intermediate process involving technical queries to other sensors to build-up the
Lo evidence for deducing the occurrence of the particular event. It will be appreciated that “more features can be explored. in proportion to the capability of the computational platform available. ~~ :
Co The system 100 can also comprise an alarm system which is raised when an event occurs and users can simultaneously apply two. types of alarm mechanisms:
Distributed mechanism: When an event occurs, the analytical engine 122 directly - sends out alarm notifications via the communication module: 124 to a designated user's mobile phone through a 3G connection. The user can. perform a verification check by playing back the cctv video feed prior. to taking an appropriate action. The system 100 advantageously allows building owners to manage. their lifts more comprehensively.
Ce
Ce Monitoririg and control can be executed remotely, for example, through a 3G based a Re communication system. In contrast; in the prior art, users have to be at designated C monitoring stations since connectivity through GPRS can only notify the users through ge : short messages. Co | | | B 8 - : : IEE RE
Bo AEE | - co Co Centralized mechanism: When an event occurs, a notification and a respective video segment are sent via the communication module 124 to a web server (not shown) ~ ata central location. This web server can then use artificial intelligence to decide which : : . - user(s): to notify based on “the rules-of-engagement it “is programmed with. The . recipient(s) of these alarms can then perform their verifications by connecting to the web
Csever LIT i. ; Figure 2A is a flow chart illustrating the system architecture/processing- steps of an audio and video awareness system for lift monitoring, generally designated as 16 reference numeral 200a, according to an example embodiment of the present invention. : The system comprises a plurality ‘of classifiers: an audio classifier 252, a vibration ~ classifier 254 and a video classifier 256. The classifiers 252, 254 and 256 can perform © either as a master and/or a slave classifier for different event detection tasks. in example embodiments, depending on the type of event, information from a first type of sensor is used in a master. classifier and information from the remaining types of sensors is called : by the master classifier from slave classifiers associated with the remaining types of sensors. The master classifier is, used to ‘suspect the occurrence of an event using the © main evidence from one sensor, while the other classifiers are used to confirm the - occurrence of the event using sub-evidences from other sensors (wherein the sub- evidences are not directly linked to the event). Advantageously, the detection of each event is processed independently. However, at any one time, the classifiers 252, 254 ~and 256 are running concurrently to detect events. In addition, the independent processing steps in each of the classifiers 252, 254 and 256 are preferably synchronized using a common time. Su oo : 30. elt co I
Co = | ‘In example embodiments, the master classifier queries the slave classifier onthe = - ~~ presence of low-level, mid-level and high-level information. no : oo
Ce : oo
LE Examples of low-level queries include: Co : | | oo - : - - . Changes in scene or the presence: of foreground objects (from video signal) - oo 17+" The basic shape of the foreground objects (video) | RE + Thepitchevel of the signal (audio/contact) EE : 5 : . Le . The bandwidth of the signal in the frequency domain | - oo
Co ) | Examples of mid-level queries include: : oo .
CT Presence of human shape/model and number of people (video) ~~ — Presence of human screams/shouts (audio) SER -
S10 ce Presence of banging impacts (contact) : oo
Ce Presence of rattling sound (contact) ; : So
BEE Examples of high-level queries include: + R Co Lo BN : « Presence of aggressive motion (video) : «Human posture estimation (video) = Presence of violent actions (video)
Co One example of the use of the audio classifier 252, functioning as a master classifier, can be for the detection of crimes against soft targets (e.g. the elderly, children and women) 212. At step 202, an audio sound is detected. At step 204, the audio signal goes through a number of sampling and pre-processing tasks that include the removal of ~~ noiseand signal re-sampling. Sh } Se Co oo EE ‘Categorization 206 is performed by the audio classifier 252 wherein the audio signal is categorized into ‘silent/non-silent’, ‘'voice/non-voice', and ‘normal/abnormal’.
Different processing threads are taken from the categorization results. For instance,
Co ‘non-voice’ audio signals are further investigated into common metadata of normal situations such as footsteps, movement sound, etc, Deviations from the normal oo situations initiate an identification of parameters that are precursors to an event. Queries : | : (low-level), e:g. indicated at numeral 260 to the video classifier 256 (slave classifier) for | - ~ foreground object detection 229 and human model detection 230 are issued to assist in ; 7 the accuracy of this categorization. If the audio signal is “silent” or “normal”; or no object is detected, further classification may not be performed.’ Techniques for foreground
Cn nn RE Co
Cn object detection and human model detection are understood in the art and will notbe oe. described in further detail. Reference is made, for example, to (i) "Adaptive Background LL "Subtraction with Multiple Feedbacks for Video. Surveillance," Inti Symp Visual © Computing, (Lake Tahoe, Nevada, USA. Dec. 5-7, 2005), pp. 380-367; and (i) Liyuan ~~ oC 5 : : Li; ‘Weimin Huang, Gu, LY.H, Leman, K, Qi Tian, “Principal color representation for oo tracking persons,” IEEE International Conference on Systems, Man and Cybernetics, 5-8
Ln Oct. 2003, vol 1, pp. 1007 - 1012), the contents of which are herein incorporated by reference; Loan oo 10° Verification of dominant signature 208 is also carried out at the audio classifier 252. The audio signal is analyzed for the presence of sounds characteristic of crimes : - against soft targets such as shouts and cries (e.g. short duration, non-repetitive high oo
Co pitch sounds with narrow spectral). Dominant Signatures in the example embodiment N } are knowledge-based parameters that define the syntax of a particular event. The syntax describes information (e.g. low/mid-level information) of a particular signal. :
Machine Learning Classification 210 using weighted features from audio and "vibration signals is also performed at the audio classifier 252. The sound is compared to sample sounds characteristic of crimes against soft targets to detect the presence of crimes against soft targets. Machine Learning classification advantageously facilitates a more comprehensive detection of signals compared to the verification of dominant signaturés alone. The machine learning classifier is preferably a Support Vector Machine (SVM) classifier with a super feature vector concatenated from MFCC, Prosody, - : Perceptual Linear Prediction Coding Cepstra (PLPCC) and MPEG-7 descriptor. : : A Support Vector Machine (SVM) classifier is understood in the art and will not be described in further detail. Reference is made, for example, to Vladimir Vapnik, The
Nature of Statistical Learning Theory. Springer-Verlag, 1995. ISBN 0-387-98780-0, the
Lo contents of which are herein incorporated by reference. - on : - Ce | N oo
MFCC is understood in the art and will not be described in further detail. R ~~ Reference is made, for example, to Fang Zheng, Guoliang Zhang and Zhanjiang Song © (2001), "Compatison of Different Implementations of MFCC," J. Computer Science &
oo - Technology, 16(6): 582-589), the contents of: which are herein incorporated by : reference.’ : 5 - o - - : | ] ©". prosody is understood in the art and will not be described in further detail.
Refererice is made, for example, to Hermansky, H. (1 990) "Perceptual linear predictive (PLP) analysis of speech", J. Acoust. Soc. Am., 87(4), pp.1738-1752, the contents of
B | which are herein incorporated by reference. : © Perceptual Linear Prediction Coding Cepstra (PLPCC) is understood in the art ~~ 10. “and will not ‘be’ described in further detail Reference is made, for example, to
Hirschberg, J., Liscombe, J., and Venditt, J. Experiments in Emotional Speech. IEEE ~~ = } Workshop on Spontaneous Speech Recognition. 2003, the contents of which are herein . Co : incorporated by reference. = iE oo oo : co "A MPEG-7 descriptor is understood in the art and will not be described in further © detail. Reference is made, for example, to B.S. Manjunath (Editor), Philippe Salembier - (Editor), and Thomas Sikora (Editor): Introduction to MPEG-7: Multimedia Content ~~ Description Interface. Wiley & Sons, April 2002 - ISBN 0-471-48678-7, the contents of - which are herein incorporated by reference.
Co SE Co oo Machine Learning features are also known as data driven features, in which the oo system learns the patterns of events by being fed with data samples of the events. In “other words, sample sounds of each event are collected. The sample sounds preferably :
Co cover a sufficiently wide variation of the conditions in the lift for a particular event. With 25° sufficient good data samples, an engine (also called a classifier) is produced to recognize the events. The engine can be created by a process where signals are fed to the engine and information about each signal is provided: This process is called the training of a classifier.’ Potential spurious signals that resemble a particular event signal can also be trained into the engine as negative samples. An engine that is created with ~~ sufficient trainings of sample data can classify a signal into one of the event categories it ~~ ©
Cs trained. in the example of screaming detection, data from sample events involving - screaming are collected, pre-processed, and sent for training using machine learning
- nx : tools such’ as the Neural-Networks, : Bayesian Networks, or Support Vector Machine. n o Mathematical representations can then be used to discriminate different events. : oo - Querying for mid/high level information e.g. indicated at numeral 260 to the video ‘classifier 256 (slave classifier) can be issued to assist in the accuracy of this ©. Classification. . For example, mid and high level information includes the number of - people in the if and the presence of aggressive behaviour respectively. In the eventthat
Co 3 "there is only one person in the lift, the system may treat it as a false alarm and may only y
Co : i provide a waning to the passenger-in the lift. Conversely, if more than one person isin 10° the lift, a high level query to an Aggressive Motion Measurement unit 233 is preferably - ‘made. A Neural Network classifier may be used to detect aggressive behaviours based
Co on the ‘shape and motion features extracted from the video. The features can include = - "motion: energy image (MEI), motion history image. (MHI), and motion rapidity image © (MRI). MRI is understood in the art and will not be described in further detail. Reference is made, for example, to Liyuan Li and Maylor KH. Leung (2001), "Suspicious Human
Action Detection for Video Surveillance," in Proc. of 3rd Intl Conference on Advanced
Concepts for Intelligent Vision Systems (ACIVS), Germany, the contents of which are herein incorporated by reference. If the system detects the presence of more than one . person and aggressive behaviour, it may imply that a crime against a soft target is 20 . occurring. Co - Co oo A crime against a soft target can then be detected in the example embodiment through - the evidence obtained via the processes of categorization, verification of - "dominant signature and machine leaming classification, assisted / confirmed using sub- ~ evidence obtained from queries from the slave classifier(s). oo One example of the use of the vibration classifier 254, functioning as a master © classifiet, can be for the detection of urination in a lift 242 and acts of vandalism 244. At . step 214, a vibration sound is detected. At step 216, the signal goes through a number oo ) of sampling and pre-processing. tasks that include the removal of noise and signal re- + + sampling. Co BE Lo
Ch aa So = JS - Categorization 206 is performed by the: vibration “classifier 254 wherein the = vibration signal is categorized into ‘silent/non-silent, and ‘normaliabnommal. Deviations
Co from the normal situations initiate ‘an’ identification of parameters that are precursors to = : so an event. Queries (low-level), e.g. indicated at numeral 264 to the video classifier 256 (slave classifier) for foreground object detection 229 and human model detection 230, ~ are issued to assist in the accuracy of this categorization. If the audio signal is “silent” or
CL “normal”; or no object is detected, further classification may not be performed. : = SRE The master classifier 254 duieries 264 the slave classifier on the presence of low- 10° level, mid-level and" high-level information in this embodiment and examples of such = ) So queries have been described above. | s : oo TL EE
TA Verification of dominant signature 208 is also carried out at the vibration classifier - 254. The vibration signal is analyzed for the presence of the characteristic sounds of : 15 - urination (e.g. low beat and periodic) and/or vandalism (e.g. high pitch scratching sounds). : Lo
Machine Learning classification 210 using weighted features from audio and vibration signals is also performed at the vibration classifier 254. The sound is compared to sample urination sounds and/or ‘sample sounds caused by acts of vandalism. As mentioned above, Machine Learning classification advantageously facilitates a more comprehensive detection of signals compared to the verification of dominant signatures. : Querying for mid/high level information e.g. indicated at numeral 264 to the video 25° classifier 256 (slave classifier) can be issued to assist in the accuracy of this - + classification. For example, with regard to urinating in the lift, if there is more than one oo person in the lift the system may treat it as a false alarm and may only provide a : warning to the passeriger in the lift. Conversely, if only one person is in the lift and is not : facing the door, the systern may treat it asa true alarm (i.e.; someone is urinating in the lift). A further query can be made to a Water Patch Detection unit 232 to detect the "presence of urine or a liquid patch (e.g. vandalism caused by strewing liquids in the lif).
Cas . at Urinating and vandalism’ inthe lift can’ then be’ detected in the example ~ oe : embodiment : through the evidence: obtained via the processes of categorization, oo = verification of dominant signature and ‘machine leaming classification, assisted /
Ce confirmed using sub-evidence obtained from queries from the slave classifier(s).- So
Ii For events such as trash dumping (abandoning objects in the lift) 238 and ~~
Co fire/smioke in the lift 240, an example detection algorithm’ involves obtaining a video ~
Te signal from the CCTV camera at step 224. The signal is processed at step 226 to - Ln remove noise. SE nn SE Sl 0
Cl The video classifier 256 performs object segmentation 228. As trash dumping ~~ - . 238 and fire/smoke in the lift 240 donot produce audio or contact signatures; the video © 3 - : classifier 256.is designated as the master classifier. No slave classifier is designated. "Event analysis at step 239 can be carried out directly by the video classifier to detect the co occurrence of trash dumping 238 or fire/smoke in the lift 240. :
Co In the ‘example embodiment described above, implementation involved both verification of dominant signature and machine learning classification. However, it will be appreciated that alternative example embodiments may be implemented with either verification of dominant. signature or machine learning classification only, as will be © described below. oo IER | oo oo Figure 2B is a flow chart illustrating the system architecture/processing steps of - an audio and video awareness system for lift monitoring, generally designated as 25° reference numeral 200b, according to an another example embodiment of the present invention. Similar to the system 200a, the present system 200b comprises a plurality of classifiers: an audio classifier 252, a vibration classifier 254 and a video classifier 256. } hs One example of the use of the audio classifier 252, functioning as a master 30. . classifier, can be for the detection of crimes against soft targets (e.g. the elderly, children © and women) 212. At step 202, an audio sound is defected. At step 204, the audio signal : goes through a number of sampling and pre-processing tasks that include the removal of : noise and signal resampling. PE oo :
ie EEE
Categorization 206 is performed by the audio classifier 252 wherein the audio : od signal is” categorized into ‘silent/nor-silent’, ‘voice/non-voice’, and ‘normal/abnormal’.
Queries (low-level), e.g. indicated at numeral 260 to the video classifier 256 (slave 5° classifier) for foreground object detection 229, and human model detection 230 are
B issued to assist in the accuracy of this categorization. If the audio signal is “silent” or . - “normal’; or no object is detected, further classification may not be performed. : 3 : no “The master classifier 252 queries 260 the slave classifier on the presence of low- ~~ level, mid-level and high-level information in this embodiment and examples of such ©
Ca : queries Have been described above in relation to the previous embodiment. Lo : I : . : - Verification of dominant signature 208 is also carried out at the audio classifier . 252. The audio signal is analyzed for the presence of sounds characteristic of crimes : 15 against soft targets such as shouts and cries (e.g. short duration, non-repetitive- high . pitch sounds with narrow spectral). } oo Querying for mid/high level information e.g. indicated at numeral 260 to the video classifier 256 (slave classifier) can be issued to assist in the accuracy of verification. For : 20 example, mid and high level information includes the number of people in the lift and the presence of aggressive behaviour. respectively. In the event that there is only one person "in the lift; the system may treat it asa false alarm and may only provide a warning to the ~ passenger in the lift. Conversely, if more than one person is in the lift, a high level query
Co to an Aggressive Motion Measurement unit is preferably made 233. A Neural Network classifier may be used to detect aggressive behaviours based on the shape and motion features extracted from the video: The features are motion energy image (MEI), motion history image (MHI), and motion rapidity image (MRI). If the system detects the : - : : presence of more than one person and aggressive behaviour, it may imply that a crime against a soft target is occurring. - oo IS Co : 0 Cn SE oo A crime against a soft target can then be detected in the example embodiment through the evidence obtained via the processes of categorization and verification of
CT Umeelonen TELL oErhgg 0 Ta E00 Ten
X E ) dominant signature, assisted / confirmed using sub-evidence obtained from queries from . - “the slave classifier(s). ANY oo 3 oo oo one example of the use of the vibration classifier 254, functioning as a master classifier, can be for the detection of urination in a lift 242 and acts of vandalism 244. At o step 214, a vibration sound is detected. At step 216, the signal goes through a number 8 2 of sampling and pre-processing tasks that include the removal of noise and signal re-
Camping nase iE a 10: - Categorization 206 is performed by the vibration classifier 254 wherein the . © © vibration signal is categorized into ‘silentinon-silent’, and ‘normal/abnormal’. Deviations from the normal situations initiate an identification of parameters that are precursors to + - “an event, ‘Queries (low-level), e.g. indicated at numeral 264 to the video classifier 256 (slave classifier) for foreground object detection 229 and human model detection 230, are issued to assist in the accuracy of this categorization. If the audio signal is “silent” or . “normal”; or no object is detected, further classification may not be performed. oo The master classifier 254 queries 264 the slave classifier on the presence of low- co level, ‘mid-level and high-level information in this embodiment and examples. of such queries have been described above in relation to the previous embodiment. Fo oo Verification of dominant signature 208 is also carried out at the vibration classifier 254. The vibration signal is analyzed for the presence of the characteristic sounds of
Co urination (e.g. low beat and periodic) and/or vandalism (e.g. high pitch scratching sounds). . . Sa I © Querying for mid/high level information e.g. indicated at numeral 264 to the video
Co classifier 256 (slave classifier) can be issued to assist in the accuracy of classification. | oo © For example, with regard to urinating in the lit, if there is more than one person in the lift, 30. the system may treat it as a false alarm and may only provide a warming to the passenger in the lift. Conversely, if only one person is in the lit, a high level query is : ‘preferably issued to a Human Posture Estimation unit 231 to determine if a passenger is facing the door. If the passenger is not facing the door and there is only one person in
Cas - Co the lft, the system may treat it as a true alarm (i.e.: someone is urinating in the lift). A : : a further query can be made to a Water Patch, Detection unit 232 to detect the presence of "urine or a liquid patch (e.g. vandalism caused by strewing liquids in the lift). ~~ - 5. - Urinating and vandalism’ in the lift can then be detected in the example i embodiment ‘through the evidence obtained via ‘the processes of categorization and : 3 verification of dominant signature, assisted / confirmed using sub-evidence obtained 8 = : from queries from the slave classifiers): - Co ; SE A Co 10° ~ - For events ‘such as trash ‘dumping (abandoning objects in the lift) 238 and "firelsmoke in the [ft 240, an example detection algorithm involves obtaining a video 3 - signal from the CCTV camera at step 224. The signal is processed at ‘step 226 to © removenoise. SE
The video classifier 256 performs object segmentation 228. As trash dumping
Co 238 and fire/smoke in the lift 240 do not produce audio or contact signatures, the video classifier 256 is designated as the master classifier. No slave classifier is designated.
So Event analysis at step 239 can be carried out directly by the video classifier to detect the ~ ocourrenice of trash dumping 238 or fire/smoke in the lit 240.
Co . - oo ) - “Figure 2C is a flow chart illustrating the system architecture and processing steps ~. of an audio and video awareness system for lift monitoring, generally designated as
Co reference numeral 200c, according to a further embodiment of the present invention.
Similar to the systems 200a/b, the present system 200c comprises a plurality of classifiers: an audio classifier 252, a vibration classifier 254 and a video classifier 256. ) Co ‘One example of the use of the audio classifier 252, functioning as a master ) | classifier, can be for the detection of crimes against soft targets (e.g. the elderly, children 30° and women) 212. At step 202, an audio sound is detected. At step 204, the audio signal
B goes through a number of sampling and pre-processing tasks that include the removal of noise and signal re-sampling. - . oo
Coa oe : . | Categorization 206 is performed by the audio classifier 252 wherein the audio ~~ signal is categorized. into ‘silent/non-silent’, ‘'voice/non-voice’, and ‘normaliabnormal’. © Queries, e.g. indicated at numeral 260 to the video classifier 256 (slave classifier) for
CL foreground object detection 229, and human model detection 230 are issued to assist in 5° the accuracy of this categorization: If the audio signal is “silent” or “normal” ”; or no : : object is detected, further classification may not be performed. ro - Co The master classifier 252 queries 260 the slave classifier on the presence of low- level, mid-level and high-level information in this embodiment and examples of such x queries have been described above in relation to the previous two embodiments. EE - : - ~ : ‘Machine Learning classification 210 using weighted features from audio and - “vibration signals is also performed at the audio classifier 252. Machine Learning
Classification advantageously facilitates a more comprehensive detection of signals compared to the verification of dominant signatures. The sound is compared to sample : sounds characteristic of crimes against soft targets to detect the presence of crimes ~~ against soft targets. . ©. Querying for mid/high level information e.g. indicated at numeral 260 to the video classifier 256 (slave classifier) can be issued to assist in the accuracy of classification.
For example, mid and high level information includes the number of people in the lift and . the presence of aggressive behaviour respectively. In the event that there is only one person in the lift, the system may. treat it as a false alarm and may only provide a : warning to the passenger in the lift. Conversely, if more than one person is in the lift, a © Neural Network classifier is used then to detect aggressive behaviours based on the oo shape and motion features extracted from the video. The features are motion energy - image (MEI), motion history image (MHI), and motion rapidity image (MRI). If the system x detects the presénce of more than one person and aggressive behaviour, it may imply oo © that a crime against a soft target is occurring. ©. g oo .]
Ca Co - A crime against a soft target can then be detected in the example embodiment through the evidence obtained via the processes of categorization and machine learning
CL 0 BE oo ) classification, assisted / confirmed using sub-evidence obtained from queries from the slave classifier(s). BE EE oo - i | ) Co One example of the use of the vibration classifier 254, functioning as a master ; classifier; can be for the detection of urination in a lift 242 and acts of vandalism 244. At i : : step 214, a vibration sound is detected. At step 216, the signal goes through a number : of sampling and pre-processing tasks that include the removal of noise and signal re- sampling. CL
Categorization 206 is performed by the vibration classifier 254 wherein the
R vibration ‘signal is categorized into ‘silent/non-silent, “and ‘normal/abnormal’. Deviations : ; from the normal situations initiate ‘an identification of parameters that are precursors to an event. Queries (low-level), e.g. indicated at numeral 264 to the video classifier 256 (slave classifier) for foreground object detection 229 and human model detection 230, are issued to assist in the accuracy of this categorization. If the audio signal is “silent” or ~~ “normal”; or no object is detected, further classification may not be performed. oo The master classifier 254 queries 264 the slave classifier on the presence of low- level, mid-level and high-level information in this embodiment and examples of such queries have been described above in relation to the previous two embodiments.
Machine Learning classification 210 using weighted features from audio and vibration signals is also performed at the vibration classifier 254. The sound is compared : to sample urination sounds and/or sample sounds caused by acts of vandalism.
So o | Querying for mid/high level information e.g. indicated at numeral 264 to the video "classifier 256 (slave classifier) can be issued to assist in the accuracy of classification:
For example, with regard to urinating in the lift, if there is more than one person in the lift, - the system may treat it as a false alarm and may only: provide a warning to the . passenger in the lift Conversely, if only one person. is in the lift, a high level query is preferably issued to a Human Posture Estimation unit 231 to determine if a passenger is oo facing the door. If the passenger is not facing the door and there is only one person in the lift, the system may treat it as a true alarm (ie: someone is urinating in the lift). A
CLI lpn er Tg Te DT “further query can be made to a Water Patch Detection unit 232 to detect the presence of “urine ora liquid patch (e.g. vandalism. caused by strewing liquids in the lift). » - Urinating and vandalism. in “the lift can then be detected in the example - embodiment through the evidence obtained via the processes of categorization and
BN machine leaming classification, assisted / confirmed using sub-evidence obtained from
Queriesfrom the slave classifier(s). © oo Co 1" For events such as trash dumping (abandoning objects in the ifty 238 and a 10: fire/smoke in the lft 240, an example detection algorithm involves obtaining a video - ‘signal from ‘the CCTV camera at step 224. The signal is processed at step 226 to fT . remove noise. Co . = :
The video classifier 256 performs object segmentation 228. As trash dumping 238 and fire/smoke in the lift 240 do not produce audio or contact signatures, the video classifier 256 is designated as the master classifier. No slave classifier is designated. Event - analysis. at step 239 can be carried out directly by the video classifier to detect the occurrence of trash dumping 238 or fire/smoke in the lift 240.
Designating master and slave classifiers in example embodiments of the present invention advantageously results in a more robust outcome since it is more resilient to the effect of noises on the sensors. ~The algorithms for the processing of video signals in the example embodiments 25° comprise: 3 Co Co ; Co .
CU al Dynamic background scene modeling - This is preferably the most fundamental
Sl algorithm for the processing of video signals and it functions to extract changes inthe background scene of the lift and segments out humans or objects in the lift. oo ‘The change is called - foreground and, for example, can be due to -objects/human(s) that come into the scene of a lift's interior. In order for the
Co system to ‘run continuously and accurately without human intervention, the © algorithm adapts to changes in the scene. For instance, when the lift door opens
CE 2 oo . and additional light enters the lift, this does not induce errors in the extraction of nn - “changes in the background scene. Co Co - en Human model detection - In a confined space of a lift, movement trajectory may
Co | hot be used to deduce whether a detected object is a human. in embodiments of oF ; the present invention, a silhouette of a foreground blob can be analyzed to check 2 LE " the presence of an. inverted omega pattern. This pattern signifies the head and
Co sewdesofamman 10° - : : «Object: basic’ shape classification - The foreground blob can be analyzed to
So - - determine if it is of one of the basic shapes such as oval, rectangular, elongated 0. (Verticallhorizontal). oo . "© « Aggressive human actions - Human motion is first extracted using computations on successive pictures from a video stream. This motion is further analyzed to oo - match motion patterns of those pertaining to fighting, battering, etc. These
So classifications are enabled by a machine learning engine (see below) trained with - oo samples of such actions. EE Co : oe
Table 1 below summarizes the preferred algorithm components from each sensor for each event, according to example embodiments. Events that involve airborne sounds, such as human voices, are preferably captured by an audio microphone. Events that involve structural sound, such as scribbling on a wall, urinating, are preferably . captured by a contact microphone. ~ . Co en Tiactk Doro RgI oN] ESTEE ET Seti GR] RES ENE atl, ST PE Ec i a a I Lo iN EEO CL CLD oe S06 ales SWRHICIIETHS so
So vibration sighalll SANNIOSIGRAl aide process Fanaa
Lo Sma Da lata as
Lo ATR NE bE A IRI SRE ba laa he
Dn a
Sr Ee me a LE ae a
SL ‘aterasnag mans GN iain re Baa ness ese ret ote 5 EC A ae ge foe a ee
Se { A NOE ed A A ye), rae GRE ESA EN <& PEER ire Sawdon Lo ill eng Re
Sr Soa maga = = cL Doctire recognitions LEN ane
Pl Sea an nn OSEIRTE000RIon. ohn
Co SHAPER HE rsa an See Li ee San co rasa Le OMT Pe A ET ee SD FUR Ree ais Poo Ban eral eR cen
Co RE eh NC ne nr eae
Cnn : mea he eas Ee oe nT oa Cahn
J foie: oa a en en Bonen ig or ra rs ee SR i Ra i a
Ce oe Ed
Ca Bedi mle ane ed sd eee as a ae
TT a La
SL y wo a it i Er a tances asd Ob Mor he b Ge i anges -
GNI ee NIE 0 Objectshape J ae Changesin
Fa Lo en = mee ' EEA Se aa rE eer a re Sa rl De SERRE GR Ee A ere Sg fed SO SER ce i Nh fe : pdm ERS aE E Gl i Le ail LS i backgroundsscene
Co ye ns [doen penaosinnal) Gp sMTCERlISGESl | IDOI ONE .. EADTEHORISIGEISH £3 Ea Vibrahiorrsinns i MEI GIO CESSIDT 5 EERE ITESSINE RE iy
Co Er I ek Dl eo DE a Se to Ea Sas IR a BPE Be valitand tooRREBRST Eel Sep ae Bk
Co © Table1 | : : ‘In the detection of events. the inventor has recognized and the embodiments of : 5 the invention exploit the following features: EE oo ~. «The nature of the events is such that the information derived from the signal of
Co . ‘one of the audio or contact microphones is generally. more informative compared
E . -to the information derived from the video signal. - : EE
Co -e In. crimes against soft targets in lifts, stronger audio signals (such as crying,
Co ‘shouting, screaming, etc) are generated compared to audio signals arising from ~~ vigorous physical actions. (such as the act of fighting back).
Ca oo - } : Ce The act of urinating; in the lift can be detected from the sound of liquid dripping oo - ‘due to the contact between the lift structure (e.g. floor/wall) and urine flow. A ) in “liquid patch can also be detected by image processing to confirm the event. In
SL other words, when a dripping sound is detected, a video signal can be 5 SS concurrently. analyzed to ‘see if a human is detected. In order to reduce the sn “number of false alarms, a greater number of concurrent verifications can be - ik i g performed, This, however, is at the cost of more computational resources. EE Co oC n oe . Some acts of vandalism (e.g. breaking, scribbling) can be more easily detected oo - SE by the sounds produced (e.g. when a sharp object is used to carve on a lift wall). co : . co Detection that is simply based on image processing is relatively more unreliable. ch + Depending on the event; one of the two microphone signals (ie. contact or = audio) will be more accurate for certain events compared to others. 15. a. SE oo oe . The confined environment of a lift allows the microphone signals to be more
CS clearly captured for: analysis; Conversely, the confined environment may not allow information such as an object's spatial position or rigorous actions to be : useful in detecting events. Thus, an Al system based on camera signal alone is oo generally ineffective. : 3 -
Bn It will be appreciated by a person skilled in the art that depending on the nature ~ of the event to be detected, different types of sensors that are more suitable than others. ~~. For example, in the above description, it is preferable that in order to detect anti social behaviours (e.g. urinating in lifts and dumping of rubbish), vandalism and crimes against soft targets, audio microphones and/or contact microphones ‘are utilized. oo So * Dynamic self-learning ’ Co oo ua 30 . Furthermore, there is additionally provided a dynamic self-leaming mechanism that allows the classifiers to automatically learn variations of an event when the system . is deployed. The system advantageously improves itself as it is exposed to a greater ~~ variety of data in actual use. -
25 CT - ) The use of multiple sensors’ in the manner described above advantageously Co - } enables the Master Classifier of a particular event to use event data to train the other i : : secondary classifiers. Data pertaining to the event detected by the system can be fed 5: - back to the Slave Classifiers that assist the Master Classifier. This incremental leaning Co : . can widen, the exposure of the system to real events that take place ata particular site - Co : Aa and makes the overall classifier smarter compared to at the initial period when tis - : - trained with mock-up data. Ce - g - Co Embodiments of the present invention utilize multiple sensors wherein each one - “of the sensors operates in a comprehensive manner (instead of point sensors such as “an acid detector which is localized to the point where the sensor is installed)
SN ‘advantageously giving rise to robustness and a relatively greater variety of events that } “can be detected. In addition, as embodiments of the present invention are software - 15 based, there is a potential for expansion to cover other situations (events) that may be : oo required by users. - Cl Co p
In addition, embodiments of the present invention advantageously enable better control and management of lifts with a more efficient and economical way for users to 20 . respond to events (e.g. via mobile devices, compared to requiring a monitoring centre).
Furthermore, embodiments of the present invention provide real-time detection of these events. Alarms can then be forwarded to relevant agencies (e.g. police), if applicable
Co and immediate action can be taken to help the victims. BE | Further advantages include cost-savings as there is a reduced need for extra - cleaning: services (e.g. due to urination and trash removal), damages from vandalism and illegitimate operation of the lift. Productivity is decreased due to the typically long and tedious investigation of an event. For instance, when an urination incident occurs, it is first leamt from a user's complaint: This is recorded through a standard administrative ; process. Investigative action may be taken some days/weeks later. The process involves
BE the localization of the video clips that are retrieved from a centralized recording centre. oo These videos are then studied frame by frame to locate the incident. Embodiments of - the present invention can provide real-time detection where a culprit can be caught
2 y I oo : quickly. Thus, investigation. is not necessary all the time. In the investigation process of events that are not detected or events that are not configured to be detected but are of - interest, users can retrieve recordings that are only to specific conditions (e.g. number of
Ll people found in the litt, their action levels, voice or. audio profile) rather than manually 3 review voluminous cctv recordings. In the urination incident for instance, users can . ; So employ embodiments of the present invention to. filter video segments where there is = 3 “only a single person in the lift, presumably where there is a higher likelihood of urination. © ) : 3 : : Figure 3is a flow chart, designated generally as reference numeral 300, oo 10 . ‘illustrating a method of event detection, according to an example embodiment of the - present invention. At step 302, respective sensor signals are obtained from a plurality of ~ ‘sensors. At step 304, the respective sensor signals are processed for the event
EE detection using respective classifier units coupled to the plurality of sensors, wherein an. © eventis detected based on the sensor signal processed in one classifier unit as a main : 15 evidence. At step 306, sub-evidence queries are issued to the other classifier units for :
Co facilitating the event detection using said one classifier unit.
The method and system of the example embodiment can be implemented on .. a computer system 400, schematically shown in Figure 4. It may be implemented as software, such as a computer program being executed within the computer system 400, and instructing the computer system 400 to conduct the method of the example . embodiment. E n | : oo ~The computer system 400 comprises a computer module 402, input modules - such as a keyboard 404 and mouse 406 and a plurality of output devices such as a display 408, and printer 410. © ol | - : The computer module 402 is connected to a computer network 412 via a’ E oo ‘suitable transceiver device 414, ‘to enable access to e.g. the Internet or other network systems such as Local Area Network (LAN) or Wide Area Network (WAN). oo The computer module 402 in the example includes a processor 418, a ~~ Random Access Memory (RAM) 420 and a Read Only: Memory (ROM) 422. The
Co I 2 IEE oo computer module 402 also includes a number of Input/Output (VO) interfaces, for oo ~~ example I/O interface 424 to the dispiay 408, and I/O interface 426 to the keyboard
Cea ov 5° - ) . The components of the computer module 402 typically communicate via an on : interconnected bus 428 and in a manner known to the person skilled in the relevant ) . Cant 3 Th Cen . = as : CT i Co = i 2 ~The application program is- typically supplied to the user of the computer ‘system 400 encoded on a data storage medium ‘such as a CD-ROM or flash = ‘memory carrier and read utilising a corresponding data storage medium drive ofa
Co - data storage device 430. The application program is read and controlled inits execution by’ the processor 418. Intermediate storage of program data maybe ~~ accomplished using RAM 420. ~~ So ST | oo i. : It will be ‘appreciated by a person skilled in the art that numerous variations . and/or modifications may be made to the present invention as shown in the specific embodiments without departing from the spirit. or scope of the ‘invention as broadly - described. The present embodiments are, therefore, to be considered in all respects to be illustrative and not restrictive. ; :

Claims (15)

  1. RE CTs 28 CLAIMS Co : 1. ~ Asystem for event detection, comprising:
  2. -_... aplurality of sensors; and: a u A - : respective classifier units coupled to the sensors for processing respective Co - Fo . sensor signals from the sensors for the event detection; Co Lan : cl - + wherein at last one of the classifier units is adapted for detectinganevent ~~. - B : based on the sensor signal processed in said at least one classifier unit as a main So Co evidence and for issuing sub-evidence queries to the other classifier units for facilitating ~~ - the event detection. ~~ Id BE "2. The system as claimed in claim 1, wherein said at least one classifier unit. oo CE ‘comprises a categorisation unit for categorising the sensor signal into event categories. ~~
  3. 3. The system as claimed in claim 2, wherein the categorisation unit is adapted to : use one or more parameters from the sub-evidence queries to the other classifier units
    *. in categorising the sensor signal into the event categories. -
  4. 4. The system as claimed in any one of the preceding claims, wherein said at least one classifier unit comprises a dominant signature verification unit for verifying a dominant signature in the sensor signal for the event detection.
  5. 5 The system as claimed in claim 4, wherein the dominant signature verification units adapted to use one or more parameters from the sub-evidence queries to the : = other classifier units in verifying the dominant signature in the sensor signal.
  6. 6 T he system as claimed in any one of the preceding claims, wherein said at least ~~ one classifier unit comprises a machine learning unit for classification processing for the ' eventdetection. - BES w = oo .
  7. 7. The system as claimed in claim 6, wherein the machine learning unit is adapted : : ~ touse one or more parameters from the sub-evidence queries to the other classifier . units in the classification processing; Ln -
  8. cl EA 8.
  9. Lo A method of event detection comprising the steps of: i IEE SE Cel . obtaining respective sensor signals from a plurality of sensors; and oo i B JT processing the respective sensor signals for the event detection using respective: =~ 5" classifier.units coupled to the plurality of sensors, wherein an event is detected basedon ~~ the sensor signal processed in one classifier unit as a main evidence; and ~~ EE 1 issiiing sub-evidence queries to the other classifier units for facilitating the event * = ; detection using said one classifier unit.
  10. SL : 107 9 3 The method as claimed in"claim 8, comprising categarising the sensor signal into EN event categories using a categorisation unit of said one classifier unit. oo a : wT 10. ~ The method as claimed in claim 9, comprising using one or more parameters oo "from the sub-evidence queries to the other classifier units in categorising the sensor ~~ signalinto the event categories. ~~ CT :
  11. 11 The method as claimed in any one claims 8 to 10, comprising verifying a dominant signature in the sensor signal for the event detection using a dominant signature verification unit of said one classifier unit comprises.
  12. 12. The method as claimed in claim 11, comprising using one or more parameters from the sub-evidence queries to the other classifier units in verifying the dominant signature in the sensor signal. + ) : :
  13. 13. The method as claimed in‘any one any one claims 8 to 12, comprising eo performing classification processing for the event detection using a machine learning : "unit of $aid one classifier unit. ~~ 3
  14. 14. The method as claimed in claim 13, comprising using one or more parameters : fromthe sub-evidence queries to the other classifier units in the classification : processing. | oe C
    ST se ©
  15. 15. Adata storage medium having stored thereon computer program code means for
    ~~. instructing-a computer system to execute a method of event detection, as claimed.in any . _-oneofclaims8to14. Co Co CT eT Le
SG2012013140A 2009-08-24 2010-08-24 Method and system for event detection SG178563A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
SG2012013140A SG178563A1 (en) 2009-08-24 2010-08-24 Method and system for event detection

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
SG200905648 2009-08-24
PCT/SG2010/000311 WO2011025460A1 (en) 2009-08-24 2010-08-24 Method and system for event detection
SG2012013140A SG178563A1 (en) 2009-08-24 2010-08-24 Method and system for event detection

Publications (1)

Publication Number Publication Date
SG178563A1 true SG178563A1 (en) 2012-03-29

Family

ID=43628266

Family Applications (1)

Application Number Title Priority Date Filing Date
SG2012013140A SG178563A1 (en) 2009-08-24 2010-08-24 Method and system for event detection

Country Status (2)

Country Link
SG (1) SG178563A1 (en)
WO (1) WO2011025460A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108053111A (en) * 2017-12-11 2018-05-18 日立楼宇技术(广州)有限公司 Elevator shunt method and system

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9397904B2 (en) 2013-12-30 2016-07-19 International Business Machines Corporation System for identifying, monitoring and ranking incidents from social media
US9965685B2 (en) 2015-06-12 2018-05-08 Google Llc Method and system for detecting an audio event for smart home devices
EP3493171A1 (en) * 2017-12-04 2019-06-05 Siemens Mobility GmbH Detection of aggressive behaviour in public transportation
WO2019110215A1 (en) 2017-12-04 2019-06-13 Siemens Mobility GmbH Automated detection of an emergency situation of one or more persons
US10354169B1 (en) 2017-12-22 2019-07-16 Motorola Solutions, Inc. Method, device, and system for adaptive training of machine learning models via detected in-field contextual sensor events and associated located and retrieved digital audio and/or video imaging
US11417128B2 (en) * 2017-12-22 2022-08-16 Motorola Solutions, Inc. Method, device, and system for adaptive training of machine learning models via detected in-field contextual incident timeline entry and associated located and retrieved digital audio and/or video imaging
US12019697B2 (en) 2018-02-16 2024-06-25 Walmart Apollo, Llc Systems and methods for identifying incidents using social media
CN111731960B (en) * 2020-06-22 2022-02-15 浙江新再灵科技股份有限公司 Elevator door opening and closing state detection method
CN116071885B (en) * 2023-04-04 2023-08-01 四川三思德科技有限公司 Anti-interference smoke induction alarm method and system for elevator

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2250156B (en) * 1990-10-06 1994-04-06 Chubb Electronics Ltd Video surveillance system
US6150927A (en) * 1998-03-30 2000-11-21 Nextbus Information Systems, Llc Anti-vandalism detector and alarm system
US20080316312A1 (en) * 2007-06-21 2008-12-25 Francisco Castillo System for capturing video of an accident upon detecting a potential impact event

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108053111A (en) * 2017-12-11 2018-05-18 日立楼宇技术(广州)有限公司 Elevator shunt method and system
CN108053111B (en) * 2017-12-11 2021-01-26 日立楼宇技术(广州)有限公司 Elevator shunting method and system

Also Published As

Publication number Publication date
WO2011025460A1 (en) 2011-03-03

Similar Documents

Publication Publication Date Title
SG178563A1 (en) Method and system for event detection
Crocco et al. Audio surveillance: A systematic review
US9761248B2 (en) Action analysis device, action analysis method, and action analysis program
Ntalampiras et al. On acoustic surveillance of hazardous situations
US9451214B2 (en) Indoor surveillance system and indoor surveillance method
Ntalampiras et al. An adaptive framework for acoustic monitoring of potential hazards
Andersson et al. Fusion of acoustic and optical sensor data for automatic fight detection in urban environments
US10109299B2 (en) Sound processing apparatus, sound processing method, and storage medium
Droghini et al. A Combined One‐Class SVM and Template‐Matching Approach for User‐Aided Human Fall Detection by Means of Floor Acoustic Features
CN114666546B (en) Monitoring method and device for communication iron tower and communication iron tower
KR101736466B1 (en) Apparatus and Method for context recognition based on acoustic information
Park et al. Sound learning–based event detection for acoustic surveillance sensors
Radhakrishnan et al. Systematic acquisition of audio classes for elevator surveillance
KR102590275B1 (en) Deep learning-based abnormal behavior detection system using de-identified data
Dadula et al. Neural network classification for detecting abnormal events in a public transport vehicle
KR20160097999A (en) Sound Detection Method Recognizing Hazard Situation
CN113781702B (en) Cash box management method and system based on internet of things
US20220148616A1 (en) System and method for controlling emergency bell based on sound
Ntalampiras Audio surveillance
EP4367653A1 (en) Threat assessment system
CN114724584A (en) Abnormal sound identification model construction method, abnormal sound detection method and system
Tripathi et al. Acoustic sensor based activity recognition using ensemble of one-class classifiers
US11379288B2 (en) Apparatus and method for event classification based on barometric pressure sensor data
He et al. Deep learning approach for audio signal classification and its application in fiber optic sensor security system
Roshan et al. Using mel‐frequency audio features from footstep sound and spatial segmentation techniques to improve frame‐based moving object detection