EP2115698A1 - Surveillance method and system using optimized object based rule checking - Google Patents

Surveillance method and system using optimized object based rule checking

Info

Publication number
EP2115698A1
EP2115698A1 EP07709185A EP07709185A EP2115698A1 EP 2115698 A1 EP2115698 A1 EP 2115698A1 EP 07709185 A EP07709185 A EP 07709185A EP 07709185 A EP07709185 A EP 07709185A EP 2115698 A1 EP2115698 A1 EP 2115698A1
Authority
EP
European Patent Office
Prior art keywords
data
sensor
analysis
rule set
object list
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP07709185A
Other languages
German (de)
French (fr)
Inventor
Mark Bloemendaal
Jelle Foks
Johannes Steensma
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ultrawaves design holding BV
Original Assignee
Ultrawaves design holding BV
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ultrawaves design holding BV filed Critical Ultrawaves design holding BV
Publication of EP2115698A1 publication Critical patent/EP2115698A1/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19602Image analysis to detect motion of the intruder, e.g. by frame subtraction
    • G08B13/19613Recognition of a predetermined image pattern or behaviour pattern indicating theft or intrusion
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19639Details of the system layout
    • G08B13/19652Systems using zones in a single scene defined for different treatment, e.g. outer zone gives pre-alarm, inner zone gives alarm

Definitions

  • the present invention relates to a surveillance method for monitoring a location, comprising acquiring sensor data from at least one sensor and processing sensor data from the at least one sensor. Furthermore, the present invention relates to a surveillance system.
  • US2003/0163289 which describes an object monitoring system comprising multiple camera's and associated processing units.
  • the processing units process the video data originating from the associated camera and data from further sensors, and generates trigger signals relating to a predetermined object under surveillance.
  • a master processor is present comprising agents which analyze the trigger signals for a specific object and generate an event signal.
  • the event signals are monitored by an event system, which determines whether or not an alarm condition exists based on the event signals.
  • the system is particularly suited to monitor static objects, such as paintings and artworks in a museum, and e.g. detect the sudden disappearance (theft) thereof.
  • a surveillance method according to the preamble defined above is provided, in which the sensor data is processed in order to obtain an extracted object list, at least one virtual object is provided, and at least one rule set is applied, the at least one rule set defining possible responses depending on the extracted object list and the at least one virtual object.
  • processing sensor data comprises indexing (or annotating) objects in the extracted object list to obtain an indexed extracted object list with a predefined set of index parameters.
  • the at least one rule set comprises first type of rules relating to the predefined set of index parameters, and second type of rules.
  • Applying the at least one rule set comprises reducing the indexed extracted object list in size by applying the first type of rules using the predefined set of index parameters, and subsequently applying the second type of rules on the reduced extracted object list.
  • the virtual object is e.g. a virtual fence referenced to the sensor data characteristic (e.g. a box or line in video footage), but may also be of a different nature, e.g. the sound of a breaking glass window.
  • the applying of rules may result in a response, e.g. generating a warning.
  • the extracted object list comprises all objects in a sensor data stream, e.g. all objects extractable from a video data stream. This as opposed to prior art systems, where only objects in a predefined region-of- interest are extracted (thus loosing information) or other systems, where only objects which generate a predefined event are extracted and further processed (e.g. tracked).
  • the objects in the extracted object list are also indexed or annotated, preferably during or directly after object extraction, which allows a more efficient retrieval of object from the extracted object list.
  • the method in a further embodiment, comprises analyzing the at least one rule set for determining the first type of rules and the second type of rules. This allows an operator to define a rule set in a high level language, which is easily understandable for a human operator, and to divide the rules in first and second type rules, which is beneficial for an efficient application of the method on the indexed extracted object list.
  • the method comprises storing results of applying the at least one rule set for application of a further rule set. Also, intermediate results may be stored for later re-use in another application of a further rule set.
  • the extracted object list is updated depending on the update rate of the at least one sensor.
  • This allows to dynamic application of the rule set, in which instant action can be taken when desired or needed.
  • the indexing of objects may be applied at the update rate or a lower frequency. This may vary from frame by frame indexing of objects, up to a limited time frame (e.g. from occurrence of an object up to disappearance of the same object), but may also apply to the entire set of sensor data (e.g. indexing after all objects are extracted and classified).
  • the extracted object list may be stored in a further embodiment, and the at least one rule set may then be applied later in time.
  • This allows to define a rule set depending on what is actually searched, which is advantageously for research and police work, e.g. when re-assessing a recorded situation.
  • This embodiment also allows to adapt a rule set and immediate rerun the analysis to check for improved response.
  • the method in a further embodiment comprises re-ordering the indexed extracted object list. This may e.g. be applied to optimize the stored indexed extracted object list for quick and efficient storage access.
  • the at least one rule set comprises multiple, independent rule sets. This allows to use the surveillance method in a multi-role fashion, in parallel operation (i.e. in real-time if needed).
  • Indexing objects in the extracted object list comprises in a further embodiment determining for each extracted object in the extracted object list associated object attributes, such as classification, color, texture, shape, position, velocity.
  • object attributes such as classification, color, texture, shape, position, velocity.
  • the object attributes may be different, e.g. in the case of audio sensors, the attributed may include frequency, frequency content, amplitude, etc.
  • Obtaining the extracted object list may comprise consecutive operations of data enhancement (e.g. image enhancement), object finding, object analysis and object tracking. As a result, an extracted object list is obtained, which may be used further in the present method.
  • data enhancement e.g. image enhancement
  • object finding e.g. object finding
  • object analysis e.g. object analysis
  • object tracking e.g. object tracking
  • the sensor data comprises video data
  • the data enhancement comprises one or more of the following data operations: noise reduction; image stabilization; contrast enhancement.
  • Object finding in a further embodiment of the present method comprises one or more of the group of data operations comprising: edge analysis, texture analysis; motion analysis; background compensation.
  • Object analysis comprises one or more of the group of data operations comprising: colour analysis, texture analysis; form analysis; object correlation.
  • Object correlation may include classification of an object (human, vehicle, aircraft,...) with a percentage score representing the likelihood that the object is correctly classified.
  • Object tracking may comprise a combination of identity analysis and trajectory analysis.
  • the present invention relates to a surveillance system comprising at least one sensor and a processing system connected to the at least one sensor, in which the processing system is arranged to execute the surveillance method according to any one of the present method embodiments.
  • the processing system in a further embodiment, comprises a local processing system located in the vicinity of the at least one sensor, and a central processing system, located remotely from the at least one sensor, and in which the local processing system is arranged to send the extracted object list (with annotations) to the central processing system.
  • the local processing system further comprises a storage device for storing raw sensor data or preprocessed sensor data.
  • a lossless video coding technique is used (or a high quality compression technique, e.g.
  • the present surveillance system may in an embodiment further comprise at least one operator console arranged for controlling the surveillance system.
  • the at least one operator console comprises a representation device, which is arranged to represent simultaneously the sensor data, objects from the extracted object list and at least one virtual object in overlay. This overlay may be used in live monitoring using the present surveillance system, but also in a post-processing mode of operation, e.g. when fine-tuning the rule sets.
  • FIG. 1 shows a schematic view of a surveillance system according to an embodiment of the present invention
  • Fig. 2 shows a schematic view of a surveillance system according to a further embodiment of the present invention
  • FIG. 3 shows a schematic view of the processing flows according to an embodiment of the present surveillance method
  • Fig. 4 shows a schematic view in more detail of a part of the flow diagram of
  • Fig. 5 shows a schematic view in more detail of a further part of the flow diagram of Fig. 3;
  • Fig. 6 shows a schematic view in more detail of a further part of the flow diagram of Fig. 3;
  • Fig. 7 shows a schematic view in more detail of a further part of the flow diagram of Fig. 3;
  • Fig. 7a shows a flow chart of an embodiment in which the rule set is applied in multiple steps
  • Fig. 8 shows a schematic view of the processing steps in the live rule checking embodiment of the present method
  • Fig. 9 shows a schematic view of the processing steps in the post-processing rule checking embodiment of the present method
  • Fig. 10 shows a view of a first application of the present method to detect intruders
  • Fig. 11 shows a view of a second application of the present method to monitor an aircraft platform
  • Fig. 12 shows a view of a third application of the present method relating to traffic management.
  • a surveillance method and system are provided for monitoring a location (or group of locations), in which use can be made of multi-sensor arrangements, distributed or centralized intelligence.
  • the implemented method is object oriented, allowing transfer of relevant data at real time while requiring only limited bandwidth resources.
  • the present invention may be applied in monitoring systems, guard systems, surveillance systems, sensor research systems, and other systems which allow to provide detailed information on scenery in an area to be monitored.
  • FIG. 1 A schematic diagram of a centralized embodiment of such a system is shown in Fig. 1.
  • a number of sensors 14 are provided, which are interfaced to a network 12 using dedicated interface units 13.
  • a central processing system 10 is connected to the network 12, and to one or more operator consoles 11, equipped with input devices (keyboard, mouse, etc.) and displays as known in the art.
  • the network 12 may be a dedicated or an ad-hoc network, and may be wired, wireless, or a combination of both.
  • the processing system 10 comprises the required interfacing circuitry, and one or more processors, such as CPU's, DSP's, etc, and associated devices, such as memory modules, which as such are known to the person skilled in the art.
  • a distributed embodiment of the intelligence of a surveillance system is shown.
  • the sensor 14 is connected to a local processing system 15, which interfaces to the network 12.
  • the one or more operator console(s) 11 may be directly interfaced to the network 12.
  • Multiple sensors 14 and associated local processing system 15 may be present in an actual surveillance system.
  • the operator console(s) 11 are connected to the network 12 via a central processing system (not shown, but similar to processing system 10 of the embodiment of Fig. 1).
  • the local processing system 15 comprises a signal converter 16, e.g. in the form of an analog to digital converter, which converts the analog signal(s) from the sensor 14 into a digital signal when necessary.
  • Processing of the digitized signal is performed by the processing system 17, which as in the previous embodiment, may comprise one or more processors (CPU, DSP, etc.) and ancillary devices.
  • the processor 17 is connected to a further hardware device 18, which may be arranged to perform compression of output data, and other functions, such as encryption, data shaping etc,, in order to allow data to be sent from the local processing system 15 into the network 12.
  • the processor 17 is connected to a local storage device 19, which is arranged to store local data (such as the raw sensor data and locally processed data).
  • Data from the local storage device 19 may be retrieved upon request, and sent via the network 12.
  • the sensors 14 may comprise any kind of sensor useful in surveillance applications, e.g. a video camera, a microphone, switches, etc.
  • a single sensor 14 may include more than one type of sensor, and provide e.g. both video data and audio data.
  • the full frame of video footage is used for object extraction (i.e. all sensor data), and not only a part of the footage (a region of interest), or only objects which generate certain predefined events, as in existing systems.
  • Object detection is accomplished using motion, texture, and contrast in the video data.
  • an extensive characterization of objects is obtained, such as color, dimension, shape, speed of an object, allowing more sophisticated classification (e.g. human, car, bicycle, etc.).
  • rules may be applied which implement a specific surveillance function.
  • behavior rule analysis may be performed, allowing a fast evaluation on complete lists of objects, or a simple detection of complex behavior of actual objects.
  • a multi-role/multi-camera analysis in which surveillance data may be used for different purposes using different rules.
  • the analysis rules may be changed after a first video analysis, and new results may be obtained without requiring processing of the raw video data anew.
  • Fig. 3 a functional flow diagram is shown of an embodiment of the surveillance method according to the present invention.
  • the video signal from the sensor 14 is converted in a digital signal in block 20.
  • the digitized video is further processed in two parallel streams.
  • the left stream implements the necessary processing for live video review and recording of the video data.
  • the digitized video data is compressed in compression block 21, and then stored in a video data store 22.
  • a lossless compression method is used, or a high quality compression technique such as MPEG4 coding, as this allows to retrieve all stored data in its original form (or at sufficient quality) at a later moment in time.
  • the video data store may be part of the local storage device 19 as shown in the Fig. 2 embodiment, and may include time stamping data (or any other kind of referencing/indexing data). Stored video data may be retrieved at any time, e.g. under the control of the operator console 11, to be able to retrieve the actual imagery of a surveillance site.
  • the right stream in the flow diagram of Fig. 3 shows the functional blocks used to obtain object data from the surveillance video data.
  • the video is enhanced using image enhancement techniques in functional block 31.
  • object are extracted or found in functional block 32.
  • the found objects may then be analyzed in functional block 33.
  • functional block 34 objects are tracked in the subsequent images of a video sequence.
  • the object data output from the object tracking functional block 34 may be submitted to rules in a live manner in functional block 40.
  • the object and associated data (characteristics, annotations), i.e. the extracted object list, are also stored in an object data storage 45 (e.g. the local storage device 19 as shown in Fig. 2, or a central storage device, e.g. part of the processing system 10 of Fig. 1).
  • data may be retrieved (e.g. using structured queries) by functional block 46, in which the recorded objects are submitted to rule checking.
  • the rule set uses predefined virtual objects, e.g. virtual fences/perimeters/lines in a video scenery, and the rules may use the mutual relationship of the virtual objects and the detected objects to provide predefined responses.
  • the responses may include, but are not limited to providing warnings, activation of other devices (e.g. other sensors 14 in vicinity), or control of the sensors 14 in use (e.g. controlling pan-tilt-zoom of a camera).
  • Fig. 4 it is shown that the video signal is first converted into the digital domain in analog to digital conversion functional block 20.
  • Analog to digital conversion of video signals (and signals from other types of sensors) is well known in the art.
  • Various methods implemented in hardware, software or a combination of both may be used. In an exemplary embodiment, this results in a digitized video data stream with 25 frames/sec, corresponding to about 10 Mpixel/sec (or a 160 Mbit/s data rate).
  • this digitized video data is compressed in compression functional block 21, e.g.
  • MPEG4 compression block 211
  • This compressed data stream may be used for live viewing of the video footage, but also for recording (locally or at a central location).
  • the image enhancement functional block 31 is shown in more detail on the right side of Fig. 4.
  • the raw video data is subjected to a noise reduction in functional block 311, and then to a digital image stabilization functional block 312.
  • the video data is subjected to a contract enhancement functional block 313. All the mentioned functions are known as such to the person skilled in the art, and again, the functional blocks 311-313 may be implemented using hardware and/or software implementations. It is noted that the video data is still at 10 Mpixel/sec and
  • Fig. 5 shows the object finding functional block 32 in more detail.
  • the video data is subjected to one or more of a number of functions or algorithms, which may include, but are not limited to, an edge analysis block 321 arranged to detect edges in the video data, a texture analysis block 322 arranged to detect areas with a similar texture, a motion analysis block 323 arranged to detect motion of (blocks) of pixels in subsequent images, and background compensation block 324 arranged to take away any possible disturbing background pixels. From all these functional blocks 321-324, areas of possible objects may be determined in functional block 325. For all the detected objects, furthermore an object shape analysis block 326 may be used to determine the shape of each object.
  • an object shape analysis block 326 may be used to determine the shape of each object.
  • the result of this object finding functional block 32 is an object list, which is updated 25 times per second in the example given. For each object, positions in the picture, boundaries and velocity is available. All the mentioned functions are known image analysis techniques as such, and again, the functional blocks 321-326 may be implemented using hardware and/or software implementations. It is noted that at this stage, the data information flow is already at a much reduced rate, i.e. orders of magnitude smaller than the original video data at 160Mbit/sec.
  • the object analysis functional block 33 is shown in more detail. From the object list with (in the given example) 25 updates/sec, a large number of characteristic features of each of the objects may be derived. For this a number of functional blocks are used, which again may be implemented in hardware and/or software. The (non-limitative) characteristics relate to color (block 331), texture (block 332), and form (block 333) analysis, and a number of correlator functional blocks.
  • the human being correlator block 334 determines the chance whether an object is a human (with an output in e.g. a percentage score). Further correlation functional blocks indicated are vehicle correlator functional block 335, and further correlator functional block 336 (e.g. aircraft correlator). The output of these functional blocks is combined in object annotation functional block 337, in which the various characteristics are assigned to the associated object in an annotated object list.
  • FIG. 7 further details of the object tracking functional block 34 are shown schematically.
  • an identity analysis and a trajectory analysis are performed in functional blocks 341, and 342, respectively.
  • the output thereof is received by identified object functional block 343, which then outputs an identified (extracted) object list, which has an update rate of 25 updates/sec.
  • the objects may then be stored or logged in the object database 45 as discussed above (e.g. an SQL database), or transferred to the liver rule checking function, indicated as live intelligence application triggering in Fig. 7.
  • the annotations found in the functional blocks described with reference to Fig. 5, 6 and/or 7 may be used to index the extracted objects. This allows to divide the extracted object list from the original (video) data stream in a number of groups depending on index parameters. E.g. it is possible to select only objects which are more than likely (>75% confidence) a car, and apply further rule sets to this (smaller) set of object data. This reduces the capacity required for transferring the data to be searched, and for processing the data.
  • This embodiment is shown in the flow diagram of Fig. 7a.
  • the extracted object list as a collection of objects extracted from an entire set of sensor data (e.g. 12 hours of video data) are represented as a whole collection.
  • indexing is applied to the extracted object list, resulting in various sets of object data characterised by one or more indexing parameters, as shown by reference number 102.
  • the indexing is used to reduce the set of (annotated) objects to a smaller group of objects 104, by first applying logical rules from the set of rules (e.g. objects with a color red). Only this smaller object data set 104 is then used to apply the more resource intensive graphical rules (e.g. find objects crossing a virtual perimeter) in block 105, eventually resulting in the desired result 106 of the search query.
  • the indexing of extracted objects is used primarily to enable acceleration of later applications of rule sets. Indexing splits the extracted object data set stored in the object data storage 45, such that groups of (partly overlapping) objects are created, which allows to read only a part of the extracted object data for further rule checking.
  • An example of such an indexing method in block 103 is annotating an object with a time stamp, e.g. when the object is first detected in the sensor data.
  • object fingerprints of extracted objects e.g. biometric data of persons
  • the indexing of objects may include, but is not limited to small/large (e.g.
  • the indexing may be used to quickly discard objects that e.g. have not been within a certain (video image) distance from another, stationary object.
  • the indexing may be applied by determining a value for a predetermined indexing parameter, e.g. the likelihood of an object being a human, a car, a truck, etc.
  • Indexing may be implemented at various stages of the data processing described in the embodiments above. In some cases, indexing may already be performed on the digital sensor data directly, e.g. when analyzing frame pixel data of a video sensor. For other types of analysis, e.g. multiple frames of a video sequence are necessary, e.g. for determining a direction of motion of an object. Also, indexing may be executed when all video footage of a predefined period has been processed, e.g. to optimize further the access to object data storage 45 using all available indexes.
  • the method as described above may be implemented for a single camera, but also for a large number of camera's and sensors 14.
  • the rule checking output may include more complex camera control operations, such as pan-tilt-zoom operations of a camera, or handover to another camera.
  • the functions described above may in this case be implemented locally in the camera 14 (see exemplary embodiment of Fig. 20), such that each video stream is processed locally, and only the object data has to be transferred over the network 12.
  • Fig. 8 a more detailed schematic is shown of the rule checking functional block 40 of Fig. 3. Extracted object lists of all camera's 14 in the surveillance system are input (real-time) to the rule checking functional block 40.
  • this functional block 40 one or more rule set functional blocks 401-403 may be present, which each provide their associated response.
  • a structure as shown schematically in Fig. 9 may be used.
  • the extracted object lists of each camera (A, B, C) are retrieved from the object database 45, and one or more rule sets may be applied to one or all of the object lists in functional blocks 461-463. Again, each rule set provides its own response.
  • the rule sets may be changed instantly (due to changing circumstances, or as a result of one of the rule sets), and the resulting response of the surveillance system is also virtually instantaneous.
  • the rule sets may be fine-tuned, and after each amendment, the same extracted object list data may be used again to see whether the fine-tuning provides a better result.
  • a rule set is advantageously divided in two types of rules, i.e. a first type of rules (logical rules) and a second type of rules ((graphical) analysis rules).
  • the logical rules apply rules related to one or more of the indexing parameters (or annotations), and allow to make a selection of a group of relevant objects from the extracted object list.
  • the analysis rules require more processing intensive functions, e.g. involving graphical processing such as determining objects in a period of video footage which cross a virtual perimeter.
  • the operator console 11 may be arranged to implement this re-arranging of the set of rules in a query, allowing the operator to enter a query in a higher level language (understandable by humans), which is then optimized according to the present invention embodiments.
  • results of a query input by an operator may be stored locally, for further fine-tuning of a search query, or for later re-use in other queries. This furthermore may enhance the execution speed of a search query.
  • a search query When executing a search involving a large amount of (video) data, usually a number of queries are performed in sequence. E.g. first a search query may be executed to find all objects crossing a line X, resulting in a first result. This result may then be added to the respective objects as a new annotation (or index).
  • these indices may advantageously be used to further improve the search query.
  • FIG. 10 shows a camera frame (indicated by dashed line) with virtual fences and virtual lines in an image from a video camera.
  • a building is located at the actual surveillance site.
  • the camera is viewing along a road (within zone A), which is bordered by a roadside (within zone B).
  • a trench is located, the middle of which is indicated by the virtual fence line C.
  • a further virtual fence line E is located.
  • the rules applied in the live rule checking functional block 40 or in recorded object rule checking functional block 46, and possible responses, may look like: • Object X in public Zone A - No suspect situation: public area
  • Zone D Intruder alert • Object X disappears in Zone D
  • a further example is shown for a surveillance system in an airport environment.
  • An aircraft parking zone on an airfield is indicated by the virtual fence Zone P inside a camera frame (indicated by dashed line).
  • the following responses are executed: Wait until the object X stops; Identify as aircraft (according to shape and size of object X; After n minutes of standstill: apply virtual object fence zones A en B (indicated by Zone IA, 2A, 3A, and IB, 2B, 3B in Fig. 11 for three different objects); and After m minutes activate "Aircraft Security Rules" for Aircraft #.
  • Aircraft Security Rules for each Aircraft # (# being 1, 2, or 3 in Fig. 11) on the aircraft parking zone may have the following form:
  • a view is shown in Fig. 12 with virtual fences for a traffic measurement and safety application.
  • a roadside is located in the actual location, along which a number of parking spaces are provided, which scenery is viewed in a camera frame indicated by a dashed line.
  • a virtual fence Zone B is raised on the roadside, and a virtual fence Zone D is raised around the parking spaces.
  • a first line A is drawn across the road in the distance, and a second line C is drawn across the road nearer to the camera position.
  • a number of rules with different purpose may be set.
  • a first rule set allows to assist in traffic management:
  • a second rule set may be applied related to safety:
  • the sensors are chosen as providing video data.
  • other sensors such as audio sensors (microphone), vibration sensors, which also are able to provide data which can be processed to obtain extracted object data.
  • audio sensors microphone
  • vibration sensors which also are able to provide data which can be processed to obtain extracted object data.
  • a virtual object may e.g. be 'Sound of braking glass' and the rule may be: Object is 'Sound of breaking glass': then activate nearest camera to instantly view the scene.

Abstract

Surveillance method and system for monitoring a location. One ore more sensors (14), e.g. camera's, are used to acquire sensor data from the location. The sensor data is processed in order to obtain an extracted object list (100), including object attributes. A number of virtual objects, such as a virtual fence, are defined, and a rule set is applied. The rule set defines possible responses depending on the list of extracted objects and the virtual objects. Rule sets may be adapted, and amended responses may be assessed immediately. A part of the rule set is applied (103) to reduce the amount of data to which the rest of the rule set is applied (105), for which more complex processing is necessary.

Description

Surveillance method and system using optimized object based rule checking
Field of the invention
The present invention relates to a surveillance method for monitoring a location, comprising acquiring sensor data from at least one sensor and processing sensor data from the at least one sensor. Furthermore, the present invention relates to a surveillance system.
Prior art Such a method and system are e.g. known from American patent application
US2003/0163289, which describes an object monitoring system comprising multiple camera's and associated processing units. The processing units process the video data originating from the associated camera and data from further sensors, and generates trigger signals relating to a predetermined object under surveillance. A master processor is present comprising agents which analyze the trigger signals for a specific object and generate an event signal. The event signals are monitored by an event system, which determines whether or not an alarm condition exists based on the event signals. The system is particularly suited to monitor static objects, such as paintings and artworks in a museum, and e.g. detect the sudden disappearance (theft) thereof.
Summary of the invention
The present invention seeks to provide a surveillance method and system with improved performance, especially alleviating or eliminating the disadvantages of the prior art methods and systems as mentioned above. According to the present invention, a surveillance method according to the preamble defined above is provided, in which the sensor data is processed in order to obtain an extracted object list, at least one virtual object is provided, and at least one rule set is applied, the at least one rule set defining possible responses depending on the extracted object list and the at least one virtual object. Furthermore, processing sensor data comprises indexing (or annotating) objects in the extracted object list to obtain an indexed extracted object list with a predefined set of index parameters. The at least one rule set comprises first type of rules relating to the predefined set of index parameters, and second type of rules. Applying the at least one rule set comprises reducing the indexed extracted object list in size by applying the first type of rules using the predefined set of index parameters, and subsequently applying the second type of rules on the reduced extracted object list.
The virtual object is e.g. a virtual fence referenced to the sensor data characteristic (e.g. a box or line in video footage), but may also be of a different nature, e.g. the sound of a breaking glass window. The applying of rules may result in a response, e.g. generating a warning. It is noted that in the present invention, the extracted object list comprises all objects in a sensor data stream, e.g. all objects extractable from a video data stream. This as opposed to prior art systems, where only objects in a predefined region-of- interest are extracted (thus loosing information) or other systems, where only objects which generate a predefined event are extracted and further processed (e.g. tracked). Furthermore, according to the present method, the objects in the extracted object list are also indexed or annotated, preferably during or directly after object extraction, which allows a more efficient retrieval of object from the extracted object list. By first applying rules of the first type, it is possible to reduce the amount of data needed for applying the rules of the second type, which further enhances the efficiency of the surveillance method.
The method, in a further embodiment, comprises analyzing the at least one rule set for determining the first type of rules and the second type of rules. This allows an operator to define a rule set in a high level language, which is easily understandable for a human operator, and to divide the rules in first and second type rules, which is beneficial for an efficient application of the method on the indexed extracted object list.
In a further embodiment, the method comprises storing results of applying the at least one rule set for application of a further rule set. Also, intermediate results may be stored for later re-use in another application of a further rule set. By storing
(intermediate) results it is possible to make more efficient use of the same sensor data for later queries using further rule sets.
In a further embodiment, the extracted object list is updated depending on the update rate of the at least one sensor. This allows to dynamic application of the rule set, in which instant action can be taken when desired or needed. Furthermore, the indexing of objects may be applied at the update rate or a lower frequency. This may vary from frame by frame indexing of objects, up to a limited time frame (e.g. from occurrence of an object up to disappearance of the same object), but may also apply to the entire set of sensor data (e.g. indexing after all objects are extracted and classified).
The extracted object list may be stored in a further embodiment, and the at least one rule set may then be applied later in time. This allows to define a rule set depending on what is actually searched, which is advantageously for research and police work, e.g. when re-assessing a recorded situation. This embodiment also allows to adapt a rule set and immediate rerun the analysis to check for improved response.
To further enhance this capability, the method in a further embodiment comprises re-ordering the indexed extracted object list. This may e.g. be applied to optimize the stored indexed extracted object list for quick and efficient storage access.
In a further embodiment, the at least one rule set comprises multiple, independent rule sets. This allows to use the surveillance method in a multi-role fashion, in parallel operation (i.e. in real-time if needed).
Indexing objects in the extracted object list comprises in a further embodiment determining for each extracted object in the extracted object list associated object attributes, such as classification, color, texture, shape, position, velocity. When other types of sensors are used, the object attributes may be different, e.g. in the case of audio sensors, the attributed may include frequency, frequency content, amplitude, etc.
Obtaining the extracted object list may comprise consecutive operations of data enhancement (e.g. image enhancement), object finding, object analysis and object tracking. As a result, an extracted object list is obtained, which may be used further in the present method.
In a further embodiment, the sensor data comprises video data, and the data enhancement comprises one or more of the following data operations: noise reduction; image stabilization; contrast enhancement. These are all preliminary steps, which aid in the further object extraction process of the present method.
Object finding in a further embodiment of the present method comprises one or more of the group of data operations comprising: edge analysis, texture analysis; motion analysis; background compensation. Object analysis comprises one or more of the group of data operations comprising: colour analysis, texture analysis; form analysis; object correlation. Object correlation may include classification of an object (human, vehicle, aircraft,...) with a percentage score representing the likelihood that the object is correctly classified. Object tracking may comprise a combination of identity analysis and trajectory analysis.
In a further aspect, the present invention relates to a surveillance system comprising at least one sensor and a processing system connected to the at least one sensor, in which the processing system is arranged to execute the surveillance method according to any one of the present method embodiments.
The processing system, in a further embodiment, comprises a local processing system located in the vicinity of the at least one sensor, and a central processing system, located remotely from the at least one sensor, and in which the local processing system is arranged to send the extracted object list (with annotations) to the central processing system. In this embodiment, only a low data rate transmission is needed between the local processing system and the central processing system, which allows to use many sensor in the surveillance system. Furthermore, it allows to use wireless network implementations, making the surveillance system much more flexible. In a further embodiment, the local processing system further comprises a storage device for storing raw sensor data or preprocessed sensor data. Advantageously, in the case of video sensors, a lossless video coding technique is used (or a high quality compression technique, e.g. MPEG4 coding). This allows to retrieve the original camera footage for later use. The present surveillance system may in an embodiment further comprise at least one operator console arranged for controlling the surveillance system. In a further embodiment, the at least one operator console comprises a representation device, which is arranged to represent simultaneously the sensor data, objects from the extracted object list and at least one virtual object in overlay. This overlay may be used in live monitoring using the present surveillance system, but also in a post-processing mode of operation, e.g. when fine-tuning the rule sets.
Short description of drawings
The present invention will be discussed in more detail below, using a number of exemplary embodiments, with reference to the attached drawings, in which
Fig. 1 shows a schematic view of a surveillance system according to an embodiment of the present invention; Fig. 2 shows a schematic view of a surveillance system according to a further embodiment of the present invention;
Fig. 3 shows a schematic view of the processing flows according to an embodiment of the present surveillance method; Fig. 4 shows a schematic view in more detail of a part of the flow diagram of
Fig. 3;
Fig. 5 shows a schematic view in more detail of a further part of the flow diagram of Fig. 3;
Fig. 6 shows a schematic view in more detail of a further part of the flow diagram of Fig. 3;
Fig. 7 shows a schematic view in more detail of a further part of the flow diagram of Fig. 3;
Fig. 7a shows a flow chart of an embodiment in which the rule set is applied in multiple steps; Fig. 8 shows a schematic view of the processing steps in the live rule checking embodiment of the present method;
Fig. 9 shows a schematic view of the processing steps in the post-processing rule checking embodiment of the present method;
Fig. 10 shows a view of a first application of the present method to detect intruders;
Fig. 11 shows a view of a second application of the present method to monitor an aircraft platform;
Fig. 12 shows a view of a third application of the present method relating to traffic management.
Detailed description of exemplary embodiments
According to the present invention, a surveillance method and system are provided for monitoring a location (or group of locations), in which use can be made of multi-sensor arrangements, distributed or centralized intelligence. The implemented method is object oriented, allowing transfer of relevant data at real time while requiring only limited bandwidth resources. The present invention may be applied in monitoring systems, guard systems, surveillance systems, sensor research systems, and other systems which allow to provide detailed information on scenery in an area to be monitored.
A schematic diagram of a centralized embodiment of such a system is shown in Fig. 1. A number of sensors 14 are provided, which are interfaced to a network 12 using dedicated interface units 13. Furthermore, a central processing system 10 is connected to the network 12, and to one or more operator consoles 11, equipped with input devices (keyboard, mouse, etc.) and displays as known in the art. The network 12 may be a dedicated or an ad-hoc network, and may be wired, wireless, or a combination of both. The processing system 10 comprises the required interfacing circuitry, and one or more processors, such as CPU's, DSP's, etc, and associated devices, such as memory modules, which as such are known to the person skilled in the art.
In an alternative embodiment, shown schematically in Fig. 2, a distributed embodiment of the intelligence of a surveillance system is shown. In this embodiment, the sensor 14 is connected to a local processing system 15, which interfaces to the network 12. In this case, the one or more operator console(s) 11 may be directly interfaced to the network 12. Multiple sensors 14 and associated local processing system 15 may be present in an actual surveillance system. For the skilled person it will be clear that other embodiments are possible, e.g. in which the operator console(s) 11 are connected to the network 12 via a central processing system (not shown, but similar to processing system 10 of the embodiment of Fig. 1).
The local processing system 15 comprises a signal converter 16, e.g. in the form of an analog to digital converter, which converts the analog signal(s) from the sensor 14 into a digital signal when necessary. Processing of the digitized signal is performed by the processing system 17, which as in the previous embodiment, may comprise one or more processors (CPU, DSP, etc.) and ancillary devices. The processor 17 is connected to a further hardware device 18, which may be arranged to perform compression of output data, and other functions, such as encryption, data shaping etc,, in order to allow data to be sent from the local processing system 15 into the network 12. Furthermore, the processor 17 is connected to a local storage device 19, which is arranged to store local data (such as the raw sensor data and locally processed data). Data from the local storage device 19 may be retrieved upon request, and sent via the network 12. The sensors 14 may comprise any kind of sensor useful in surveillance applications, e.g. a video camera, a microphone, switches, etc. A single sensor 14 may include more than one type of sensor, and provide e.g. both video data and audio data.
For surveillance applications, especially when used for large events covering a large geographical area, a lot of video data may be available from camera's located in the area. In known systems, all of the video data was observed by human operators, which requires a lot of time and effort. Improved systems are known, in which the video data is digitized, and the digitized video data is analyzed. However, such a system still requires a lot of human effort and time, especially when some analysis of the video data has to be repeated. Further improvements are known, e.g. from the publication US2003/0163289, in which object data is extracted from the video data, and events related to the object are detected (e.g. providing an alarm when a painting in a museum has suddenly disappeared). However, using known systems, it is still difficult and expensive to analyze a large amount of surveillance data. It would be a tremendous advantage when more detailed searches could be performed in surveillance data without the necessity of spending more (computer and human) time. The ability of repeatedly searching surveillance data without the need to process the raw data over and over again is also highly desired.
In the surveillance method according to embodiments of the present invention, the full frame of video footage is used for object extraction (i.e. all sensor data), and not only a part of the footage (a region of interest), or only objects which generate certain predefined events, as in existing systems. Object detection is accomplished using motion, texture, and contrast in the video data. Furthermore, an extensive characterization of objects is obtained, such as color, dimension, shape, speed of an object, allowing more sophisticated classification (e.g. human, car, bicycle, etc.). Using the objects and the associated characteristics thereof, rules may be applied which implement a specific surveillance function. As an example, behavior rule analysis may be performed, allowing a fast evaluation on complete lists of objects, or a simple detection of complex behavior of actual objects. Furthermore, it is possible to implement a multi-role/multi-camera analysis, in which surveillance data may be used for different purposes using different rules. The analysis rules may be changed after a first video analysis, and new results may be obtained without requiring processing of the raw video data anew. In Fig. 3 a functional flow diagram is shown of an embodiment of the surveillance method according to the present invention. The video signal from the sensor 14 is converted in a digital signal in block 20. The digitized video is further processed in two parallel streams. The left stream implements the necessary processing for live video review and recording of the video data. For this, the digitized video data is compressed in compression block 21, and then stored in a video data store 22. Advantageously, a lossless compression method is used, or a high quality compression technique such as MPEG4 coding, as this allows to retrieve all stored data in its original form (or at sufficient quality) at a later moment in time. The video data store may be part of the local storage device 19 as shown in the Fig. 2 embodiment, and may include time stamping data (or any other kind of referencing/indexing data). Stored video data may be retrieved at any time, e.g. under the control of the operator console 11, to be able to retrieve the actual imagery of a surveillance site.
The right stream in the flow diagram of Fig. 3 shows the functional blocks used to obtain object data from the surveillance video data. First, the video is enhanced using image enhancement techniques in functional block 31. Then, object are extracted or found in functional block 32. The found objects may then be analyzed in functional block 33. In functional block 34, objects are tracked in the subsequent images of a video sequence. The object data output from the object tracking functional block 34 may be submitted to rules in a live manner in functional block 40. The object and associated data (characteristics, annotations), i.e. the extracted object list, are also stored in an object data storage 45 (e.g. the local storage device 19 as shown in Fig. 2, or a central storage device, e.g. part of the processing system 10 of Fig. 1). From this object data storage 45, data may be retrieved (e.g. using structured queries) by functional block 46, in which the recorded objects are submitted to rule checking. In one embodiment, the rule set uses predefined virtual objects, e.g. virtual fences/perimeters/lines in a video scenery, and the rules may use the mutual relationship of the virtual objects and the detected objects to provide predefined responses. The responses may include, but are not limited to providing warnings, activation of other devices (e.g. other sensors 14 in vicinity), or control of the sensors 14 in use (e.g. controlling pan-tilt-zoom of a camera).
The functional blocks 21, 31-34, 40 and 46 are now explained in more detail with reference to the detailed functional block diagrams of Figs. 4-9. In Fig. 4, it is shown that the video signal is first converted into the digital domain in analog to digital conversion functional block 20. Analog to digital conversion of video signals (and signals from other types of sensors) is well known in the art. Various methods implemented in hardware, software or a combination of both may be used. In an exemplary embodiment, this results in a digitized video data stream with 25 frames/sec, corresponding to about 10 Mpixel/sec (or a 160 Mbit/s data rate). In a first stream, this digitized video data is compressed in compression functional block 21, e.g. using MPEG4 compression (block 211), resulting in a compressed digital footage of the surveillance site of 10 Mpixel/sec, but now reduced to 3 Mbit/sec. Compression may also be implemented using various combinations of hardware and/or software implementations, as known to the person skilled in the art. This compressed data stream may be used for live viewing of the video footage, but also for recording (locally or at a central location).
The image enhancement functional block 31 is shown in more detail on the right side of Fig. 4. First, the raw video data is subjected to a noise reduction in functional block 311, and then to a digital image stabilization functional block 312. Furthermore, the video data is subjected to a contract enhancement functional block 313. All the mentioned functions are known as such to the person skilled in the art, and again, the functional blocks 311-313 may be implemented using hardware and/or software implementations. It is noted that the video data is still at 10 Mpixel/sec and
160Mbit/sec in the mentioned example, and that the functional blocks are arranged to allow processing of video data at such rates.
Fig. 5 shows the object finding functional block 32 in more detail. The video data is subjected to one or more of a number of functions or algorithms, which may include, but are not limited to, an edge analysis block 321 arranged to detect edges in the video data, a texture analysis block 322 arranged to detect areas with a similar texture, a motion analysis block 323 arranged to detect motion of (blocks) of pixels in subsequent images, and background compensation block 324 arranged to take away any possible disturbing background pixels. From all these functional blocks 321-324, areas of possible objects may be determined in functional block 325. For all the detected objects, furthermore an object shape analysis block 326 may be used to determine the shape of each object. The result of this object finding functional block 32 is an object list, which is updated 25 times per second in the example given. For each object, positions in the picture, boundaries and velocity is available. All the mentioned functions are known image analysis techniques as such, and again, the functional blocks 321-326 may be implemented using hardware and/or software implementations. It is noted that at this stage, the data information flow is already at a much reduced rate, i.e. orders of magnitude smaller than the original video data at 160Mbit/sec.
In Fig. 6, the object analysis functional block 33 is shown in more detail. From the object list with (in the given example) 25 updates/sec, a large number of characteristic features of each of the objects may be derived. For this a number of functional blocks are used, which again may be implemented in hardware and/or software. The (non-limitative) characteristics relate to color (block 331), texture (block 332), and form (block 333) analysis, and a number of correlator functional blocks. The human being correlator block 334 determines the chance whether an object is a human (with an output in e.g. a percentage score). Further correlation functional blocks indicated are vehicle correlator functional block 335, and further correlator functional block 336 (e.g. aircraft correlator). The output of these functional blocks is combined in object annotation functional block 337, in which the various characteristics are assigned to the associated object in an annotated object list.
In Figure 7, further details of the object tracking functional block 34 are shown schematically. In consecutive images or fields of the video data, or more specifically at this stage, in the consecutive updates of the annotated object list, an identity analysis and a trajectory analysis are performed in functional blocks 341, and 342, respectively. The output thereof is received by identified object functional block 343, which then outputs an identified (extracted) object list, which has an update rate of 25 updates/sec. The objects may then be stored or logged in the object database 45 as discussed above (e.g. an SQL database), or transferred to the liver rule checking function, indicated as live intelligence application triggering in Fig. 7.
The annotations found in the functional blocks described with reference to Fig. 5, 6 and/or 7 may be used to index the extracted objects. This allows to divide the extracted object list from the original (video) data stream in a number of groups depending on index parameters. E.g. it is possible to select only objects which are more than likely (>75% confidence) a car, and apply further rule sets to this (smaller) set of object data. This reduces the capacity required for transferring the data to be searched, and for processing the data. This embodiment is shown in the flow diagram of Fig. 7a. In block 100, the extracted object list as a collection of objects extracted from an entire set of sensor data (e.g. 12 hours of video data) are represented as a whole collection. In block 101, indexing is applied to the extracted object list, resulting in various sets of object data characterised by one or more indexing parameters, as shown by reference number 102. In block 103, the indexing is used to reduce the set of (annotated) objects to a smaller group of objects 104, by first applying logical rules from the set of rules (e.g. objects with a color red). Only this smaller object data set 104 is then used to apply the more resource intensive graphical rules (e.g. find objects crossing a virtual perimeter) in block 105, eventually resulting in the desired result 106 of the search query.
The indexing of extracted objects is used primarily to enable acceleration of later applications of rule sets. Indexing splits the extracted object data set stored in the object data storage 45, such that groups of (partly overlapping) objects are created, which allows to read only a part of the extracted object data for further rule checking. An example of such an indexing method in block 103 is annotating an object with a time stamp, e.g. when the object is first detected in the sensor data. In some instances, it is possible to create object fingerprints of extracted objects (e.g. biometric data of persons), which can then later be used to screen a large number of extracted objects. The indexing of objects may include, but is not limited to small/large (e.g. maximum vertical size in pixels, minimum vertical size in pixels), long/short, rectangular/round, dark/light, time parameters (time of day, object duration (i.e. time of occurrence, time of disappearance)), quick/slow moving objects, parameters related to original video footage (e.g. maximum position in video frame (left/right/top/bottom). In the latter example, the indexing may be used to quickly discard objects that e.g. have not been within a certain (video image) distance from another, stationary object. Furthermore, the indexing may be applied by determining a value for a predetermined indexing parameter, e.g. the likelihood of an object being a human, a car, a truck, etc.
Indexing (or annotation) may be implemented at various stages of the data processing described in the embodiments above. In some cases, indexing may already be performed on the digital sensor data directly, e.g. when analyzing frame pixel data of a video sensor. For other types of analysis, e.g. multiple frames of a video sequence are necessary, e.g. for determining a direction of motion of an object. Also, indexing may be executed when all video footage of a predefined period has been processed, e.g. to optimize further the access to object data storage 45 using all available indexes.
The method as described above may be implemented for a single camera, but also for a large number of camera's and sensors 14. When multiple camera's are used, the rule checking output (Live Response, or Post processing response) may include more complex camera control operations, such as pan-tilt-zoom operations of a camera, or handover to another camera. The functions described above may in this case be implemented locally in the camera 14 (see exemplary embodiment of Fig. 20), such that each video stream is processed locally, and only the object data has to be transferred over the network 12.
In Fig. 8, a more detailed schematic is shown of the rule checking functional block 40 of Fig. 3. Extracted object lists of all camera's 14 in the surveillance system are input (real-time) to the rule checking functional block 40. In this functional block 40, one or more rule set functional blocks 401-403 may be present, which each provide their associated response.
For the case of off-line video surveillance, e.g. for research implementations of recorded video footage, a structure as shown schematically in Fig. 9 may be used. The extracted object lists of each camera (A, B, C) are retrieved from the object database 45, and one or more rule sets may be applied to one or all of the object lists in functional blocks 461-463. Again, each rule set provides its own response.
Both in the live embodiment and in the off-line embodiment, the rule sets may be changed instantly (due to changing circumstances, or as a result of one of the rule sets), and the resulting response of the surveillance system is also virtually instantaneous. In the case of the off-line embodiment, the rule sets may be fine-tuned, and after each amendment, the same extracted object list data may be used again to see whether the fine-tuning provides a better result.
In order to optimize the rule set application to the (stored) extracted object list, a rule set is advantageously divided in two types of rules, i.e. a first type of rules (logical rules) and a second type of rules ((graphical) analysis rules). The logical rules apply rules related to one or more of the indexing parameters (or annotations), and allow to make a selection of a group of relevant objects from the extracted object list. The analysis rules require more processing intensive functions, e.g. involving graphical processing such as determining objects in a period of video footage which cross a virtual perimeter. By re-arranging a rule set in two parts, and first applying the indexing rules to obtain a smaller amount of data to which the analysis rules are applied, a further improvement in processing speed may be obtained. It is noted that not only the processing power of the processing system 10 or processor 17 (see Fig. 1 and 2) is important in this aspect, but also the access time to retrieve the necessary data from the object data storage 45 or the storage device 19.
The operator console 11 may be arranged to implement this re-arranging of the set of rules in a query, allowing the operator to enter a query in a higher level language (understandable by humans), which is then optimized according to the present invention embodiments.
Furthermore, results of a query input by an operator may be stored locally, for further fine-tuning of a search query, or for later re-use in other queries. This furthermore may enhance the execution speed of a search query. When executing a search involving a large amount of (video) data, usually a number of queries are performed in sequence. E.g. first a search query may be executed to find all objects crossing a line X, resulting in a first result. This result may then be added to the respective objects as a new annotation (or index). When performing a further search later in time (e.g. 'find objects crossing line X, and having a speed below Y') these indices may advantageously be used to further improve the search query. A number of possible set-ups of the surveillance system and method according to the present inventions are now discussed with reference to the schematic diagrams of Figs. 10-12.
With reference to Fig. 10, a set-up and rule set is discussed for a virtual fencing system, which allows to detect an intruder. Fig. 10 shows a camera frame (indicated by dashed line) with virtual fences and virtual lines in an image from a video camera. In this example, on the right side of the picture (within zone E), a building is located at the actual surveillance site. The camera is viewing along a road (within zone A), which is bordered by a roadside (within zone B). Along the roadside, a trench is located, the middle of which is indicated by the virtual fence line C. At the border of the picture, a further virtual fence line E is located.
The rules applied in the live rule checking functional block 40 or in recorded object rule checking functional block 46, and possible responses, may look like: • Object X in public Zone A - No suspect situation: public area
- Possible registration because of "hazard assessment"
• Object X transits from Zone A to Zone B
- Possible intruder situation - Pre-alert and close inspection
• Object X transits from Zone B across Line C
- Object passes area border from outside: Intruder alert
• Object X in Zone D
- Intruder in Zone D : Intruder alert • Object X disappears in Zone D
- Intruder behind vehicles in front of building: last position known
- Intruder disappears outside camera view, heading north-west
• Object X transits from zone D over Line E
- Intruder transits from Zone D to area outside the camera view heading north-west
In Fig. 11, a further example is shown for a surveillance system in an airport environment. An aircraft parking zone on an airfield is indicated by the virtual fence Zone P inside a camera frame (indicated by dashed line). When a new object X is detected in Zone P, the following responses are executed: Wait until the object X stops; Identify as aircraft (according to shape and size of object X; After n minutes of standstill: apply virtual object fence zones A en B (indicated by Zone IA, 2A, 3A, and IB, 2B, 3B in Fig. 11 for three different objects); and After m minutes activate "Aircraft Security Rules" for Aircraft #.
The Aircraft Security Rules for each Aircraft # (# being 1, 2, or 3 in Fig. 11) on the aircraft parking zone may have the following form:
• Object X in Zone #A
- Possible Intruder situation
- Pre-signaling and close PTZ inspection
• Object X transits from Zone #A to Zone #B - Object crosses security border from outside: Aircraft intruder alert
- Track object in Zone P with a PTZ (Pan-Tilt-Zoom) camera
• Object X transits from Zone #B to Zone #A
- Object crosses security border from inside : Stowaway alert - Track object in Zone P with PTZ camera
In a further example of rules which may be applied to objects extracted from video data, a view is shown in Fig. 12 with virtual fences for a traffic measurement and safety application. A roadside is located in the actual location, along which a number of parking spaces are provided, which scenery is viewed in a camera frame indicated by a dashed line. A virtual fence Zone B is raised on the roadside, and a virtual fence Zone D is raised around the parking spaces. Furthermore, a first line A is drawn across the road in the distance, and a second line C is drawn across the road nearer to the camera position. At the same time, a number of rules with different purpose may be set. A first rule set allows to assist in traffic management:
• Object passes Line A and Line C
- Compute average speed from distance between the lines and the time interval (and register licence plate when average speed is over limit)
• Object(s) in Zone B stand still - Pile up situation
A second rule set may be applied related to safety:
• Object in Zone D
- Correlation with human shape > 50% (human object detected)
- Motion up and afterwards down or Motion down, afterwards up (behavioral pattern of a person looking to break into one of the parked cars)
- Possible car burglar (after which a PZT-camera may be used to obtain detailed imagery of the burglar)
In the above embodiments, the sensors are chosen as providing video data. However, it is also possible to use other sensors, such as audio sensors (microphone), vibration sensors, which also are able to provide data which can be processed to obtain extracted object data. E.g. for sound data from a microphone, it may be determined that the extracted object is 'breaking glass', and further object annotations may be provided for proper rule checking, e.g. to allow to discern between a breaking glass bottle and a breaking glass window. A virtual object may e.g. be 'Sound of braking glass' and the rule may be: Object is 'Sound of breaking glass': then activate nearest camera to instantly view the scene.

Claims

1. Surveillance method for monitoring a location, comprising:
- acquiring sensor data from at least one sensor; - processing sensor data from the at least one sensor, in order to obtain an extracted object list;
- defining at least one virtual object; and
- applying at least one rule set, the at least one rule set defining possible responses depending on the extracted object list and the at least one virtual object, in which processing sensor data comprises indexing objects in the extracted object list to obtain an indexed extracted object list with a predefined set of index parameters, in which the at least one rule set comprises first type of rules relating to the predefined set of index parameters, and second type of rules, and in which applying the at least one rule set comprises - reducing the indexed extracted object list in size by applying the first type of rules using the predefined set of index parameters, and
- subsequently applying the second type of rules on the reduced extracted object list.
2. Method according to claim 1, further comprising analyzing the at least one rule set for determining the first type of rules and the second type of rules.
3. Method according to claim 2, further comprising storing results of applying the at least one rule set for application of a further rule set.
4. Method according to claim 1, 2 or 3, in which the extracted object list is updated depending on the update rate of the at least one sensor.
5. Method according to claim 4, in which indexing of objects is applied at the update rate or lower frequency.
6. Method according to any one of claims 1-5, in which the extracted object list is stored, and the at least one rule set is applied later in time.
7. Method according to claim 6, further comprising re-ordering the indexed extracted object list.
8. Method according to any one of claims 1-7, in which the at least one rule set comprises multiple, independent rule sets.
9. Method according to any one of claims 1-8, in which indexing objects in the extracted object list comprises determining for each extracted object in the extracted object list associated object attributes.
10. Method according to any one of claims 1-9, in which obtaining the extracted object list comprises consecutive operations of data enhancement, object finding, object analysis and object tracking.
11. Method according to claim 10, in which the sensor data comprises video data, and the data enhancement comprises one or more of the following data operations: noise reduction; image stabilization; contrast enhancement.
12. Method according to claim 10 or 11, in which object finding comprises one or more of the group of data operations comprising: edge analysis, texture analysis; motion analysis; background compensation.
13. Method according to any one of claims 10-12, in which object analysis comprises one or more of the group of data operations comprising: colour analysis, texture analysis; form analysis; object correlation.
14. Method according to any one of claims 10-13, in which object tracking comprises a combination of identity analysis and trajectory analysis.
15. Surveillance system comprising at least one sensor and a processing system connected to the at least one sensor, in which the processing system is arranged to execute the surveillance method according to any one of claims 1-14.
16. Surveillance system according to claim 15, in which the processing system comprises a local processing system located in the vicinity of the at least one sensor, and a central processing system, located remotely from the at least one sensor, and in which the local processing system is arranged to send the extracted object list to the central processing system.
17. Surveillance system according to claim 15 or 16, in which the local processing system further comprises a storage device for storing raw sensor data or preprocessed sensor data.
18. Surveillance system according to any one of claims 15-17, further comprising at least one operator console arranged for controlling the surveillance system.
19. Surveillance system according to claim 18, in which the at least one operator console comprises a representation device, which is arranged to represent simultaneously the sensor data, objects from the extracted object list and at least one virtual object in overlay.
EP07709185A 2007-01-31 2007-01-31 Surveillance method and system using optimized object based rule checking Withdrawn EP2115698A1 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/NL2007/050037 WO2008094029A1 (en) 2007-01-31 2007-01-31 Surveillance method and system using optimized object based rule checking

Publications (1)

Publication Number Publication Date
EP2115698A1 true EP2115698A1 (en) 2009-11-11

Family

ID=38515452

Family Applications (1)

Application Number Title Priority Date Filing Date
EP07709185A Withdrawn EP2115698A1 (en) 2007-01-31 2007-01-31 Surveillance method and system using optimized object based rule checking

Country Status (2)

Country Link
EP (1) EP2115698A1 (en)
WO (1) WO2008094029A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2893521A1 (en) * 2012-09-07 2015-07-15 Siemens Schweiz AG Methods and apparatus for establishing exit/entry criteria for a secure location
CN104077311B (en) 2013-03-28 2017-11-14 国际商业机器公司 Vehicle location indexing means and device
JP2016197795A (en) * 2015-04-03 2016-11-24 日立オートモティブシステムズ株式会社 Imaging device

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050169367A1 (en) * 2000-10-24 2005-08-04 Objectvideo, Inc. Video surveillance system employing video primitives

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050169367A1 (en) * 2000-10-24 2005-08-04 Objectvideo, Inc. Video surveillance system employing video primitives

Also Published As

Publication number Publication date
WO2008094029A1 (en) 2008-08-07

Similar Documents

Publication Publication Date Title
US20090315712A1 (en) Surveillance method and system using object based rule checking
JP5400138B2 (en) System, method, computer program and computer-readable medium for electronic monitoring
JP6088541B2 (en) Cloud-based video surveillance management system
JP3876288B2 (en) State recognition system and state recognition display generation method
KR101935399B1 (en) Wide Area Multi-Object Monitoring System Based on Deep Neural Network Algorithm
Zabłocki et al. Intelligent video surveillance systems for public spaces–a survey
KR100973930B1 (en) Parking ticket system having a multi-function camera, an electronic display and a broadcasting device
US20060200307A1 (en) Vehicle identification and tracking system
KR102039277B1 (en) Pedestrian face recognition system and method thereof
KR102144531B1 (en) Method for automatic monitoring selectively based in metadata of object employing analysis of images of deep learning
KR101492473B1 (en) Context-aware cctv intergrated managment system with user-based
JP5047382B2 (en) System and method for classifying moving objects during video surveillance
KR102282800B1 (en) Method for trackig multi target employing ridar and camera
CN110677619A (en) Intelligent monitoring video processing method
WO2008094029A1 (en) Surveillance method and system using optimized object based rule checking
KR20100077662A (en) Inteligent video surveillance system and method thereof
JP5712401B2 (en) Behavior monitoring system, behavior monitoring program, and behavior monitoring method
KR101453386B1 (en) Vehicle Intelligent Search System and Operating Method thereof
KR101686851B1 (en) Integrated control system using cctv camera
KR101669885B1 (en) method of detecting pedestrian collision and IoT apparatus thereof
KR102434154B1 (en) Method for tracking multi target in traffic image-monitoring-system
CN114143506A (en) Workshop monitoring management system based on intelligent camera
JP2005346545A (en) Monitoring device and program
Sahu Automated vehicle detection and tracking using Raspberry-Pi
US11727580B2 (en) Method and system for gathering information of an object moving in an area of interest

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20090831

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC NL PL PT RO SE SI SK TR

17Q First examination report despatched

Effective date: 20100330

DAX Request for extension of the european patent (deleted)
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20110707