WO2022113453A1 - Abnormality detection device and abnormality detection method - Google Patents

Abnormality detection device and abnormality detection method Download PDF

Info

Publication number
WO2022113453A1
WO2022113453A1 PCT/JP2021/031589 JP2021031589W WO2022113453A1 WO 2022113453 A1 WO2022113453 A1 WO 2022113453A1 JP 2021031589 W JP2021031589 W JP 2021031589W WO 2022113453 A1 WO2022113453 A1 WO 2022113453A1
Authority
WO
WIPO (PCT)
Prior art keywords
feature amount
node
unit
graph data
abnormality detection
Prior art date
Application number
PCT/JP2021/031589
Other languages
French (fr)
Japanese (ja)
Inventor
全 孔
智明 吉永
Original Assignee
株式会社日立製作所
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社日立製作所 filed Critical 株式会社日立製作所
Publication of WO2022113453A1 publication Critical patent/WO2022113453A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B25/00Alarm systems in which the location of the alarm condition is signalled to a central station, e.g. fire or police telegraphic systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Definitions

  • the present invention relates to an apparatus and a method for monitoring a monitored location and detecting an abnormality.
  • a person who behaves suspiciously is discovered based on images and images taken by a surveillance camera installed in a surveillance target place such as a public facility where an unspecified number of people enter and exit, and that person is regarded as a suspicious person.
  • a security system is used to prevent damage caused by threats such as crime and terrorist acts.
  • Judgment of such a suspicious person is difficult and highly depends on the professional skills of the guards.
  • the burden on the guards is heavy.
  • Patent Document 1 is known as a technique for automatically finding a suspicious person from an image of a surveillance camera.
  • an abnormal behavior detecting device for detecting an abnormal behavior of a behavioral body reflected in a time-series image the pattern or characteristic of the abnormal behavior is not defined in advance, and the behavior different from the normal defined behavior is taken.
  • a technique for extracting a behavioral body as an abnormal body candidate and determining whether or not the abnormal body candidate is an abnormal body is disclosed.
  • an object of the present invention is to provide a technique for accurately detecting suspicious behavior or abnormal behavior from images or images of various people or objects and detecting the abnormality.
  • the abnormality detection system detects a plurality of elements in the monitored place based on an image or an image obtained by photographing a predetermined monitored place, and the attributes of each element and the relationship between the elements.
  • the graph data generation unit that generates graph data representing the sex
  • the spatiotemporal feature amount calculation unit that calculates the spatiotemporal feature amount of the graph data generated by the graph data generation unit
  • the spatiotemporal feature amount calculation unit is provided.
  • the present invention it is possible to accurately detect suspicious behavior or abnormal behavior from images or images of various people or objects and detect the abnormality.
  • xxx table various information may be described by the expression of "xxx table”, but various information may be expressed by a data structure other than the table.
  • the "xxx table” may be referred to as "xxx information” in order to show that various types of information do not depend on the data structure.
  • a reference code (or a common part in the reference code) is used when the same type of element is not distinguished, and when the same type of element is described separately, the element ID (or the element ID) is used.
  • Element reference code may be used.
  • a process may be described with a "program” or its process as the subject, but the program is executed by a processor (for example, a CPU (Central Processing Unit)) to perform a defined process.
  • the subject of the process may be a processor because it is performed while appropriately using a storage resource (for example, a memory) and / or a communication interface device (for example, a communication port).
  • the processor operates as a functional unit that realizes a predetermined function by operating according to a program.
  • a device and system including a processor is a device and system including these functional parts.
  • FIG. 1 is a block diagram showing a configuration of an abnormality detection system according to the first embodiment of the present invention.
  • the abnormality detection system 1 of the present embodiment is a system that detects a threat or a sign thereof generated in the monitored place as an abnormality based on an image or an image obtained by photographing a predetermined monitored place with a surveillance camera. ..
  • the video or image used in the abnormality detection system 1 is a video or a moving image taken by a surveillance camera at a predetermined frame rate, and each is composed of a combination of a plurality of images acquired in time series. ..
  • the images and images handled by the abnormality detection system 1 will be collectively referred to as “images” and described.
  • the abnormality detection system 1 includes a camera moving image input unit 10, a graph data generation unit 20, a graph database 30, a graph data visualization editing unit 60, a node feature amount extraction unit 70, and an edge feature amount extraction unit 80.
  • Node feature amount storage unit 90 edge feature amount storage unit 100, spatiotemporal feature amount calculation unit 110, node feature amount acquisition unit 120, abnormality detection unit 130, threat sign degree storage unit 140, judgment basis presentation unit 150, and elements. It is configured to include a contribution storage unit 160.
  • the camera moving image input unit 10 the graph data generation unit 20, the graph data visualization editing unit 60, the node feature amount extraction unit 70, the edge feature amount extraction unit 80, the spatiotemporal feature amount calculation unit 110, and the node feature.
  • Each functional block of the quantity acquisition unit 120, the abnormality detection unit 130, and the judgment basis presentation unit 150 is realized by, for example, executing a predetermined program by a computer, and is realized by, for example, a graph database 30, a node feature amount storage unit 90, and an edge feature amount storage.
  • the unit 100, the threat sign storage unit 140, and the element contribution storage unit 160 are realized by using a storage device such as an HDD (Hard Disk Drive) or an SSD (Solid State Drive).
  • a part or all of these functional blocks may be realized by using GPU (Graphics Processing Unit) or FPGA (Field Programmable Gate Array).
  • the camera moving image input unit 10 acquires video (moving image) data taken by a surveillance camera (not shown) and inputs it to the graph data generation unit 20.
  • the graph data generation unit 20 extracts one or more elements to be monitored from various subjects reflected in the image based on the image data input from the camera moving image input unit 10, and the attributes and elements for each element. Generate graph data showing the relationship between them.
  • the element to be monitored extracted by the graph data generation unit 20 is a movement or movement at a monitoring target place where the surveillance camera is installed among various people and objects reflected in the image captured by the surveillance camera. A person or object that is stationary. However, it is preferable to exclude objects that are permanently installed in the monitored area and buildings in which the monitored area exists from the elements to be monitored.
  • the graph data generation unit 20 sets a plurality of time ranges for the video by dividing the time-series video data into predetermined time intervals ⁇ t, and generates graph data for each time range. Then, each generated graph data is recorded in the graph database 30 and output to the graph data visualization editing unit 60. The details of the graph data generation unit 20 will be described later with reference to FIGS. 2 and 3.
  • the graph database 30 stores the graph data generated by the graph data generation unit 20.
  • the graph database 30 has a node database 40 and an edge database 50.
  • the node database 40 stores node data representing the attributes of each element in the graph data
  • the edge database 50 stores edge data representing the relationships between the elements in the graph data. The details of the graph database 30, the node database 40, and the edge database 50 will be described later with reference to FIGS. 4, 5, and 6.
  • the graph data visualization editing unit 60 visualizes the graph data generated by the graph data generation unit 20 and presents it to the user, and accepts the user to edit the graph data.
  • the edited graph data is stored in the graph database 30. The details of the graph data visualization editing unit 60 will be described later with reference to FIG. 7.
  • the node feature amount extraction unit 70 extracts the node feature amount of each graph data based on the node data stored in the node database 40.
  • the node feature amount extracted by the node feature amount extraction unit 70 is a numerical value of the features possessed by the attributes of each element in each graph data, and is extracted for each node constituting each graph data.
  • the node feature amount extraction unit 70 stores the extracted node feature amount information in the node feature amount storage unit 90, and stores the weight used for calculating the node feature amount in the element contribution storage unit 160. The details of the node feature amount extraction unit 70 will be described later with reference to FIGS. 8 and 9.
  • the edge feature amount extraction unit 80 extracts the edge feature amount of each graph data based on the edge data stored in the edge database 50.
  • the edge feature amount extracted by the edge feature amount extraction unit 80 is a numerical value of the features having a relationship between the elements in each graph data, and is extracted for each edge constituting each graph data.
  • the edge feature amount extraction unit 80 stores the extracted edge feature amount information in the edge feature amount storage unit 100, and stores the weight used for calculating the edge feature amount in the element contribution storage unit 160. The details of the edge feature amount extraction unit 80 will be described later with reference to FIG.
  • the spatiotemporal feature amount calculation unit 110 calculates the spatiotemporal feature amount of the graph data based on the node feature amount and the edge feature amount of each graph accumulated in the node feature amount storage unit 90 and the edge feature amount storage unit 100, respectively. do.
  • the spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110 is the temporal and spatial feature amount of each graph data generated by the graph data generation unit 20 for each predetermined time interval ⁇ t with respect to the time-series video data. It is a numerical value of various features, and is calculated for each node that composes each graph data.
  • the spatiotemporal feature amount calculation unit 110 sets the feature amounts of other nodes adjacent to each node in the spatial direction and the temporal direction with respect to the node feature amount accumulated for each node, and the adjacent node.
  • a folding operation is performed in which weights are applied to the feature amounts of the edges set in between. By repeating such a convolution operation a plurality of times, it is possible to calculate the spatiotemporal feature amount that reflects the potential relationship between the feature amount of each node and the adjacent node.
  • the spatiotemporal feature amount calculation unit 110 updates the node feature amount accumulated in the node feature amount storage unit 90, reflecting the calculated spatiotemporal feature amount. The details of the spatiotemporal feature amount calculation unit 110 will be described later with reference to FIGS. 11, 12, and 13.
  • the node feature amount acquisition unit 120 acquires the node feature amount stored in the node feature amount storage unit 90 reflecting the spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110, and causes the abnormality detection unit 130 to acquire the node feature amount. input.
  • the abnormality detection unit 130 calculates the threat sign degree of each element reflected in the image captured by the surveillance camera based on the node feature amount input from the node feature amount acquisition unit 120.
  • the threat sign degree is a value indicating the degree to which the behavior or characteristic of a person or object corresponding to each element is considered to correspond to a threat such as a crime or terrorist act or a sign thereof. Then, if there is a person or a suspicious object that behaves suspiciously, it is detected based on the calculation result of the threat sign degree of each element.
  • the node feature amount input from the node feature amount acquisition unit 120 reflects the spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110 as described above.
  • the anomaly detection unit 130 detects an abnormality in the monitoring location where the surveillance camera is installed by calculating the threat sign degree of each element based on the spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110. It is something to do.
  • the abnormality detection unit 130 stores the calculated threat sign degree and abnormality detection result of each element in the threat sign degree storage unit 140. The details of the abnormality detection unit 130 will be described later with reference to FIGS. 14 and 15.
  • the determination basis presentation unit 150 stores each graph data stored in the graph database 30, the threat predictive degree for each element of each graph data stored in the threat predictive degree storage unit 140, and the element contribution degree storage unit 160.
  • An abnormality detection screen showing the processing result of the abnormality detection system 1 is presented to the user based on the weighting coefficient at the time of calculating the node feature amount and the edge feature amount.
  • the abnormality detection screen includes information on a person or object detected as a suspicious person or a suspicious object by the abnormality detection unit 130, as well as information indicating the grounds for the abnormality detection unit 130 to make the determination.
  • the user By looking at the abnormality detection screen presented by the determination basis presentation unit 150, the user detects which person or object is a suspicious person or suspicious object among various people or objects reflected in the image for what reason. You can check if it was done.
  • the details of the determination basis presentation unit 150 will be described later with reference to FIGS. 16, 17, and 18.
  • FIG. 2 is a block diagram showing the configuration of the graph data generation unit 20.
  • the graph data generation unit 20 includes an entity detection processing unit 21, an in-video co-reference analysis unit 22, and a relationship detection processing unit 23.
  • the entity detection processing unit 21 performs entity detection processing on the video data input from the camera moving image input unit 10.
  • the entity detection process performed by the entity detection processing unit 21 is a process of detecting a person or an object corresponding to a monitored element from a video and estimating the attribute of each element.
  • the entity detection processing unit 21 includes a person / object detection processing unit 210, a person / object tracking processing unit 211, and a person / object attribute estimation unit 212.
  • the person / object detection processing unit 210 uses a predetermined algorithm or tool (for example, OpenCV, Faster R-CNN, etc.) for each time range in which the time-series video data is divided by a predetermined time interval ⁇ t, and in the video. Detects people and objects reflected in the image as elements to be monitored. Then, a unique ID is assigned to each detected element as a node ID, a frame surrounding the area in the image of each element is set, and frame information regarding the position and size of the frame is acquired.
  • a predetermined algorithm or tool for example, OpenCV, Faster R-CNN, etc.
  • the person / object tracking processing unit 211 uses a predetermined object tracking algorithm or tool (for example, Deepsort) based on the frame information of each element acquired by the person / object detection processing unit 210 to display time-series video data. Track each element. Then, the tracking information indicating the result of the tracking process of each element is acquired and associated with the node ID of each element.
  • a predetermined object tracking algorithm or tool for example, Deepsort
  • the person / object attribute estimation unit 212 estimates the attributes of each element based on the tracking information of each element acquired by the person / object tracking processing unit 211.
  • the entropy of each frame extracted by sampling video data at a predetermined sampling rate eg: 1 fps
  • the attribute estimation of each element is performed using the image information of the person or the object in the frame having the highest calculated entropy value.
  • Attribute estimation is performed, for example, using a pre-learned attribute estimation model, such as the appearance or behavioral characteristics of a person or object, such as gender, age, clothing, maskedness, size, color, stay. Time etc. are estimated. Once the attributes of each element can be estimated, the attribute information is linked to the node ID of each element.
  • a pre-learned attribute estimation model such as the appearance or behavioral characteristics of a person or object, such as gender, age, clothing, maskedness, size, color, stay. Time etc.
  • various persons and objects reflected in the image are detected as elements to be monitored by the processing of each block described above, and the characteristics of each person and each object are acquired as attributes for each element.
  • a unique node ID is assigned to each element.
  • tracking information and attribute information of each element are set in association with the node ID.
  • These pieces of information are stored in the node database 40 as node data representing the characteristics of each element.
  • the in-video co-reference analysis unit 22 performs in-video co-reference analysis on the node data acquired by the entity detection processing unit 21.
  • the in-video co-reference analysis performed by the in-video co-reference analysis unit 22 is a process of correcting the node ID given to each element in the node data by mutually referencing the images of each frame in the video. be.
  • different node IDs may be erroneously assigned to the same person or object, and the frequency of occurrence varies depending on the performance of the algorithm.
  • the in-video co-reference analysis unit 22 corrects such an error in the node ID by performing the in-video co-reference analysis.
  • the in-video co-reference analysis unit 22 includes a maximum entropy frame sampling processing unit 220, a tracking matching processing unit 221 and a node ID updating unit 222.
  • the maximum entropy frame sampling processing unit 220 samples the frame having the highest entropy value in the video data, and reads the node data of each element detected in that frame from the node database 40. Then, based on the read node data, the template image of each element is acquired by extracting the image area corresponding to each element in the image of the frame.
  • the tracking matching processing unit 221 is based on the template image acquired by the maximum entropy frame sampling processing unit 220 and the tracking information included in the node data of each element read from the node database 40, between the frames. Perform template matching.
  • the range in which each element exists in the image of each frame is estimated from the tracking information, and template matching using the template image is performed within the estimated image range.
  • the node ID update unit 222 updates the node ID assigned to each element based on the result of the template matching of each element performed by the tracking matching processing unit 221.
  • the node data of each element stored in the node database 40 can be obtained. Align.
  • each time range set at the time interval ⁇ t interval is used.
  • Generate node data for each element in the graph data is stored in the node database 40 together with the graph ID uniquely set for each graph data.
  • the relationship detection processing unit 23 performs the relationship detection processing on the video data input from the camera moving image input unit 10 based on the node data whose node ID has been updated by the in-video co-reference analysis unit 22.
  • the relationship detection process performed by the relationship detection processing unit 23 is a process of detecting mutual relationships with respect to a person or an object detected as a monitored element by the entity detection processing unit 21.
  • the relationship detection processing unit 23 includes a person / object relationship detection processing unit 230 and a person behavior detection processing unit 231.
  • the person / object relationship detection processing unit 230 detects the relationship between the person and the object reflected in the image based on the node data of each element read from the node database 40.
  • actions such as "carrying”, “opening”, and “leaving” that a person performs on an object such as luggage are detected as the relationship between the two. do.
  • the person behavior detection processing unit 231 detects the interaction behavior between people reflected in the video based on the node data of each element read from the node database 40.
  • actions such as "conversation” and “delivery” performed by a plurality of people together are detected as interaction actions between each person.
  • the relationship detection processing unit 23 detects the action performed by one person on another person or object with respect to the person or object detected as the monitored element by the entity detection processing unit 21 by the processing of each block described above. And the behavior is acquired as a mutual relationship. This information is stored in the edge database 50 as edge data representing the relationship between each element.
  • FIG. 3 is a diagram showing an outline of the processing performed by the graph data generation unit 20.
  • the graph data generation unit 20 is the object 3 carried by the person 2 and the person 2 from the image taken by the camera moving image input unit 10 by the entity detection process performed by the entity detection processing unit 21. Are detected and these are tracked in the video.
  • the relationship detection process performed by the relationship detection processing unit 23 detects the relationship between the person 2 and the object 3. Then, based on these processing results, graph data composed of a plurality of nodes and edges is generated for each fixed time interval ⁇ t.
  • the person 2 is represented by the node P1 and the object 3 is represented by the node O1, and attribute information indicating the characteristics thereof is set for each of these nodes.
  • an edge of "carrying" indicating the relationship between the person 2 and the object 3 is set between the node P1 and the node O1.
  • the information of the graph data thus generated is stored in the graph database 30.
  • FIG. 4 is a diagram showing an example of the data structure of the graph database 30.
  • the graph database 30 is represented by, for example, a data table including columns 301 to 304.
  • Column 301 stores a series of reference numbers set for each row of the data table.
  • a graph ID unique to each graph data is stored in the column 302.
  • the start time and end time of the time range corresponding to each graph data are stored in columns 303 and 304, respectively.
  • the start time and end time are calculated from the shooting start time and shooting end time recorded in the video used to generate each graph data, and the difference is equal to the above-mentioned time interval ⁇ t.
  • the graph database 30 is configured by storing this information row by row for each graph data.
  • FIG. 5 is a diagram showing an example of a data structure of the node database 40.
  • the node database 40 is composed of a node attribute table 41 shown in FIG. 5A, a tracking information table 42 shown in FIG. 5B, and a frame information table 43 shown in FIG. 5C.
  • the node attribute table 41 is represented by, for example, a data table including columns 411 to 414.
  • Column 411 stores a series of reference numbers set for each row of the data table.
  • the graph ID of the graph data to which each node belongs is stored in the column 412.
  • the value of this graph ID is associated with the value of the graph ID stored in the column 302 in the data table of FIG. 4, whereby each node is associated with the graph data.
  • Column 413 stores a node ID unique to each node.
  • Column 414 stores the attribute information acquired for the element represented by each node.
  • the node attribute table 41 is configured by storing this information row by row for each node.
  • the tracking information table 42 is represented by, for example, a data table including columns 421 to 424.
  • Column 421 stores a series of reference numbers set for each row of the data table.
  • the node ID of the node targeted by each tracking information is stored. The value of this node ID is associated with the value of the node ID stored in column 413 in the data table of FIG. 5A, whereby each tracking information is associated with the node.
  • Column 423 stores a track ID unique to each tracking information.
  • Column 424 stores a list of frame IDs of each frame in which the element represented by the node is reflected in the video.
  • the tracking information table 42 is configured by storing this information row by row for each tracking information.
  • the frame information table 43 is represented by, for example, a data table including columns 431 to 434.
  • Column 431 stores a series of reference numbers set for each row of the data table.
  • the track ID of the tracking information to which each frame information belongs is stored in the column 432.
  • the value of the track ID is associated with the value of the track ID stored in the column 423 in the data table of FIG. 5B, whereby each frame information and the tracking information are associated with each other.
  • a frame ID unique to each frame information is stored in the column 433.
  • the column 434 stores information indicating the position of each element in the frame represented by the frame information and the type of each element (person, object, etc.).
  • the frame information table 43 is configured by storing this information row by row for each frame information.
  • FIG. 6 is a diagram showing an example of the data structure of the edge database 50.
  • the edge database 50 is represented by, for example, a data table containing columns 501-506.
  • Column 501 stores a series of reference numbers set for each row of the data table.
  • the graph ID of the graph data to which each edge belongs is stored in the column 502.
  • the value of this graph ID is associated with the value of the graph ID stored in the column 302 in the data table of FIG. 4, whereby each edge is associated with the graph data.
  • the nodes 503 and 504 store the node IDs of the nodes located at the start point and the end point of each edge, respectively.
  • the values of these node IDs are associated with the values of the node IDs stored in column 413 in the data table of FIG.
  • each edge represents the relationship between which nodes. Is specified.
  • Column 505 stores an edge ID unique to each edge.
  • Column 506 stores, as edge information representing the relationship between the elements represented by the edge, the content of the action performed by the person corresponding to the start point node on another person or object corresponding to the end point node.
  • the edge database 50 is configured by storing this information row by row for each edge.
  • FIG. 7 is an explanatory diagram of the graph data visualization editing unit 60.
  • the graph data visualization editing unit 60 displays, for example, the graph data editing screen 61 shown in FIG. 7 on a display (not shown) and presents it to the user.
  • the user can arbitrarily edit the graph data by performing a predetermined operation.
  • the graph data 610 generated by the graph data generation unit 20 is visualized and displayed on the graph data editing screen 61.
  • the user selects an arbitrary node or edge on the screen to display the node information boxes 611 and 612 showing the detailed information of the node and the edge information box 613 showing the detailed information of the edge. Can be done.
  • the attribute information of each node is displayed in these information boxes 611 to 613.
  • the user can edit the content of each attribute information shown in the underlined portion by selecting arbitrary attribute information in the information boxes 611 to 613.
  • a node addition button 614 and an edge addition button 615 are displayed together with the graph data 610.
  • the user can add a node or an edge to the graph data 610 at an arbitrary position by selecting the node addition button 614 or the edge addition button 615 on the screen. Further, by selecting an arbitrary node or edge in the graph data 610 and performing a predetermined operation (for example, dragging or right-clicking the mouse), the node or edge can be moved or deleted.
  • the graph data visualization editing unit 60 can appropriately edit the contents of the generated graph data by the user's operation as described above. Then, the graph database 30 is updated to reflect the edited graph data.
  • FIG. 8 is a block diagram showing the configuration of the node feature amount extraction unit 70.
  • the node feature amount extraction unit 70 includes a maximum entropy frame sampling processing unit 71, a person / object area image acquisition unit 72, an image feature amount calculation unit 73, an attribute information acquisition unit 74, and an attribute information feature amount calculation. It is configured to include a unit 75, a feature amount coupling processing unit 76, an attribute weight calculation attention mechanism 77, and a node feature amount calculation unit 78.
  • the maximum entropy frame sampling processing unit 71 reads the node data of each node from the node database 40, and samples the frame having the maximum entropy in the video for each node.
  • the person / object area image acquisition unit 72 acquires the area image of the person or object corresponding to the element represented by each node from the frame sampled by the maximum entropy frame sampling processing unit 71.
  • the image feature amount calculation unit 73 calculates the image feature amount for each element represented by each node from the area image of each person or each object acquired by the person / object area image acquisition unit 72.
  • a DNN Deep Neural Network
  • MSCOCO Large-scale image data set
  • the image feature amount is calculated by extracting the output from the intermediate layer. If the image feature amount can be calculated for the area image of each element, another method may be used.
  • the attribute information acquisition unit 74 reads the node information of each node from the node database 40 and acquires the attribute information of each node.
  • the attribute information feature amount calculation unit 75 calculates the feature amount of the attribute information for each element represented by each node from the attribute information acquired by the attribute information acquisition unit 74.
  • a predetermined language processing algorithm for example, word2Vec
  • the attribute items for example, age, clothes, presence / absence of wearing a mask
  • the feature amount combination processing unit 76 performs a combination process of combining the image feature amount calculated by the image feature amount calculation unit 73 and the feature amount of the attribute information calculated by the attribute information feature amount calculation unit 75.
  • the feature amount for the feature of the whole person or object represented by the image feature amount and the feature amount for each attribute item of the person or object represented by the attribute information are set as vector components according to the feature amount of each of these items. Create a feature vector for each element.
  • the attribute weight calculation attention mechanism 77 acquires the weight for each item of the feature amount for the feature amount combined by the feature amount combination processing unit 76.
  • the weights learned in advance are acquired for each vector component of the feature amount vector.
  • Attribute weight calculation The weight information acquired by the attention mechanism 77 is stored in the element contribution storage unit 160 as an element contribution indicating the contribution of each node feature amount to the threat sign degree calculated by the anomaly detection unit 130. Will be done.
  • the node feature amount calculation unit 78 performs weighting processing by multiplying the feature amount combined by the feature amount combination processing unit 76 by the weight acquired by the attribute weight calculation attention mechanism 77, and calculates the node feature amount. calculate. That is, the node feature amount is calculated by summing the values obtained by multiplying each vector component of the feature amount vector by the weight set by the attribute weight calculation attention mechanism 77.
  • the node feature amount representing the attribute feature amount for each element is generated for each graph data generated for each time range set in the time interval ⁇ t interval by the processing of each block described above. Be extracted.
  • the extracted node feature amount information is stored in the node feature amount storage unit 90.
  • FIG. 9 is a diagram showing an outline of the processing performed by the node feature amount extraction unit 70.
  • the node feature amount extraction unit 70 calculates the image feature amount by the image feature amount calculation unit 73 for the frame having the maximum entropy of the person 2 in the video corresponding to each graph data.
  • the attribute information feature amount calculation unit 75 calculates the feature amount for each attribute item of the attribute information of the node P1 corresponding to the person 2, so that the "whole body feature amount”, “mask”, “skin color”, and “skin color” are calculated.
  • the feature amount of the node P1 is obtained for each item such as "stay time".
  • the feature amount of the node P1 is extracted by performing the weighting operation for each of these items by the node feature amount calculation unit 78 using the weight acquired by the attribute weight calculation attention mechanism 77.
  • the feature amount of each node of the graph data can be obtained.
  • the weight acquired by the attribute weight calculation attention mechanism 77 is stored in the element contribution storage unit 160 as the element contribution.
  • FIG. 10 is a block diagram showing the configuration of the edge feature amount extraction unit 80.
  • the edge feature amount extraction unit 80 includes an edge information acquisition unit 81, an edge feature amount calculation unit 82, an edge weight calculation attention mechanism 83, and a weighting calculation unit 84.
  • the edge information acquisition unit 81 reads and acquires the edge information of each edge from the edge database 50.
  • the edge feature amount calculation unit 82 calculates the edge feature amount, which is the feature amount of the relationship between the elements represented by each edge, from the edge information acquired by the edge information acquisition unit 81.
  • the edge feature amount is calculated by using a predetermined language processing algorithm (for example, word2Vec) for text data such as "passing” and "conversation” representing the action contents set as edge information. ..
  • the edge weight calculation attention mechanism 83 acquires the weight for the edge feature amount calculated by the edge feature amount calculation unit 82.
  • the weight learned in advance is acquired for the edge feature amount.
  • the weight information acquired by the edge weight calculation attention mechanism 83 is stored in the element contribution storage unit 160 as an element contribution representing the contribution of the edge feature amount to the threat sign degree calculated by the abnormality detection unit 130.
  • the weighting calculation unit 84 performs weighting processing by multiplying the edge feature amount calculated by the edge feature amount calculation unit 82 by the weight acquired by the edge weight calculation attention mechanism 83, and the weighted edge feature. Calculate the amount.
  • the edge feature amount extraction unit 80 the edge feature amount representing the feature amount of the relationship between the elements for each graph data generated for each time range set in the time interval ⁇ t interval by the processing of each block described above. Is extracted.
  • the extracted edge feature amount information is stored in the edge feature amount storage unit 100.
  • FIG. 11 is a block diagram showing the configuration of the spatiotemporal feature amount calculation unit 110.
  • the spatiotemporal feature amount calculation unit 110 includes a plurality of residual convolution calculation blocks 111 and a node feature amount update unit 112.
  • Each residual convolution calculation block 111 corresponds to a predetermined number of stages, receives the calculation result of the residual convolution calculation block 111 in the previous stage, executes the convolution operation, and inputs the convolution operation to the residual convolution calculation block 111 in the subsequent stage.
  • the node feature amount and the edge feature amount read from the node feature amount storage unit 90 and the edge feature amount storage unit 100 are input to the residual convolution calculation block 111 in the front stage, and the residual convolution calculation block 111 in the final stage is input.
  • the calculation result of is input to the node feature amount update unit 112.
  • GNN Graph Neural Network
  • each residual convolution calculation block 111 includes two space convolution calculation processing units 1110 and one time convolution calculation processing unit 1111.
  • the space convolution calculation processing unit 1110 calculates the outer product of the feature amount of the adjacent node adjacent to each node in the graph data and the feature amount of the edge set between each node and the adjacent node as the convolution calculation in the spatial direction. Then, a weighting operation using a weight matrix of D ⁇ D size is performed on this outer product.
  • the value of the degree D of the weight matrix is defined as the length of the feature amount of each node. This ensures a variety of learning using a learnable weighted linear transformation. Further, since the weight matrix can be designed without being restricted by the number of nodes and edges constituting the graph data, the weighting operation can be performed using the optimum weight matrix.
  • the spatial convolution calculation processing unit 1110 performs a weighting calculation twice for each node constituting the graph data. As a result, the convolution operation in the spatial direction is realized.
  • the time-convolution calculation processing unit 1111 performs a time-direction convolution calculation on the feature amount of each node for which the spatial-direction convolution calculation is performed by the two space convolution calculation processing units 1110.
  • the node adjacent to each node in the time direction that is, the feature amount of the node representing the same person or object as the node in the graph data generated for the video in the adjacent time range, and the adjacent node.
  • the outer product with the set feature amount of the edge is calculated, and the weighting operation similar to that of the space convolution calculation processing unit 1110 is performed on this outer product. This realizes a convolution operation in the time direction.
  • the operation result of the residual convolution operation block 111 is obtained. Desired. By performing such an operation, it is possible to perform a convolution operation in which the features of both the adjacent nodes adjacent to each other in the spatial direction and the temporal direction and the edges between the adjacent nodes are added to the features of each node at the same time.
  • the node feature amount update unit 112 updates the feature amount of each node accumulated in the node feature amount storage unit 90 by using the calculation result output from the residual convolution calculation block 111 in the final stage. As a result, the spatiotemporal features calculated for each node constituting the graph data are reflected in the features of each node.
  • the spatiotemporal feature amount calculation unit 110 can calculate the spatiotemporal feature amount of each graph data using GNN by the processing of each block described above, and can update the node feature amount by reflecting it on the node feature amount. ..
  • the GNN learning in the spatiotemporal feature calculation unit 110 it is preferable to learn the residual function with reference to the input of any layer. In this way, even if the layer at the time of learning is deep, a gradient explosion or a gradient explosion occurs. The problem of vanishing gradient can be prevented. Therefore, it is possible to calculate the node features that reflect more accurate spatiotemporal information.
  • FIG. 12 is a diagram showing an example of a mathematical formula representing arithmetic processing in the space convolution arithmetic processing unit 1110.
  • the feature amount after spatial convolution is calculated, and the time convolution calculation is further performed to perform the spatiotemporal feature amount. Is calculated and reflected in the node features.
  • the convolution operation performed by the space convolution calculation processing unit 1110 and the convolution calculation performed by the time convolution calculation processing unit 1111 are expressed by the following mathematical formulas (1) and (2), respectively.
  • O represents the concatenation or average pooling
  • represents the non-linear activation function
  • l represents the layer number of the GNN corresponding to the space convolution arithmetic processing unit 1110.
  • k represents the layer number of the GNN corresponding to the time convolution calculation processing unit 1111.
  • H N ⁇ D represents a matrix of spatial node features
  • N represents the number of nodes in the graph data
  • D represents the length (order) of the node features.
  • M i L ⁇ D represents a matrix of time node features for the i-th node
  • L represents the length of time.
  • EN ⁇ N ⁇ P represents a matrix of edge features
  • Eij represents an edge feature (order P) connecting the i-th node and the j-th node.
  • Eij 0.
  • Fi 1 ⁇ D represents a matrix of time node features for the i -th node.
  • Fij represents the existence or nonexistence of the jth node in the jth graph data.
  • Q1 ⁇ L represents a convolutional kernel for weighting the relationship between the nodes in the time direction
  • WS l is D regarding the node features in the spatial direction. It represents a weighting matrix of ⁇ D size
  • WTk represents a weighting matrix of D ⁇ D size regarding the node features in the time direction.
  • FIG. 13 is a diagram showing an outline of the processing performed by the spatiotemporal feature amount calculation unit 110.
  • the dotted line represents the space convolution calculation by the space convolution calculation processing unit 1110
  • the broken line represents the time convolution calculation by the time convolution calculation processing unit 1111.
  • node 3 in the t-th graph data has a space corresponding to the feature amounts of adjacent nodes 1 and 4 and the feature amount of the edge set between these adjacent nodes.
  • Features are added by spatial convolution.
  • the time feature amount corresponding to the feature amount of the node 3 in the immediately preceding t-1st graph data and the feature amount of the node 3 in the immediately following t + 1st graph data is added by the time convolution operation.
  • the spatiotemporal feature amount of the t-th graph data for the node 3 is calculated and reflected in the feature amount of the node 3.
  • FIG. 14 is a block diagram showing the configuration of the abnormality detection unit 130.
  • the abnormality detection unit 130 includes a feature quantity distribution clustering unit 131, a center point distance calculation unit 132, and an abnormality determination unit 133.
  • the feature amount distribution clustering unit 131 performs clustering processing of the feature amount of each node acquired from the node feature amount storage unit 90 by the node feature amount acquisition unit 120, and obtains the distribution of the node feature amount.
  • the distribution of the node features is obtained by plotting the features of each node on a two-dimensional map.
  • the center point distance calculation unit 132 calculates the distance from the center point of each node feature amount in the distribution of the node feature amount obtained by the feature amount distribution clustering unit 131. As a result, the features of each node reflecting the spatiotemporal features are compared with each other. The distance from the center point of each node feature amount calculated by the center point distance calculation unit 132 is stored in the threat sign degree storage unit 140 as a threat sign degree indicating the degree of threat of the element corresponding to each node.
  • the abnormality determination unit 133 determines the threat sign degree of each node based on the distance calculated by the center point distance calculation unit 132. As a result, if there is a node with a threat sign of more than a predetermined value, the element corresponding to that node is determined to be a suspicious person or a suspicious object, an abnormality in the monitored location is detected, and the user is notified. .. Notification to the user is performed using, for example, an alarm device (not shown). At this time, the position of an element determined to be a suspicious person or a suspicious object may be highlighted in the image of the surveillance camera.
  • the abnormality detection result by the abnormality determination unit 133 is stored in the threat sign degree storage unit 140 in association with the threat sign degree.
  • the anomaly detection unit 130 detects an abnormality in the monitored location based on the spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110 by the processing of each block described above, and also detects the spatiotemporal feature for each element.
  • the quantities can be compared with each other, and the threat predictiveness for each element can be obtained based on the comparison result.
  • FIG. 15 is a diagram showing an outline of the processing performed by the abnormality detection unit 130.
  • the abnormality detection unit 130 plots the node features for which the spatiotemporal features have been determined for each node of the graph data including the nodes P3, P6, and O2 on a two-dimensional map. Find the distribution of node features. Then, the central point of the distribution of the obtained node feature amount is obtained, and the distance from this center point to each node feature amount is calculated to obtain the threat sign degree of each node.
  • the element corresponding to the node whose threat sign degree is equal to or higher than the predetermined value for example, the person corresponding to the node P6 whose node feature amount is outside the distribution circle 4 on the distribution map is a suspicious person or a suspicious object. Judgment is made and an abnormality is detected.
  • FIG. 16 is a block diagram showing the configuration of the determination basis presentation unit 150.
  • the judgment basis presentation unit 150 includes a basis confirmation target selection unit 151, a subgraph extraction processing unit 152, a person attribute threat contribution presentation unit 153, an object attribute threat contribution presentation unit 154, and an action history contribution presentation. It is configured to include a unit 155 and a verbalization summary generation unit 156.
  • the basis confirmation target selection unit 151 acquires the threat sign degree stored in the threat sign degree storage unit 140, and includes the node in which the abnormality is detected by the abnormality detection unit 130 based on the acquired threat sign degree of each node. Select any part of the graph data as the target for confirming the basis for abnormality detection.
  • the part related to the node with the highest threat sign may be automatically selected, or an arbitrary node may be specified according to the user's operation and the part related to that node may be selected. good.
  • the subgraph extraction processing unit 152 acquires the graph data stored in the graph database 30, and extracts the portion selected by the basis confirmation target selection unit 151 in the acquired graph data as a subgraph indicating the basis confirmation target for abnormality detection. do. For example, the node with the highest threat sign or the node specified by the user, each node connected to the node, and each edge are extracted as a subgraph.
  • the person attribute threat contribution presentation unit 153 calculates and visualizes the contribution of the person's attribute to the threat sign degree. Present to the user. For example, the element contribution degree stored in the element contribution degree storage unit 160 for various attribute items (gender, age, clothes, whether or not a mask is worn, staying time, etc.) represented by the attribute information included in the node information of the node. That is, the contribution of each attribute item is calculated based on the weight of each attribute item with respect to the node feature amount. Then, a predetermined number of attribute items are selected from the one with the highest calculated contribution, and the content and contribution of each attribute item are presented in a predetermined layout on the abnormality detection screen.
  • attribute items for various attribute items (gender, age, clothes, whether or not a mask is worn, staying time, etc.) represented by the attribute information included in the node information of the node. That is, the contribution of each attribute item is calculated based on the weight of each attribute item with respect to the node feature amount.
  • a predetermined number of attribute items are
  • the object attribute threat contribution presentation unit 154 calculates and visualizes the contribution of the object attribute to the threat sign degree. Present to the user. For example, for various attribute items (size, color, staying time, etc.) represented by the attribute information included in the node information of the node, the element contribution degree stored in the element contribution degree storage unit 160, that is, the attribute for the node feature amount. Calculate the contribution of each attribute item based on the weight of each item. Then, a predetermined number of attribute items are selected from the one with the highest calculated contribution, and the content and contribution of each attribute item are presented in a predetermined layout on the abnormality detection screen.
  • attribute items size, color, staying time, etc.
  • the action history contribution presentation unit 155 is an action performed between the person or object and another person or object when the node included in the subgraph extracted by the subgraph extraction processing unit 152 represents a person or an object. Calculates the degree of contribution to the threat sign by, visualizes it, and presents it to the user. For example, for each edge connected to the node, the contribution of each edge is calculated based on the element contribution stored in the element contribution storage unit 160, that is, the weight for the edge feature amount. Then, a predetermined number of edges are selected from the one with the highest calculated contribution, and the action content and contribution represented by each edge are presented in a predetermined layout on the abnormality detection screen.
  • the verbalization summary generation unit 156 verbalizes the contents presented by the person attribute threat contribution presentation unit 153, the object attribute threat contribution presentation unit 154, and the action history contribution presentation unit 155, respectively, and is the basis for abnormality detection. Generate a text (summary) that expresses concisely. Then, the generated summary is displayed at a predetermined position on the abnormality detection screen.
  • the threat predictive degree and the threat predictive degree calculated for the element such as a person or an object in which the abnormality is detected by the abnormality detection unit 130 by the processing of each block described above
  • An abnormality detection screen including at least information on the characteristics or behaviors of the element having a high degree of contribution to the above can be presented to the user as a screen showing the determination basis of the abnormality detection unit 130.
  • FIG. 17 is a diagram showing an outline of the processing performed by the evidence confirmation target selection unit 151 and the subgraph extraction processing unit 152.
  • (a) shows an example of visualizing the graph data before the subgraph extraction
  • (b) shows an example of visualizing the graph data after the subgraph extraction.
  • the basis confirmation target selection unit 151 is connected to the designated node and the node. Select each node and each edge as the target for checking the basis for abnormality detection.
  • the subgraph extraction processing unit 152 extracts the nodes and edges selected by the basis confirmation target selection unit 151 as subgraphs, highlights the extracted subgraphs, and grays out and displays the parts other than the subgraphs of the graph data. By doing so, the subgraph is visualized.
  • the portion including the designated node O2, the nodes P2 and P4 adjacent to the node O2, and the edges set between the nodes O2, P2, and P4 are selected by the grounds confirmation target selection unit 151. It is extracted as a subgraph from the subgraph extraction processing unit 152. Then, as shown in FIG. 17 (b), these extracted nodes and edges are highlighted, and the other parts are grayed out, so that the subgraph is visualized.
  • FIG. 18 is a diagram showing an example of an abnormality detection screen displayed by the determination basis presentation unit 150.
  • the threat sign degree is shown as a threat level
  • the characteristics and the contribution degree of each action to the threat sign degree are shown. .. Specifically, the contribution to each item of "mask”, “stay time”, and “upper body color” is shown for the person photographed by the camera 2, and “left behind” for the object photographed by the camera 1. The degree of contribution to each item of "stay time” and "delivery” is shown.
  • a summary generated by the verbalization summary generation unit 156 is displayed. Further, a video showing a suspicious action taken by a person and the shooting time thereof are displayed as an action timeline.
  • the abnormality detection screen 180 shown in FIG. 18 is an example, and if the abnormality detection result by the abnormality detection unit 130 and its basis can be presented in an easy-to-understand manner for the user, the abnormality detection screen is displayed with other contents and screen layout. May be good.
  • the abnormality detection system 1 detects a plurality of elements in a monitored place based on an image or an image obtained by photographing a predetermined monitored place, and determines the attributes of each element and the relationship between the elements.
  • the graph data generation unit 20 that generates the graph data to be represented
  • the spatiotemporal feature amount calculation unit 110 that calculates the spatiotemporal feature amount of the graph data generated by the graph data generation unit 20, and the spatiotemporal feature amount calculation unit 110.
  • An abnormality detecting unit 130 for detecting an abnormality in a monitored place is provided based on the spatiotemporal feature amount. By doing so, it is possible to accurately detect suspicious behavior or abnormal behavior from images or images of various people or objects and detect the abnormality.
  • the graph data generation unit 20 detects a person or object reflected in a video or image as an element by the entity detection processing unit 21, acquires the characteristics of the detected person or object as an element, and acquires the characteristics of the detected person or object as an attribute for each element.
  • the relationship detection processing unit 23 acquires actions taken by a person with respect to another person or object as relationships between elements. Since this is done, the information necessary for generating the graph data can be reliably acquired from the video or the image, and the graph data can be generated.
  • the graph data generation unit 20 combines a plurality of nodes representing the attributes of each element and a plurality of edges representing the relationships between the elements to generate graph data as shown in FIG. 3, for example. By doing so, it is possible to generate graph data that clearly expresses the attributes of each element and the relationships between the elements.
  • the graph data generation unit 20 generates graph data for each of a plurality of time ranges obtained by dividing the video or image acquired in time series into predetermined time intervals ⁇ t. Since this is done, it is possible to generate appropriate graph data in consideration of the balance between the amount of data and the accuracy of abnormality detection for a video or image that changes over time.
  • the abnormality detection system 1 has a node feature amount extraction unit 70 that extracts a node feature amount representing an attribute feature amount for each element for each graph data generated for each predetermined time range, and a node feature amount extraction unit 70 for each time range.
  • Each of the generated graph data is further provided with an edge feature amount extraction unit 80 for extracting an edge feature amount representing the feature amount of the relationship between the elements.
  • the spatiotemporal feature amount calculation unit 110 calculates the spatiotemporal feature amount based on the node feature amount and the edge feature amount of each graph data extracted by the node feature amount extraction unit 70 and the edge feature amount extraction unit 80, respectively. Since this is done, it is possible to calculate the spatiotemporal feature amount that reflects the potential relationship between the feature amount of each node and the adjacent node.
  • the abnormality detection system 1 further includes a graph data visualization editing unit 60 that visualizes graph data, presents it to the user, and accepts the user to edit the graph data. By doing so, even if inappropriate graph data is erroneously generated, the user can correct it to the correct graph data.
  • the spatiotemporal feature amount calculation unit 110 calculates the spatiotemporal feature amount for each element.
  • the anomaly detection unit 130 compares the spatiotemporal feature amounts of each element calculated by the spatiotemporal feature amount calculation unit 110 with each other by the feature amount distribution clustering unit 131 and the center point distance calculation unit 132, and the abnormality determination unit 133 uses the abnormality determination unit 133. Anomalies are detected by calculating the degree of threat for each element based on the comparison result. Since this is done, when there is a person or a suspicious object that behaves suspiciously, it can be reliably determined and the abnormality can be detected.
  • the degree of threat calculated for the element and the degree of contribution to the degree of threat are determined.
  • An abnormality detection screen 180 including at least high information on the characteristics or behavior of the element is presented to the user as a screen showing the determination basis of the abnormality detection unit 130.
  • the abnormality detection screen 180 further includes information on the degree of contribution to the degree of threat for each feature or action of the element. By doing so, it is possible to make the user understand in an easy-to-understand manner what characteristics and behaviors of the person or object detected as a suspicious person or a suspicious object are emphasized in the judgment of the threat.
  • the computer constituting the abnormality detection system 1 detects a plurality of elements in the monitored place based on the video or image obtained by photographing the predetermined monitored place (the entity detection processing unit 21 performs the process). Entity detection processing to be executed), processing to generate graph data showing attributes of each element and relationships between elements (processing of graph data generation unit 20), and processing to calculate spatiotemporal feature amount of graph data (time).
  • the processing of the spatial feature amount calculation unit 110) and the processing of detecting an abnormality in the monitored place based on the spatiotemporal feature amount (processing of the abnormality detecting unit 130) are executed. Since this is done, it is possible to accurately detect suspicious behaviors and abnormal behaviors and detect abnormalities from images or images of various people and objects photographed by processing using a computer.
  • the present invention is not limited to the above embodiment, and can be carried out by using any component within the range not deviating from the gist thereof.
  • the embodiments and modifications described above are merely examples, and the present invention is not limited to these contents as long as the features of the invention are not impaired. Further, although various embodiments and modifications have been described above, the present invention is not limited to these contents. Other aspects considered within the scope of the technical idea of the present invention are also included within the scope of the present invention.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Emergency Management (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)
  • Alarm Systems (AREA)
  • Closed-Circuit Television Systems (AREA)

Abstract

This abnormality detection system comprises: a graph data generation unit for detecting a plurality of elements at a prescribed monitoring target place and generating graph data indicating attributes of the respective elements and a relationship among the elements, on the basis of a video or an image acquired by photographing the monitoring target place; a spatial-temporal feature amount calculation unit for calculating the spatial-temporal feature amount of the graph data generated by the graph data generation unit; and an abnormality detection unit that detects an abnormality at the monitoring target place on the basis of the spatial-temporal feature amount calculated by the spatial-temporal feature amount calculation unit.

Description

異常検知装置、異常検知方法Anomaly detection device, anomaly detection method
 本発明は、監視対象場所を監視して異常を検知する装置および方法に関する。 The present invention relates to an apparatus and a method for monitoring a monitored location and detecting an abnormality.
 従来、不特定多数の人間が出入りする公共施設等の監視対象場所に設置された監視カメラにより撮影された映像や画像を元に、不審な行動をとる人物を発見し、その人物を不審者として扱うことで、犯罪やテロ行為等の脅威による被害を未然に防ぐ警備システムが利用されている。このような警備システムでは、監視カメラが撮影した映像や画像に映り込んだ多数の人物から不審者を発見する必要があり、従来では警備員が映像や画像を実際に見て判断することで不審者の発見を行っていた。こうした不審者の判断は難易度が高く、警備員の専門的なスキルに高度に依存する。また、監視中には常に映像や画像を見続ける必要があるため、警備員の負担も大きい。 Conventionally, a person who behaves suspiciously is discovered based on images and images taken by a surveillance camera installed in a surveillance target place such as a public facility where an unspecified number of people enter and exit, and that person is regarded as a suspicious person. By handling it, a security system is used to prevent damage caused by threats such as crime and terrorist acts. In such a security system, it is necessary to detect a suspicious person from a large number of people reflected in the images and images taken by the surveillance camera, and in the past, the guards actually saw the images and images to make a suspicious decision. Was discovering the person. Judgment of such a suspicious person is difficult and highly depends on the professional skills of the guards. In addition, since it is necessary to constantly watch images and images during monitoring, the burden on the guards is heavy.
 近年、日本をはじめとする先進諸国では、労働人口の減少による警備員の不足が深刻化しており、上記のように専門的なスキルを持った警備員の確保が困難となっている。また、上記の警備システムでは映像を監視する警備員を常駐させる必要があり、そのためのコストが高いという課題もある。そこで、これらの課題を解決するために、監視カメラの映像や画像から不審者を自動的に発見する技術が求められている。 In recent years, in developed countries such as Japan, the shortage of security guards has become serious due to the decrease in the working population, and it has become difficult to secure security guards with specialized skills as described above. Further, in the above-mentioned security system, it is necessary to have a security guard who monitors the image permanently stationed, and there is also a problem that the cost for that is high. Therefore, in order to solve these problems, there is a demand for a technique for automatically detecting a suspicious person from the images and images of a surveillance camera.
 監視カメラの画像から不審者を自動的に発見する技術として、例えば特許文献1の技術が知られている。特許文献1には、時系列画像に映った行動体の異常行動を検出する異常行動検出装置において、事前に異常行動のパターンや特徴を定義せず、通常と定義した行動とは異なる行動をとった行動体を異常体候補として抽出し、その異常体候補が異常体であるかどうかを判断する技術が開示されている。 For example, the technique of Patent Document 1 is known as a technique for automatically finding a suspicious person from an image of a surveillance camera. In Patent Document 1, in an abnormal behavior detecting device for detecting an abnormal behavior of a behavioral body reflected in a time-series image, the pattern or characteristic of the abnormal behavior is not defined in advance, and the behavior different from the normal defined behavior is taken. A technique for extracting a behavioral body as an abnormal body candidate and determining whether or not the abnormal body candidate is an abnormal body is disclosed.
日本国特許第6692086号Japanese Patent No. 6692086
 特許文献1の技術では、ある時点で行動体が行った行動のみに注目して、異常な行動であるか否かを判断している。しかしながら、不審者がとり得る実際の行動では、同じ映像内に写った他の要素との関連性や、時間ごとに行われた行動間の関連性などを考慮してその行動を解釈しなければ、異常な行動であるか否かを正しく判断することが困難である。したがって、特許文献1の技術では、不審者を異常体として確実に判断することができない可能性がある。 In the technique of Patent Document 1, whether or not the behavior is abnormal is determined by paying attention only to the behavior performed by the behavioral body at a certain point in time. However, in the actual action that a suspicious person can take, the action must be interpreted in consideration of the relationship with other elements shown in the same image and the relationship between the actions performed at each time. , It is difficult to correctly judge whether or not the behavior is abnormal. Therefore, the technique of Patent Document 1 may not be able to reliably determine a suspicious person as an abnormal body.
 そこで、本発明では、様々な人物や物体を撮影した映像または画像から、不審な行動や異常な行動を正確に発見して異常を検知する技術を提供することを目的とする。 Therefore, an object of the present invention is to provide a technique for accurately detecting suspicious behavior or abnormal behavior from images or images of various people or objects and detecting the abnormality.
 本発明による異常検知システムは、所定の監視対象場所を撮影して得られた映像または画像に基づいて、前記監視対象場所における複数の要素を検知し、前記要素ごとの属性および前記要素間の関係性を表すグラフデータを生成するグラフデータ生成部と、前記グラフデータ生成部により生成された前記グラフデータの時空間特徴量を算出する時空間特徴量算出部と、前記時空間特徴量算出部により算出された前記時空間特徴量に基づいて、前記監視対象場所における異常を検知する異常検知部と、を備える。 The abnormality detection system according to the present invention detects a plurality of elements in the monitored place based on an image or an image obtained by photographing a predetermined monitored place, and the attributes of each element and the relationship between the elements. The graph data generation unit that generates graph data representing the sex, the spatiotemporal feature amount calculation unit that calculates the spatiotemporal feature amount of the graph data generated by the graph data generation unit, and the spatiotemporal feature amount calculation unit. An abnormality detecting unit for detecting an abnormality in the monitored place based on the calculated spatiotemporal feature amount is provided.
 本発明によれば、様々な人物や物体を撮影した映像または画像から、不審な行動や異常な行動を正確に発見して異常を検知することができる。 According to the present invention, it is possible to accurately detect suspicious behavior or abnormal behavior from images or images of various people or objects and detect the abnormality.
本発明の第1の実施形態に係る異常検知システムの構成を示すブロック図である。It is a block diagram which shows the structure of the abnormality detection system which concerns on 1st Embodiment of this invention. グラフデータ生成部の構成を示すブロック図である。It is a block diagram which shows the structure of the graph data generation part. グラフデータ生成部が行う処理の概要を示す図である。It is a figure which shows the outline of the processing performed by the graph data generation part. グラフデータベースのデータ構造例を示す図である。It is a figure which shows the data structure example of a graph database. ノードデータベースのデータ構造例を示す図である。It is a figure which shows the data structure example of a node database. エッジデータベースのデータ構造例を示す図である。It is a figure which shows the data structure example of an edge database. グラフデータ可視化編集部の説明図である。It is explanatory drawing of the graph data visualization editorial department. ノード特徴量抽出部の構成を示すブロック図である。It is a block diagram which shows the structure of a node feature amount extraction part. ノード特徴量抽出部が行う処理の概要を示す図である。It is a figure which shows the outline of the processing performed by a node feature amount extraction unit. エッジ特徴量抽出部の構成を示すブロック図である。It is a block diagram which shows the structure of the edge feature amount extraction part. 時空間特徴量算出部の構成を示すブロック図である。It is a block diagram which shows the structure of the spatiotemporal feature amount calculation part. 時空間特徴量算出部における演算処理を表す数式の一例を示す図である。It is a figure which shows an example of the mathematical expression which represents the arithmetic processing in the space-time feature amount calculation part. 時空間特徴量算出部が行う処理の概要を示す図である。It is a figure which shows the outline of the processing performed by the spatiotemporal feature amount calculation unit. 異常検知部の構成を示すブロック図である。It is a block diagram which shows the structure of an abnormality detection part. 異常検知部が行う処理の概要を示す図である。It is a figure which shows the outline of the processing performed by an abnormality detection unit. 判定根拠提示部の構成を示すブロック図である。It is a block diagram which shows the structure of the judgment basis presentation part. 根拠確認対象選択部およびサブグラフ抽出処理部が行う処理の概要を示す図である。It is a figure which shows the outline of the processing performed by the basis confirmation target selection unit and the subgraph extraction processing unit. 判定根拠提示部により表示される異常検知画面の例を示す図である。It is a figure which shows the example of the abnormality detection screen displayed by the judgment basis presentation part.
 以下、図面を参照して本発明の実施形態を説明する。説明の明確化のため、以下の記載及び図面は、適宜、省略及び簡略化がなされている。本発明が本実施形態に制限されることは無く、本発明の思想に合致するあらゆる応用例が本発明の技術的範囲に含まれる。特に限定しない限り、各構成要素は複数でも単数でも構わない。 Hereinafter, embodiments of the present invention will be described with reference to the drawings. In order to clarify the explanation, the following description and drawings are omitted or simplified as appropriate. The present invention is not limited to this embodiment, and any application example consistent with the idea of the present invention is included in the technical scope of the present invention. Unless otherwise specified, each component may be plural or singular.
 以下の説明では、例えば、「xxx表」の表現にて各種情報を説明することがあるが、各種情報は表以外のデータ構造で表現されていてもよい。各種情報がデータ構造に依存しないことを示すために、「xxx表」を「xxx情報」と呼ぶことがある。 In the following description, for example, various information may be described by the expression of "xxx table", but various information may be expressed by a data structure other than the table. The "xxx table" may be referred to as "xxx information" in order to show that various types of information do not depend on the data structure.
 また、以下の説明では、同種の要素を区別しないで説明する場合には、参照符号(又は参照符号における共通部分)を使用し、同種の要素を区別して説明する場合は、要素のID(又は要素の参照符号)を使用することがある。 Further, in the following description, a reference code (or a common part in the reference code) is used when the same type of element is not distinguished, and when the same type of element is described separately, the element ID (or the element ID) is used. Element reference code) may be used.
 以下の説明では、「プログラム」あるいはそのプロセスを主語として処理を説明する場合があるが、プログラムは、プロセッサ(例えば、CPU(Central Processing Unit))によって実行されることで、定められた処理を、適宜に記憶資源(例えば、メモリ)及び/又は通信インタフェース装置(例えば、通信ポート)を用いながら行うため、処理の主語がプロセッサであってもよい。プロセッサは、プログラムに従って動作することによって、所定の機能を実現する機能部として動作する。プロセッサを含む装置及びシステムは、これらの機能部を含む装置及びシステムである。 In the following description, a process may be described with a "program" or its process as the subject, but the program is executed by a processor (for example, a CPU (Central Processing Unit)) to perform a defined process. The subject of the process may be a processor because it is performed while appropriately using a storage resource (for example, a memory) and / or a communication interface device (for example, a communication port). The processor operates as a functional unit that realizes a predetermined function by operating according to a program. A device and system including a processor is a device and system including these functional parts.
[第1の実施形態]
 以下、本発明の第1の実施形態について説明する。
[First Embodiment]
Hereinafter, the first embodiment of the present invention will be described.
 図1は、本発明の第1の実施形態に係る異常検知システムの構成を示すブロック図である。本実施形態の異常検知システム1は、監視カメラにより所定の監視対象場所を撮影して得られた映像または画像に基づいて、監視対象場所において発生する脅威やその予兆を異常として検知するシステムである。なお、異常検知システム1において用いられる映像または画像とは、監視カメラにより所定のフレームレートで撮影された映像または動画像であり、いずれも時系列で取得された複数の画像の組み合わせによって構成される。以下では、異常検知システム1が取り扱う映像と画像をまとめて、単に「映像」と称して説明する。 FIG. 1 is a block diagram showing a configuration of an abnormality detection system according to the first embodiment of the present invention. The abnormality detection system 1 of the present embodiment is a system that detects a threat or a sign thereof generated in the monitored place as an abnormality based on an image or an image obtained by photographing a predetermined monitored place with a surveillance camera. .. The video or image used in the abnormality detection system 1 is a video or a moving image taken by a surveillance camera at a predetermined frame rate, and each is composed of a combination of a plurality of images acquired in time series. .. In the following, the images and images handled by the abnormality detection system 1 will be collectively referred to as “images” and described.
 図1に示すように、異常検知システム1は、カメラ動画像入力部10、グラフデータ生成部20、グラフデータベース30、グラフデータ可視化編集部60、ノード特徴量抽出部70、エッジ特徴量抽出部80、ノード特徴量蓄積部90、エッジ特徴量蓄積部100、時空間特徴量算出部110、ノード特徴量取得部120、異常検知部130、脅威予兆度保存部140、判定根拠提示部150、および要素寄与度保存部160を備えて構成される。異常検知システム1において、カメラ動画像入力部10、グラフデータ生成部20、グラフデータ可視化編集部60、ノード特徴量抽出部70、エッジ特徴量抽出部80、時空間特徴量算出部110、ノード特徴量取得部120、異常検知部130、判定根拠提示部150の各機能ブロックは、例えばコンピュータが所定のプログラムを実行することにより実現され、グラフデータベース30、ノード特徴量蓄積部90、エッジ特徴量蓄積部100、脅威予兆度保存部140、要素寄与度保存部160は、HDD(Hard Disk Drive)やSSD(Solid State Drive)等の記憶装置を用いて実現される。なお、これらの機能ブロックの一部または全部を、GPU(Graphics Processing Unit)やFPGA(Field Programmable Gate Array)を用いて実現してもよい。 As shown in FIG. 1, the abnormality detection system 1 includes a camera moving image input unit 10, a graph data generation unit 20, a graph database 30, a graph data visualization editing unit 60, a node feature amount extraction unit 70, and an edge feature amount extraction unit 80. , Node feature amount storage unit 90, edge feature amount storage unit 100, spatiotemporal feature amount calculation unit 110, node feature amount acquisition unit 120, abnormality detection unit 130, threat sign degree storage unit 140, judgment basis presentation unit 150, and elements. It is configured to include a contribution storage unit 160. In the abnormality detection system 1, the camera moving image input unit 10, the graph data generation unit 20, the graph data visualization editing unit 60, the node feature amount extraction unit 70, the edge feature amount extraction unit 80, the spatiotemporal feature amount calculation unit 110, and the node feature. Each functional block of the quantity acquisition unit 120, the abnormality detection unit 130, and the judgment basis presentation unit 150 is realized by, for example, executing a predetermined program by a computer, and is realized by, for example, a graph database 30, a node feature amount storage unit 90, and an edge feature amount storage. The unit 100, the threat sign storage unit 140, and the element contribution storage unit 160 are realized by using a storage device such as an HDD (Hard Disk Drive) or an SSD (Solid State Drive). In addition, a part or all of these functional blocks may be realized by using GPU (Graphics Processing Unit) or FPGA (Field Programmable Gate Array).
 カメラ動画像入力部10は、不図示の監視カメラにより撮影された映像(動画像)のデータを取得し、グラフデータ生成部20に入力する。 The camera moving image input unit 10 acquires video (moving image) data taken by a surveillance camera (not shown) and inputs it to the graph data generation unit 20.
 グラフデータ生成部20は、カメラ動画像入力部10から入力された映像データに基づいて、映像に映り込んだ様々な被写体から監視対象の要素を単数または複数抽出し、その要素ごとの属性および要素間の関係性を表すグラフデータを生成する。ここで、グラフデータ生成部20において抽出される監視対象の要素とは、監視カメラにより撮影された映像に映り込んだ様々な人物や物体のうち、監視カメラが設置された監視対象場所において移動または静止している人物や物体のことである。ただし、監視対象場所に常設されている物体や、監視対象場所が存在する建造物などは、監視対象の要素から除外することが好ましい。 The graph data generation unit 20 extracts one or more elements to be monitored from various subjects reflected in the image based on the image data input from the camera moving image input unit 10, and the attributes and elements for each element. Generate graph data showing the relationship between them. Here, the element to be monitored extracted by the graph data generation unit 20 is a movement or movement at a monitoring target place where the surveillance camera is installed among various people and objects reflected in the image captured by the surveillance camera. A person or object that is stationary. However, it is preferable to exclude objects that are permanently installed in the monitored area and buildings in which the monitored area exists from the elements to be monitored.
 グラフデータ生成部20は、時系列の映像データを所定の時刻区間Δtごとに区切ることで映像に対して複数の時間範囲を設定し、その時間範囲ごとにグラフデータを生成する。そして、生成した各グラフデータをグラフデータベース30に記録するとともに、グラフデータ可視化編集部60に出力する。なお、グラフデータ生成部20の詳細は、後で図2、図3を参照して説明する。 The graph data generation unit 20 sets a plurality of time ranges for the video by dividing the time-series video data into predetermined time intervals Δt, and generates graph data for each time range. Then, each generated graph data is recorded in the graph database 30 and output to the graph data visualization editing unit 60. The details of the graph data generation unit 20 will be described later with reference to FIGS. 2 and 3.
 グラフデータベース30には、グラフデータ生成部20により生成されたグラフデータが格納される。グラフデータベース30は、ノードデータベース40およびエッジデータベース50を有している。ノードデータベース40には、グラフデータにおいて各要素の属性を表すノードのデータが格納され、エッジデータベース50には、グラフデータにおいて各要素間の関係性を表すエッジのデータが格納される。なお、グラフデータベース30、ノードデータベース40およびエッジデータベース50の詳細は、後で図4、図5、図6を参照して説明する。 The graph database 30 stores the graph data generated by the graph data generation unit 20. The graph database 30 has a node database 40 and an edge database 50. The node database 40 stores node data representing the attributes of each element in the graph data, and the edge database 50 stores edge data representing the relationships between the elements in the graph data. The details of the graph database 30, the node database 40, and the edge database 50 will be described later with reference to FIGS. 4, 5, and 6.
 グラフデータ可視化編集部60は、グラフデータ生成部20により生成されたグラフデータを可視化してユーザに提示するとともに、ユーザによるグラフデータの編集を受け付ける。編集後のグラフデータは、グラフデータベース30に格納される。なお、グラフデータ可視化編集部60の詳細は、後で図7を参照して説明する。 The graph data visualization editing unit 60 visualizes the graph data generated by the graph data generation unit 20 and presents it to the user, and accepts the user to edit the graph data. The edited graph data is stored in the graph database 30. The details of the graph data visualization editing unit 60 will be described later with reference to FIG. 7.
 ノード特徴量抽出部70は、ノードデータベース40に格納されたノードデータに基づいて、各グラフデータのノード特徴量を抽出する。ノード特徴量抽出部70が抽出するノード特徴量とは、各グラフデータにおける要素ごとの属性が有する特徴を数値化したものであり、各グラフデータを構成するノードごとに抽出される。ノード特徴量抽出部70は、抽出したノード特徴量の情報をノード特徴量蓄積部90に格納するとともに、ノード特徴量の算出に用いた重みを要素寄与度保存部160に格納する。なお、ノード特徴量抽出部70の詳細は、後で図8、図9を参照して説明する。 The node feature amount extraction unit 70 extracts the node feature amount of each graph data based on the node data stored in the node database 40. The node feature amount extracted by the node feature amount extraction unit 70 is a numerical value of the features possessed by the attributes of each element in each graph data, and is extracted for each node constituting each graph data. The node feature amount extraction unit 70 stores the extracted node feature amount information in the node feature amount storage unit 90, and stores the weight used for calculating the node feature amount in the element contribution storage unit 160. The details of the node feature amount extraction unit 70 will be described later with reference to FIGS. 8 and 9.
 エッジ特徴量抽出部80は、エッジデータベース50に格納されたエッジデータに基づいて、各グラフデータのエッジ特徴量を抽出する。エッジ特徴量抽出部80が抽出するエッジ特徴量とは、各グラフデータにおける要素間の関係性が有する特徴を数値化したものであり、各グラフデータを構成するエッジごとに抽出される。エッジ特徴量抽出部80は、抽出したエッジ特徴量の情報をエッジ特徴量蓄積部100に格納するとともに、エッジ特徴量の算出に用いた重みを要素寄与度保存部160に格納する。なお、エッジ特徴量抽出部80の詳細は、後で図10を参照して説明する。 The edge feature amount extraction unit 80 extracts the edge feature amount of each graph data based on the edge data stored in the edge database 50. The edge feature amount extracted by the edge feature amount extraction unit 80 is a numerical value of the features having a relationship between the elements in each graph data, and is extracted for each edge constituting each graph data. The edge feature amount extraction unit 80 stores the extracted edge feature amount information in the edge feature amount storage unit 100, and stores the weight used for calculating the edge feature amount in the element contribution storage unit 160. The details of the edge feature amount extraction unit 80 will be described later with reference to FIG.
 時空間特徴量算出部110は、ノード特徴量蓄積部90とエッジ特徴量蓄積部100にそれぞれ蓄積された各グラフのノード特徴量およびエッジ特徴量に基づいて、グラフデータの時空間特徴量を算出する。時空間特徴量算出部110が算出する時空間特徴量とは、グラフデータ生成部20において時系列の映像データに対して所定の時刻区間Δtごとに生成された各グラフデータの時間的および空間的な特徴を数値化したものであり、各グラフデータを構成するノードごとに算出される。時空間特徴量算出部110は、各ノードについて蓄積されたノード特徴量に対して、空間方向と時間方向のそれぞれにおいて各ノードと隣接関係にある他のノードの特徴量と、当該隣接ノードとの間に設定されているエッジの特徴量とにそれぞれ重み付けを行って加える畳み込み操作を行う。こうした畳み込み操作を複数回繰り返して行うことにより、各ノードの特徴量に隣接ノードとの潜在的な関係性を反映した時空間特徴量を算出することができる。時空間特徴量算出部110は、算出した時空間特徴量を反映して、ノード特徴量蓄積部90に蓄積されたノード特徴量を更新する。なお、時空間特徴量算出部110の詳細は、後で図11、図12、図13を参照して説明する。 The spatiotemporal feature amount calculation unit 110 calculates the spatiotemporal feature amount of the graph data based on the node feature amount and the edge feature amount of each graph accumulated in the node feature amount storage unit 90 and the edge feature amount storage unit 100, respectively. do. The spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110 is the temporal and spatial feature amount of each graph data generated by the graph data generation unit 20 for each predetermined time interval Δt with respect to the time-series video data. It is a numerical value of various features, and is calculated for each node that composes each graph data. The spatiotemporal feature amount calculation unit 110 sets the feature amounts of other nodes adjacent to each node in the spatial direction and the temporal direction with respect to the node feature amount accumulated for each node, and the adjacent node. A folding operation is performed in which weights are applied to the feature amounts of the edges set in between. By repeating such a convolution operation a plurality of times, it is possible to calculate the spatiotemporal feature amount that reflects the potential relationship between the feature amount of each node and the adjacent node. The spatiotemporal feature amount calculation unit 110 updates the node feature amount accumulated in the node feature amount storage unit 90, reflecting the calculated spatiotemporal feature amount. The details of the spatiotemporal feature amount calculation unit 110 will be described later with reference to FIGS. 11, 12, and 13.
 ノード特徴量取得部120は、時空間特徴量算出部110により算出された時空間特徴量が反映されてノード特徴量蓄積部90に蓄積されているノード特徴量を取得し、異常検知部130に入力する。 The node feature amount acquisition unit 120 acquires the node feature amount stored in the node feature amount storage unit 90 reflecting the spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110, and causes the abnormality detection unit 130 to acquire the node feature amount. input.
 異常検知部130は、ノード特徴量取得部120から入力されたノード特徴量に基づいて、監視カメラにより撮影された映像に映り込んだ各要素の脅威予兆度を算出する。脅威予兆度とは、各要素に対応する人物や物体の行動や特徴が、犯罪やテロ行為等の脅威またはその予兆に該当すると考えられる度合いを示す値である。そして、各要素の脅威予兆度の算出結果に基づいて、不審な行動をとる人物や不審物が存在する場合には、これを検知する。ここで、ノード特徴量取得部120から入力されるノード特徴量には、前述のように、時空間特徴量算出部110により算出された時空間特徴量が反映されている。すなわち、異常検知部130は、時空間特徴量算出部110により算出された時空間特徴量に基づいて各要素の脅威予兆度を算出することで、監視カメラが設置された監視場所における異常を検知するものである。異常検知部130は、算出した各要素の脅威予兆度と異常検知結果を脅威予兆度保存部140に格納する。なお、異常検知部130の詳細は、後で図14、図15を参照して説明する。 The abnormality detection unit 130 calculates the threat sign degree of each element reflected in the image captured by the surveillance camera based on the node feature amount input from the node feature amount acquisition unit 120. The threat sign degree is a value indicating the degree to which the behavior or characteristic of a person or object corresponding to each element is considered to correspond to a threat such as a crime or terrorist act or a sign thereof. Then, if there is a person or a suspicious object that behaves suspiciously, it is detected based on the calculation result of the threat sign degree of each element. Here, the node feature amount input from the node feature amount acquisition unit 120 reflects the spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110 as described above. That is, the anomaly detection unit 130 detects an abnormality in the monitoring location where the surveillance camera is installed by calculating the threat sign degree of each element based on the spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110. It is something to do. The abnormality detection unit 130 stores the calculated threat sign degree and abnormality detection result of each element in the threat sign degree storage unit 140. The details of the abnormality detection unit 130 will be described later with reference to FIGS. 14 and 15.
 判定根拠提示部150は、グラフデータベース30に格納された各グラフデータと、脅威予兆度保存部140に格納された各グラフデータの要素ごとの脅威予兆度と、要素寄与度保存部160に格納されたノード特徴量およびエッジ特徴量の算出時の重み付け係数とに基づいて、異常検知システム1の処理結果を示す異常検知画面をユーザに提示する。この異常検知画面には、異常検知部130により不審者または不審物として検知された人物や物体の情報とともに、異常検知部130がその判定を下した根拠を示す情報も含まれている。ユーザは、判定根拠提示部150により提示された異常検知画面を見ることで、映像に映り込んだ様々な人物や物体のうち、どの人物または物体がどのような理由で不審者または不審物として検知されたのかを確認することができる。なお、判定根拠提示部150の詳細は、後で図16、図17、図18を参照して説明する。 The determination basis presentation unit 150 stores each graph data stored in the graph database 30, the threat predictive degree for each element of each graph data stored in the threat predictive degree storage unit 140, and the element contribution degree storage unit 160. An abnormality detection screen showing the processing result of the abnormality detection system 1 is presented to the user based on the weighting coefficient at the time of calculating the node feature amount and the edge feature amount. The abnormality detection screen includes information on a person or object detected as a suspicious person or a suspicious object by the abnormality detection unit 130, as well as information indicating the grounds for the abnormality detection unit 130 to make the determination. By looking at the abnormality detection screen presented by the determination basis presentation unit 150, the user detects which person or object is a suspicious person or suspicious object among various people or objects reflected in the image for what reason. You can check if it was done. The details of the determination basis presentation unit 150 will be described later with reference to FIGS. 16, 17, and 18.
 続いて、上記の各機能ブロックの詳細を以下に説明する。 Next, the details of each of the above functional blocks will be described below.
 図2は、グラフデータ生成部20の構成を示すブロック図である。図2(a)に示すように、グラフデータ生成部20は、エンティティ検知処理部21、映像内共参照解析部22および関係性検知処理部23を備えて構成される。 FIG. 2 is a block diagram showing the configuration of the graph data generation unit 20. As shown in FIG. 2A, the graph data generation unit 20 includes an entity detection processing unit 21, an in-video co-reference analysis unit 22, and a relationship detection processing unit 23.
 エンティティ検知処理部21は、カメラ動画像入力部10から入力される映像データに対してエンティティ検知処理を実施する。エンティティ検知処理部21が行うエンティティ検知処理とは、映像から監視対象要素に該当する人物や物体を検知し、各要素の属性を推定する処理のことである。図2(b)に示すように、エンティティ検知処理部21は、人物/物体検知処理部210、人物/物体追跡処理部211および人物/物体属性推定部212を備える。 The entity detection processing unit 21 performs entity detection processing on the video data input from the camera moving image input unit 10. The entity detection process performed by the entity detection processing unit 21 is a process of detecting a person or an object corresponding to a monitored element from a video and estimating the attribute of each element. As shown in FIG. 2B, the entity detection processing unit 21 includes a person / object detection processing unit 210, a person / object tracking processing unit 211, and a person / object attribute estimation unit 212.
 人物/物体検知処理部210は、時系列の映像データを所定の時刻区間Δtごとに区切った各時間範囲について、所定のアルゴリズムやツール(例えばOpenCVやFaster R-CNNなど)を用いて、映像内に映り込んだ人物や物体を監視対象の要素として検知する。そして、検知した各要素に対してユニークなIDをノードIDとして付与するとともに、各要素の映像内の領域を囲う枠を設定し、その枠の位置や大きさに関する枠情報を取得する。 The person / object detection processing unit 210 uses a predetermined algorithm or tool (for example, OpenCV, Faster R-CNN, etc.) for each time range in which the time-series video data is divided by a predetermined time interval Δt, and in the video. Detects people and objects reflected in the image as elements to be monitored. Then, a unique ID is assigned to each detected element as a node ID, a frame surrounding the area in the image of each element is set, and frame information regarding the position and size of the frame is acquired.
 人物/物体追跡処理部211は、人物/物体検知処理部210により取得された各要素の枠情報に基づき、所定の物体追跡アルゴリズムやツール(例えばDeepsortなど)を用いて、時系列の映像データにおける各要素の追跡処理を行う。そして、各要素の追跡処理の結果を示す追跡情報を取得し、各要素のノードIDに紐付ける。 The person / object tracking processing unit 211 uses a predetermined object tracking algorithm or tool (for example, Deepsort) based on the frame information of each element acquired by the person / object detection processing unit 210 to display time-series video data. Track each element. Then, the tracking information indicating the result of the tracking process of each element is acquired and associated with the node ID of each element.
 人物/物体属性推定部212は、人物/物体追跡処理部211により取得された各要素の追跡情報に基づいて、各要素の属性推定を行う。ここでは、例えば所定のサンプリングレート(例:1fps)で映像データをサンプリングすることで抽出された各フレームのエントロピーを計算する。各フレームのエントロピーは、例えば各フレームの検知結果の信頼度をpとすると(p∈{0,1})、H=plog(1-p)で計算される。そして、算出したエントロピーの値が最も高いフレームにおける人物や物体の画像情報を用いて、各要素の属性推定を行う。属性の推定は、例えば、事前に学習した属性推定モデルを用いて行われ、人物や物体の外見的または行動的な特徴、例えば性別、年齢、服装、マスク着用の有無、大きさ、色、滞在時間などが推定される。各要素の属性が推定できたら、その属性情報を各要素のノードIDに紐付ける。 The person / object attribute estimation unit 212 estimates the attributes of each element based on the tracking information of each element acquired by the person / object tracking processing unit 211. Here, for example, the entropy of each frame extracted by sampling video data at a predetermined sampling rate (eg: 1 fps) is calculated. The entropy of each frame is calculated by H = plog (1-p), for example, assuming that the reliability of the detection result of each frame is p (p ∈ {0,1}). Then, the attribute estimation of each element is performed using the image information of the person or the object in the frame having the highest calculated entropy value. Attribute estimation is performed, for example, using a pre-learned attribute estimation model, such as the appearance or behavioral characteristics of a person or object, such as gender, age, clothing, maskedness, size, color, stay. Time etc. are estimated. Once the attributes of each element can be estimated, the attribute information is linked to the node ID of each element.
 エンティティ検知処理部21では、以上説明した各ブロックの処理により、映像に映り込んだ様々な人物や物体が監視対象の要素としてそれぞれ検知され、各人物や各物体の特徴が要素ごとの属性として取得されるとともに、要素ごとにユニークなノードIDが付与される。そして、ノードIDに紐付けて各要素の追跡情報や属性情報が設定される。これらの情報は、各要素の特徴を表すノードデータとして、ノードデータベース40に格納される。 In the entity detection processing unit 21, various persons and objects reflected in the image are detected as elements to be monitored by the processing of each block described above, and the characteristics of each person and each object are acquired as attributes for each element. At the same time, a unique node ID is assigned to each element. Then, tracking information and attribute information of each element are set in association with the node ID. These pieces of information are stored in the node database 40 as node data representing the characteristics of each element.
 映像内共参照解析部22は、エンティティ検知処理部21により取得されたノードデータに対して映像内共参照解析を実施する。映像内共参照解析部22が行う映像内共参照解析とは、映像内の各フレームの画像を相互に参照することで、ノードデータにおいて各要素に付与されたノードIDを修正する処理のことである。エンティティ検知処理部21が行うエンティティ検知処理では、同一の人物や物体に対して異なるノードIDが誤って付与されることがあり、その発生頻度はアルゴリズムの性能によって変わる。映像内共参照解析部22は、映像内共参照解析を実施することで、こうしたノードIDの誤りを修正する。図2(c)に示すように、映像内共参照解析部22は、最大エントロピーフレームサンプリング処理部220、追跡マッチング処理部221およびノードID更新部222を備える。 The in-video co-reference analysis unit 22 performs in-video co-reference analysis on the node data acquired by the entity detection processing unit 21. The in-video co-reference analysis performed by the in-video co-reference analysis unit 22 is a process of correcting the node ID given to each element in the node data by mutually referencing the images of each frame in the video. be. In the entity detection process performed by the entity detection processing unit 21, different node IDs may be erroneously assigned to the same person or object, and the frequency of occurrence varies depending on the performance of the algorithm. The in-video co-reference analysis unit 22 corrects such an error in the node ID by performing the in-video co-reference analysis. As shown in FIG. 2C, the in-video co-reference analysis unit 22 includes a maximum entropy frame sampling processing unit 220, a tracking matching processing unit 221 and a node ID updating unit 222.
 最大エントロピーフレームサンプリング処理部220は、映像データにおいてエントロピーの値が最も高いフレームをサンプリングし、そのフレームにおいて検知された各要素のノードデータをノードデータベース40から読み出す。そして、読み出したノードデータに基づき、当該フレームの画像内で各要素に対応する画像領域を抽出することで、各要素のテンプレート画像を取得する。 The maximum entropy frame sampling processing unit 220 samples the frame having the highest entropy value in the video data, and reads the node data of each element detected in that frame from the node database 40. Then, based on the read node data, the template image of each element is acquired by extracting the image area corresponding to each element in the image of the frame.
 追跡マッチング処理部221は、最大エントロピーフレームサンプリング処理部220により取得されたテンプレート画像と、ノードデータベース40から読み出された各要素のノードデータに含まれる追跡情報とに基づいて、各フレーム間でのテンプレートマッチングを行う。ここでは、追跡情報から各要素が各フレームの画像においてどの範囲に存在するかを推定し、推定した画像範囲内でテンプレート画像を用いたテンプレートマッチングを行う。 The tracking matching processing unit 221 is based on the template image acquired by the maximum entropy frame sampling processing unit 220 and the tracking information included in the node data of each element read from the node database 40, between the frames. Perform template matching. Here, the range in which each element exists in the image of each frame is estimated from the tracking information, and template matching using the template image is performed within the estimated image range.
 ノードID更新部222は、追跡マッチング処理部221により行われた各要素のテンプレートマッチングの結果に基づいて、各要素に付与されたノードIDを更新する。ここでは、テンプレートマッチングにより複数のフレーム間で互いに同一の人物または物体としてマッチングされた要素に対して、共通のノードIDを付与することで、ノードデータベース40に格納されている各要素のノードデータを整合させる。そして、整合されたノードデータを一定の時刻区間Δtごとに区切って属性情報と追跡情報をそれぞれ分割し、各要素のノードIDに紐付けることで、時刻区間Δt間隔で設定された時間範囲ごとのグラフデータにおける各要素のノードデータを生成する。こうして生成されたノードデータは、グラフデータごとにユニークに設定されるグラフIDとともに、ノードデータベース40に格納される。 The node ID update unit 222 updates the node ID assigned to each element based on the result of the template matching of each element performed by the tracking matching processing unit 221. Here, by assigning a common node ID to the elements matched as the same person or object among a plurality of frames by template matching, the node data of each element stored in the node database 40 can be obtained. Align. Then, by dividing the matched node data into fixed time interval Δt, dividing the attribute information and tracking information, and linking them to the node ID of each element, each time range set at the time interval Δt interval is used. Generate node data for each element in the graph data. The node data generated in this way is stored in the node database 40 together with the graph ID uniquely set for each graph data.
 関係性検知処理部23は、映像内共参照解析部22によりノードIDを更新されたノードデータに基づき、カメラ動画像入力部10から入力される映像データに対して関係性検知処理を実施する。関係性検知処理部23が行う関係性検知処理とは、エンティティ検知処理部21により監視対象要素として検知された人物や物体に対して、相互の関係性を検知する処理のことである。図2(d)に示すように、関係性検知処理部23は、人物・物体関係検知処理部230および人物行動検知処理部231を備える。 The relationship detection processing unit 23 performs the relationship detection processing on the video data input from the camera moving image input unit 10 based on the node data whose node ID has been updated by the in-video co-reference analysis unit 22. The relationship detection process performed by the relationship detection processing unit 23 is a process of detecting mutual relationships with respect to a person or an object detected as a monitored element by the entity detection processing unit 21. As shown in FIG. 2D, the relationship detection processing unit 23 includes a person / object relationship detection processing unit 230 and a person behavior detection processing unit 231.
 人物・物体関係検知処理部230は、ノードデータベース40から読み込んだ各要素のノードデータに基づき、映像内に映り込んだ人物と物体の関係を検知する。ここでは、例えば事前に学習済みの人物・物体関係検知モデルを用いて、人物が荷物等の物体に対して行う「運ぶ」、「開ける」、「置き去り」等の行動を、両者の関係として検知する。 The person / object relationship detection processing unit 230 detects the relationship between the person and the object reflected in the image based on the node data of each element read from the node database 40. Here, for example, using a person / object relationship detection model that has been learned in advance, actions such as "carrying", "opening", and "leaving" that a person performs on an object such as luggage are detected as the relationship between the two. do.
 人物行動検知処理部231は、ノードデータベース40から読み込んだ各要素のノードデータに基づき、映像内に映り込んだ人物間のインタラクション行動を検知する。ここでは、例えば事前に学習済みの人物インタラクション行動検知モデルを用いて、複数の人物が一緒に行う「会話」、「受け渡し」などの行動を、各人物間のインタラクション行動として検知する。 The person behavior detection processing unit 231 detects the interaction behavior between people reflected in the video based on the node data of each element read from the node database 40. Here, for example, using a person interaction behavior detection model that has been learned in advance, actions such as "conversation" and "delivery" performed by a plurality of people together are detected as interaction actions between each person.
 関係性検知処理部23では、以上説明した各ブロックの処理により、エンティティ検知処理部21により監視対象要素として検知された人物や物体について、ある人物が他の人物または物体に対して行う行動が検知され、その行動が相互の関係性として取得される。この情報は、各要素間の関係性を表すエッジデータとして、エッジデータベース50に格納される。 The relationship detection processing unit 23 detects the action performed by one person on another person or object with respect to the person or object detected as the monitored element by the entity detection processing unit 21 by the processing of each block described above. And the behavior is acquired as a mutual relationship. This information is stored in the edge database 50 as edge data representing the relationship between each element.
 図3は、グラフデータ生成部20が行う処理の概要を示す図である。図3に示すように、グラフデータ生成部20は、エンティティ検知処理部21が行うエンティティ検知処理により、カメラ動画像入力部10により撮影された映像から、人物2と人物2が運んでいる物体3を検知し、これらを映像内で追跡する。また、関係性検知処理部23が行う関係性検知処理により、人物2と物体3との関係性を検知する。そして、これらの処理結果に基づき、一定の時刻区間Δtごとに、複数のノードとエッジで構成されるグラフデータを生成する。このグラフデータでは、例えば人物2はノードP1、物体3はノードO1でそれぞれ表され、これらのノードに対して、その特徴を示す属性情報がそれぞれ設定される。また、ノードP1とノードO1の間に、人物2と物体3の関係性を示す「運ぶ」というエッジが設定される。こうして生成されたグラフデータの情報が、グラフデータベース30に格納される。 FIG. 3 is a diagram showing an outline of the processing performed by the graph data generation unit 20. As shown in FIG. 3, the graph data generation unit 20 is the object 3 carried by the person 2 and the person 2 from the image taken by the camera moving image input unit 10 by the entity detection process performed by the entity detection processing unit 21. Are detected and these are tracked in the video. Further, the relationship detection process performed by the relationship detection processing unit 23 detects the relationship between the person 2 and the object 3. Then, based on these processing results, graph data composed of a plurality of nodes and edges is generated for each fixed time interval Δt. In this graph data, for example, the person 2 is represented by the node P1 and the object 3 is represented by the node O1, and attribute information indicating the characteristics thereof is set for each of these nodes. Further, an edge of "carrying" indicating the relationship between the person 2 and the object 3 is set between the node P1 and the node O1. The information of the graph data thus generated is stored in the graph database 30.
 図4は、グラフデータベース30のデータ構造例を示す図である。図4に示すように、グラフデータベース30は、例えば列301~304を含むデータテーブルにより表現される。列301には、データテーブルの各行に対して設定される一連の整理番号が格納される。列302には、各グラフデータに固有のグラフIDが格納される。列303、304には、各グラフデータに対応する時間範囲の開始時刻と終了時刻がそれぞれ格納される。なお、開始時刻と終了時刻は、各グラフデータの生成に使用された映像において記録された撮影開始時刻と撮影終了時刻からそれぞれ計算され、その差は前述の時刻区間Δtに等しい。これらの情報が各グラフデータについて行ごとに格納されることで、グラフデータベース30が構成される。 FIG. 4 is a diagram showing an example of the data structure of the graph database 30. As shown in FIG. 4, the graph database 30 is represented by, for example, a data table including columns 301 to 304. Column 301 stores a series of reference numbers set for each row of the data table. A graph ID unique to each graph data is stored in the column 302. The start time and end time of the time range corresponding to each graph data are stored in columns 303 and 304, respectively. The start time and end time are calculated from the shooting start time and shooting end time recorded in the video used to generate each graph data, and the difference is equal to the above-mentioned time interval Δt. The graph database 30 is configured by storing this information row by row for each graph data.
 図5は、ノードデータベース40のデータ構造例を示す図である。ノードデータベース40は、図5(a)に示すノード属性テーブル41と、図5(b)に示す追跡情報テーブル42と、図5(c)に示すフレーム情報テーブル43とによって構成される。 FIG. 5 is a diagram showing an example of a data structure of the node database 40. The node database 40 is composed of a node attribute table 41 shown in FIG. 5A, a tracking information table 42 shown in FIG. 5B, and a frame information table 43 shown in FIG. 5C.
 図5(a)に示すように、ノード属性テーブル41は、例えば列411~414を含むデータテーブルにより表現される。列411には、データテーブルの各行に対して設定される一連の整理番号が格納される。列412には、各ノードが属するグラフデータのグラフIDが格納される。このグラフIDの値は、図4のデータテーブルにおいて列302に格納されたグラフIDの値と対応付けられており、これによって各ノードとグラフデータとの紐付けが行われる。列413には、各ノードに固有のノードIDが格納される。列414には、各ノードが表す要素に対して取得された属性情報が格納される。これらの情報が各ノードについて行ごとに格納されることで、ノード属性テーブル41が構成される。 As shown in FIG. 5A, the node attribute table 41 is represented by, for example, a data table including columns 411 to 414. Column 411 stores a series of reference numbers set for each row of the data table. The graph ID of the graph data to which each node belongs is stored in the column 412. The value of this graph ID is associated with the value of the graph ID stored in the column 302 in the data table of FIG. 4, whereby each node is associated with the graph data. Column 413 stores a node ID unique to each node. Column 414 stores the attribute information acquired for the element represented by each node. The node attribute table 41 is configured by storing this information row by row for each node.
 図5(b)に示すように、追跡情報テーブル42は、例えば列421~424を含むデータテーブルにより表現される。列421には、データテーブルの各行に対して設定される一連の整理番号が格納される。列422には、各追跡情報が追跡対象としたノードのノードIDが格納される。このノードIDの値は、図5(a)のデータテーブルにおいて列413に格納されたノードIDの値と対応付けられており、これによって各追跡情報とノードとの紐付けが行われる。列423には、各追跡情報に固有のトラックIDが格納される。列424には、当該ノードが表す要素が映像内で映り込んでいる各フレームのフレームIDのリストが格納される。これらの情報が各追跡情報について行ごとに格納されることで、追跡情報テーブル42が構成される。 As shown in FIG. 5B, the tracking information table 42 is represented by, for example, a data table including columns 421 to 424. Column 421 stores a series of reference numbers set for each row of the data table. In column 422, the node ID of the node targeted by each tracking information is stored. The value of this node ID is associated with the value of the node ID stored in column 413 in the data table of FIG. 5A, whereby each tracking information is associated with the node. Column 423 stores a track ID unique to each tracking information. Column 424 stores a list of frame IDs of each frame in which the element represented by the node is reflected in the video. The tracking information table 42 is configured by storing this information row by row for each tracking information.
 図5(c)に示すように、フレーム情報テーブル43は、例えば列431~434を含むデータテーブルにより表現される。列431には、データテーブルの各行に対して設定される一連の整理番号が格納される。列432には、各フレーム情報が属する追跡情報のトラックIDが格納される。このトラックIDの値は、図5(b)のデータテーブルにおいて列423に格納されたトラックIDの値と対応付けられており、これによって各フレーム情報と追跡情報との紐付けが行われる。列433には、各フレーム情報に固有のフレームIDが格納される。列434には、当該フレーム情報が表すフレーム内での各要素の位置と、各要素の種類(人物、物体など)とを表す情報が格納される。これらの情報が各フレーム情報について行ごとに格納されることで、フレーム情報テーブル43が構成される。 As shown in FIG. 5C, the frame information table 43 is represented by, for example, a data table including columns 431 to 434. Column 431 stores a series of reference numbers set for each row of the data table. The track ID of the tracking information to which each frame information belongs is stored in the column 432. The value of the track ID is associated with the value of the track ID stored in the column 423 in the data table of FIG. 5B, whereby each frame information and the tracking information are associated with each other. A frame ID unique to each frame information is stored in the column 433. The column 434 stores information indicating the position of each element in the frame represented by the frame information and the type of each element (person, object, etc.). The frame information table 43 is configured by storing this information row by row for each frame information.
 図6は、エッジデータベース50のデータ構造例を示す図である。図6に示すように、エッジデータベース50は、例えば列501~506を含むデータテーブルにより表現される。列501には、データテーブルの各行に対して設定される一連の整理番号が格納される。列502には、各エッジが属するグラフデータのグラフIDが格納される。このグラフIDの値は、図4のデータテーブルにおいて列302に格納されたグラフIDの値と対応付けられており、これによって各エッジとグラフデータとの紐付けが行われる。列503、504には、各エッジの始点と終点に位置するノードのノードIDがそれぞれ格納される。これらのノードIDの値は、図5(a)のデータテーブルにおいて列413に格納されたノードIDの値とそれぞれ対応付けられており、これによって各エッジがどのノード間の関係性を表しているのかが特定される。列505には、各エッジに固有のエッジIDが格納される。列506には、当該エッジが表す要素間の関係性を表すエッジ情報として、始点ノードに対応する人物が終点ノードに対応する他の人物または物体に対して行う行動の内容が格納される。これらの情報が各エッジについて行ごとに格納されることで、エッジデータベース50が構成される。 FIG. 6 is a diagram showing an example of the data structure of the edge database 50. As shown in FIG. 6, the edge database 50 is represented by, for example, a data table containing columns 501-506. Column 501 stores a series of reference numbers set for each row of the data table. The graph ID of the graph data to which each edge belongs is stored in the column 502. The value of this graph ID is associated with the value of the graph ID stored in the column 302 in the data table of FIG. 4, whereby each edge is associated with the graph data. The nodes 503 and 504 store the node IDs of the nodes located at the start point and the end point of each edge, respectively. The values of these node IDs are associated with the values of the node IDs stored in column 413 in the data table of FIG. 5A, whereby each edge represents the relationship between which nodes. Is specified. Column 505 stores an edge ID unique to each edge. Column 506 stores, as edge information representing the relationship between the elements represented by the edge, the content of the action performed by the person corresponding to the start point node on another person or object corresponding to the end point node. The edge database 50 is configured by storing this information row by row for each edge.
 図7は、グラフデータ可視化編集部60の説明図である。グラフデータ可視化編集部60は、例えば図7に示すグラフデータ編集画面61を不図示のディスプレイに表示してユーザに提示する。このグラフデータ編集画面61において、ユーザは所定の操作を行うことにより、グラフデータを任意に編集することができる。例えば、グラフデータ編集画面61には、グラフデータ生成部20により生成されたグラフデータ610が可視化して表示される。このグラフデータ610において、ユーザは任意のノードまたはエッジを画面上で選択することにより、ノードの詳細情報を示すノード情報ボックス611、612や、エッジの詳細情報を示すエッジ情報ボックス613を表示させることができる。これらの情報ボックス611~613には、各ノードの属性情報が表示されている。ユーザは、情報ボックス611~613内で任意の属性情報を選択することにより、下線部に示した各属性情報の内容を編集することができる。 FIG. 7 is an explanatory diagram of the graph data visualization editing unit 60. The graph data visualization editing unit 60 displays, for example, the graph data editing screen 61 shown in FIG. 7 on a display (not shown) and presents it to the user. On the graph data editing screen 61, the user can arbitrarily edit the graph data by performing a predetermined operation. For example, the graph data 610 generated by the graph data generation unit 20 is visualized and displayed on the graph data editing screen 61. In the graph data 610, the user selects an arbitrary node or edge on the screen to display the node information boxes 611 and 612 showing the detailed information of the node and the edge information box 613 showing the detailed information of the edge. Can be done. The attribute information of each node is displayed in these information boxes 611 to 613. The user can edit the content of each attribute information shown in the underlined portion by selecting arbitrary attribute information in the information boxes 611 to 613.
 また、グラフデータ編集画面61には、グラフデータ610とともに、ノード追加ボタン614およびエッジ追加ボタン615が表示される。ユーザはノード追加ボタン614またはエッジ追加ボタン615を画面上で選択することにより、グラフデータ610に対して任意の位置にノードまたはエッジを追加することができる。さらに、グラフデータ610において任意のノードまたはエッジを選択し、所定の操作(例えばマウスのドラッグや右クリック)を行うことで、そのノードまたはエッジを移動または削除することもできる。 Further, on the graph data editing screen 61, a node addition button 614 and an edge addition button 615 are displayed together with the graph data 610. The user can add a node or an edge to the graph data 610 at an arbitrary position by selecting the node addition button 614 or the edge addition button 615 on the screen. Further, by selecting an arbitrary node or edge in the graph data 610 and performing a predetermined operation (for example, dragging or right-clicking the mouse), the node or edge can be moved or deleted.
 グラフデータ可視化編集部60は、以上説明したようなユーザの操作により、生成されたグラフデータの内容を適宜編集することができる。そして、編集後のグラフデータを反映してグラフデータベース30を更新する。 The graph data visualization editing unit 60 can appropriately edit the contents of the generated graph data by the user's operation as described above. Then, the graph database 30 is updated to reflect the edited graph data.
 図8は、ノード特徴量抽出部70の構成を示すブロック図である。図7に示すように、ノード特徴量抽出部70は、最大エントロピーフレームサンプリング処理部71、人物・物体領域画像取得部72、画像特徴量計算部73、属性情報取得部74、属性情報特徴量計算部75、特徴量結合処理部76、属性重み計算アテンション機構77およびノード特徴量計算部78を備えて構成される。 FIG. 8 is a block diagram showing the configuration of the node feature amount extraction unit 70. As shown in FIG. 7, the node feature amount extraction unit 70 includes a maximum entropy frame sampling processing unit 71, a person / object area image acquisition unit 72, an image feature amount calculation unit 73, an attribute information acquisition unit 74, and an attribute information feature amount calculation. It is configured to include a unit 75, a feature amount coupling processing unit 76, an attribute weight calculation attention mechanism 77, and a node feature amount calculation unit 78.
 最大エントロピーフレームサンプリング処理部71は、ノードデータベース40から各ノードのノードデータを読み出し、各ノードについて映像内で最大エントロピーを有するフレームをサンプリングする。 The maximum entropy frame sampling processing unit 71 reads the node data of each node from the node database 40, and samples the frame having the maximum entropy in the video for each node.
 人物・物体領域画像取得部72は、最大エントロピーフレームサンプリング処理部71によりサンプリングされたフレームから、各ノードが表す要素に対応する人物や物体の領域画像を取得する。 The person / object area image acquisition unit 72 acquires the area image of the person or object corresponding to the element represented by each node from the frame sampled by the maximum entropy frame sampling processing unit 71.
 画像特徴量計算部73は、人物・物体領域画像取得部72により取得された各人物または各物体の領域画像から、各ノードが表す要素ごとの画像特徴量を計算する。ここでは例えば、予め大規模の画像データセット(例えばMS COCOなど)を用いて学習済みの物体分類用のDNN(Deep Neural Network)を使用し、このDNNに各要素の領域画像を入力したときの中間層からの出力を抽出することで、画像特徴量を計算する。なお、各要素の領域画像に対して画像特徴量が計算できれば、他の方法を用いてもよい。 The image feature amount calculation unit 73 calculates the image feature amount for each element represented by each node from the area image of each person or each object acquired by the person / object area image acquisition unit 72. Here, for example, when a DNN (Deep Neural Network) for object classification that has been learned in advance using a large-scale image data set (for example, MSCOCO) is used and a region image of each element is input to this DNN. The image feature amount is calculated by extracting the output from the intermediate layer. If the image feature amount can be calculated for the area image of each element, another method may be used.
 属性情報取得部74は、ノードデータベース40から各ノードのノード情報を読み出し、各ノードの属性情報を取得する。 The attribute information acquisition unit 74 reads the node information of each node from the node database 40 and acquires the attribute information of each node.
 属性情報特徴量計算部75は、属性情報取得部74により取得された属性情報から、各ノードが表す要素ごとの属性情報の特徴量を計算する。ここでは例えば、属性情報を構成するテキストデータに対して、所定の言語処理アルゴリズム(例えばword2Vecなど)を用いることにより、属性情報が表す各要素の属性項目(性別、年齢、服装、マスク着用の有無、大きさ、色、滞在時間など)ごとに特徴量を計算する。なお、各要素の属性情報に対して属性情報特徴量が計算できれば、他の方法を用いてもよい。 The attribute information feature amount calculation unit 75 calculates the feature amount of the attribute information for each element represented by each node from the attribute information acquired by the attribute information acquisition unit 74. Here, for example, by using a predetermined language processing algorithm (for example, word2Vec) for the text data constituting the attribute information, the attribute items (gender, age, clothes, presence / absence of wearing a mask) of each element represented by the attribute information are used. , Size, color, staying time, etc.) Calculate the feature amount. If the attribute information feature amount can be calculated for the attribute information of each element, another method may be used.
 特徴量結合処理部76は、画像特徴量計算部73により計算された画像特徴量と、属性情報特徴量計算部75により計算された属性情報の特徴量とを結合する結合処理を行う。ここでは、例えば画像特徴量が表す人物または物体全体での特徴に対する特徴量と、属性情報が表す人物または物体の属性項目ごとの特徴量とをベクトル成分として、これらの各項目の特徴量に応じた特徴量ベクトルを要素ごとに作成する。 The feature amount combination processing unit 76 performs a combination process of combining the image feature amount calculated by the image feature amount calculation unit 73 and the feature amount of the attribute information calculated by the attribute information feature amount calculation unit 75. Here, for example, the feature amount for the feature of the whole person or object represented by the image feature amount and the feature amount for each attribute item of the person or object represented by the attribute information are set as vector components according to the feature amount of each of these items. Create a feature vector for each element.
 属性重み計算アテンション機構77は、特徴量結合処理部76により結合された特徴量に対して、その特徴量の項目ごとの重みを取得する。ここでは、例えば特徴量ベクトルの各ベクトル成分に対して、事前に学習された重みをそれぞれ取得する。属性重み計算アテンション機構77が取得した重みの情報は、異常検知部130により算出される脅威予兆度に対するノード特徴量の項目ごとの寄与度を表す要素寄与度として、要素寄与度保存部160に格納される。 The attribute weight calculation attention mechanism 77 acquires the weight for each item of the feature amount for the feature amount combined by the feature amount combination processing unit 76. Here, for example, the weights learned in advance are acquired for each vector component of the feature amount vector. Attribute weight calculation The weight information acquired by the attention mechanism 77 is stored in the element contribution storage unit 160 as an element contribution indicating the contribution of each node feature amount to the threat sign degree calculated by the anomaly detection unit 130. Will be done.
 ノード特徴量計算部78は、特徴量結合処理部76により結合された特徴量に対して、属性重み計算アテンション機構77により取得された重みを乗算することで、重み付け処理を行い、ノード特徴量を計算する。すなわち、特徴量ベクトルの各ベクトル成分に対して、属性重み計算アテンション機構77により設定された重みをそれぞれ乗算した値を合計することで、ノード特徴量を計算する。 The node feature amount calculation unit 78 performs weighting processing by multiplying the feature amount combined by the feature amount combination processing unit 76 by the weight acquired by the attribute weight calculation attention mechanism 77, and calculates the node feature amount. calculate. That is, the node feature amount is calculated by summing the values obtained by multiplying each vector component of the feature amount vector by the weight set by the attribute weight calculation attention mechanism 77.
 ノード特徴量抽出部70では、以上説明した各ブロックの処理により、時刻区間Δt間隔で設定された時間範囲ごとに生成された各グラフデータについて、要素ごとの属性の特徴量を表すノード特徴量が抽出される。抽出されたノード特徴量の情報は、ノード特徴量蓄積部90に格納される。 In the node feature amount extraction unit 70, the node feature amount representing the attribute feature amount for each element is generated for each graph data generated for each time range set in the time interval Δt interval by the processing of each block described above. Be extracted. The extracted node feature amount information is stored in the node feature amount storage unit 90.
 図9は、ノード特徴量抽出部70が行う処理の概要を示す図である。図9に示すように、ノード特徴量抽出部70は、各グラフデータに対応する映像内で人物2のエントロピーが最大のフレームに対して、画像特徴量計算部73により画像特徴量を計算するとともに、属性情報特徴量計算部75により、人物2に対応するノードP1の属性情報の各属性項目に対して特徴量を計算することで、「全身特徴量」、「マスク」、「肌色」、「滞在時間」などの各項目に対するノードP1の特徴量を求める。そして、属性重み計算アテンション機構77により取得された重みを用いて、これらの各項目に対する重み付け演算をノード特徴量計算部78により行うことで、ノードP1の特徴量を抽出する。同様の計算を他の各ノードに対して行うことで、グラフデータの各ノードの特徴量が求められる。なお、属性重み計算アテンション機構77により取得された重みは、要素寄与度として要素寄与度保存部160に格納される。 FIG. 9 is a diagram showing an outline of the processing performed by the node feature amount extraction unit 70. As shown in FIG. 9, the node feature amount extraction unit 70 calculates the image feature amount by the image feature amount calculation unit 73 for the frame having the maximum entropy of the person 2 in the video corresponding to each graph data. , The attribute information feature amount calculation unit 75 calculates the feature amount for each attribute item of the attribute information of the node P1 corresponding to the person 2, so that the "whole body feature amount", "mask", "skin color", and "skin color" are calculated. The feature amount of the node P1 is obtained for each item such as "stay time". Then, the feature amount of the node P1 is extracted by performing the weighting operation for each of these items by the node feature amount calculation unit 78 using the weight acquired by the attribute weight calculation attention mechanism 77. By performing the same calculation for each of the other nodes, the feature amount of each node of the graph data can be obtained. The weight acquired by the attribute weight calculation attention mechanism 77 is stored in the element contribution storage unit 160 as the element contribution.
 図10は、エッジ特徴量抽出部80の構成を示すブロック図である。図10に示すように、エッジ特徴量抽出部80は、エッジ情報取得部81、エッジ特徴量計算部82、エッジ重み計算アテンション機構83および重み付け計算部84を備えて構成される。 FIG. 10 is a block diagram showing the configuration of the edge feature amount extraction unit 80. As shown in FIG. 10, the edge feature amount extraction unit 80 includes an edge information acquisition unit 81, an edge feature amount calculation unit 82, an edge weight calculation attention mechanism 83, and a weighting calculation unit 84.
 エッジ情報取得部81は、エッジデータベース50から各エッジのエッジ情報を読み出して取得する。 The edge information acquisition unit 81 reads and acquires the edge information of each edge from the edge database 50.
 エッジ特徴量計算部82は、エッジ情報取得部81により取得されたエッジ情報から、各エッジが表す要素間の関係性の特徴量であるエッジ特徴量を計算する。ここでは例えば、エッジ情報として設定された行動内容を表す「受け渡し」、「会話」などのテキストデータに対して、所定の言語処理アルゴリズム(例えばword2Vecなど)を用いることにより、エッジ特徴量を計算する。 The edge feature amount calculation unit 82 calculates the edge feature amount, which is the feature amount of the relationship between the elements represented by each edge, from the edge information acquired by the edge information acquisition unit 81. Here, for example, the edge feature amount is calculated by using a predetermined language processing algorithm (for example, word2Vec) for text data such as "passing" and "conversation" representing the action contents set as edge information. ..
 エッジ重み計算アテンション機構83は、エッジ特徴量計算部82により計算されたエッジ特徴量に対する重みを取得する。ここでは、例えばエッジ特徴量に対して、事前に学習された重みを取得する。エッジ重み計算アテンション機構83が取得した重みの情報は、異常検知部130により算出される脅威予兆度に対するエッジ特徴量の寄与度を表す要素寄与度として、要素寄与度保存部160に格納される。 The edge weight calculation attention mechanism 83 acquires the weight for the edge feature amount calculated by the edge feature amount calculation unit 82. Here, for example, the weight learned in advance is acquired for the edge feature amount. The weight information acquired by the edge weight calculation attention mechanism 83 is stored in the element contribution storage unit 160 as an element contribution representing the contribution of the edge feature amount to the threat sign degree calculated by the abnormality detection unit 130.
 重み付け計算部84は、エッジ特徴量計算部82により計算されたエッジ特徴量に対して、エッジ重み計算アテンション機構83により取得された重みを乗算することで、重み付け処理を行い、重み付け後のエッジ特徴量を計算する。 The weighting calculation unit 84 performs weighting processing by multiplying the edge feature amount calculated by the edge feature amount calculation unit 82 by the weight acquired by the edge weight calculation attention mechanism 83, and the weighted edge feature. Calculate the amount.
 エッジ特徴量抽出部80では、以上説明した各ブロックの処理により、時刻区間Δt間隔で設定された時間範囲ごとに生成された各グラフデータについて、要素間の関係性の特徴量を表すエッジ特徴量が抽出される。抽出されたエッジ特徴量の情報は、エッジ特徴量蓄積部100に格納される。 In the edge feature amount extraction unit 80, the edge feature amount representing the feature amount of the relationship between the elements for each graph data generated for each time range set in the time interval Δt interval by the processing of each block described above. Is extracted. The extracted edge feature amount information is stored in the edge feature amount storage unit 100.
 図11は、時空間特徴量算出部110の構成を示すブロック図である。図11に示すように、時空間特徴量算出部110は、複数の残余畳み込み演算ブロック111と、ノード特徴量更新部112とを備えて構成される。各残余畳み込み演算ブロック111は所定の段数にそれぞれ対応しており、前段の残余畳み込み演算ブロック111の演算結果を受けて畳み込み演算を実行し、後段の残余畳み込み演算ブロック111に入力する。なお、最前段の残余畳み込み演算ブロック111には、ノード特徴量蓄積部90とエッジ特徴量蓄積部100からそれぞれ読み込まれたノード特徴量およびエッジ特徴量が入力され、最終段の残余畳み込み演算ブロック111の演算結果はノード特徴量更新部112に入力される。これにより、GNN(Graph Neural Network)を用いた時空間特徴量の算出を実現している。 FIG. 11 is a block diagram showing the configuration of the spatiotemporal feature amount calculation unit 110. As shown in FIG. 11, the spatiotemporal feature amount calculation unit 110 includes a plurality of residual convolution calculation blocks 111 and a node feature amount update unit 112. Each residual convolution calculation block 111 corresponds to a predetermined number of stages, receives the calculation result of the residual convolution calculation block 111 in the previous stage, executes the convolution operation, and inputs the convolution operation to the residual convolution calculation block 111 in the subsequent stage. The node feature amount and the edge feature amount read from the node feature amount storage unit 90 and the edge feature amount storage unit 100 are input to the residual convolution calculation block 111 in the front stage, and the residual convolution calculation block 111 in the final stage is input. The calculation result of is input to the node feature amount update unit 112. As a result, the calculation of spatiotemporal features using GNN (Graph Neural Network) is realized.
 時空間特徴量算出部110は、複数の残余畳み込み演算ブロック111のそれぞれにおいて前述のような畳み込み操作を行う。この畳み込み操作を実現するため、各残余畳み込み演算ブロック111は、2つの空間畳み込み演算処理部1110と、1つの時間畳み込み演算処理部1111とを備えて構成される。 The spatiotemporal feature amount calculation unit 110 performs the above-mentioned convolution operation in each of the plurality of residual convolution calculation blocks 111. In order to realize this convolution operation, each residual convolution calculation block 111 includes two space convolution calculation processing units 1110 and one time convolution calculation processing unit 1111.
 空間畳み込み演算処理部1110は、空間方向の畳み込み演算として、グラフデータにおいて各ノードに隣接する隣接ノードの特徴量と、各ノードと隣接ノードの間に設定されたエッジの特徴量との外積を計算し、この外積に対して、D×Dサイズの重み行列を用いた重み付け演算を行う。ここで、重み行列の次数Dの値は、各ノードの特徴量の長さとして定義される。これにより、学習可能な重み付き線形変換を用いて、学習の多様性を保証するようにしている。また、グラフデータを構成するノードとエッジの数による制約を受けずに重み行列の設計が可能となるため、最適な重み行列を用いて重み付け演算を実施することができる。 The space convolution calculation processing unit 1110 calculates the outer product of the feature amount of the adjacent node adjacent to each node in the graph data and the feature amount of the edge set between each node and the adjacent node as the convolution calculation in the spatial direction. Then, a weighting operation using a weight matrix of D × D size is performed on this outer product. Here, the value of the degree D of the weight matrix is defined as the length of the feature amount of each node. This ensures a variety of learning using a learnable weighted linear transformation. Further, since the weight matrix can be designed without being restricted by the number of nodes and edges constituting the graph data, the weighting operation can be performed using the optimum weight matrix.
 残余畳み込み演算ブロック111では、グラフデータを構成する各ノードに対して、空間畳み込み演算処理部1110による重み付け演算を2回行う。これにより、空間方向の畳み込み演算を実現する。 In the residual convolution calculation block 111, the spatial convolution calculation processing unit 1110 performs a weighting calculation twice for each node constituting the graph data. As a result, the convolution operation in the spatial direction is realized.
 時間畳み込み演算処理部1111は、2つの空間畳み込み演算処理部1110により空間方向の畳み込み演算を実施された各ノードの特徴量に対して、時間方向の畳み込み演算を行う。ここでは、各ノードに時間方向で隣接するノード、すなわち隣接する時間範囲の映像に対して生成されたグラフデータにおいて当該ノードと同じ人物または物体を表すノードの特徴量と、その隣接ノードに対して設定されたエッジの特徴量との外積を計算し、この外積に対して、空間畳み込み演算処理部1110と同様の重み付け演算を行う。これにより、時間方向の畳み込み演算を実現する。 The time-convolution calculation processing unit 1111 performs a time-direction convolution calculation on the feature amount of each node for which the spatial-direction convolution calculation is performed by the two space convolution calculation processing units 1110. Here, for the node adjacent to each node in the time direction, that is, the feature amount of the node representing the same person or object as the node in the graph data generated for the video in the adjacent time range, and the adjacent node. The outer product with the set feature amount of the edge is calculated, and the weighting operation similar to that of the space convolution calculation processing unit 1110 is performed on this outer product. This realizes a convolution operation in the time direction.
 以上説明した空間方向および時間方向の畳み込み演算によって計算された時空間特徴量と、残余畳み込み演算ブロック111に入力されたノード特徴量とが加算されることで、残余畳み込み演算ブロック111の演算結果が求められる。こうした演算を行うことにより、空間方向と時間方向でそれぞれ隣接する隣接ノードおよび隣接ノード間のエッジの両方の特徴量を、同時に各ノードの特徴量に加えていく畳み込み操作が可能となる。 By adding the spatiotemporal feature amount calculated by the spatial and temporal convolution operations described above and the node feature amount input to the residual convolution operation block 111, the operation result of the residual convolution operation block 111 is obtained. Desired. By performing such an operation, it is possible to perform a convolution operation in which the features of both the adjacent nodes adjacent to each other in the spatial direction and the temporal direction and the edges between the adjacent nodes are added to the features of each node at the same time.
 ノード特徴量更新部112は、最終段の残余畳み込み演算ブロック111から出力された演算結果を用いて、ノード特徴量蓄積部90に蓄積された各ノードの特徴量を更新する。これにより、グラフデータを構成するノードごとに算出された時空間特徴量を各ノードの特徴量に反映させる。 The node feature amount update unit 112 updates the feature amount of each node accumulated in the node feature amount storage unit 90 by using the calculation result output from the residual convolution calculation block 111 in the final stage. As a result, the spatiotemporal features calculated for each node constituting the graph data are reflected in the features of each node.
 時空間特徴量算出部110では、以上説明した各ブロックの処理により、GNNを用いて各グラフデータの時空間特徴量を算出し、ノード特徴量に反映してノード特徴量を更新することができる。なお、時空間特徴量算出部110におけるGNNの学習では、任意の層の入力を参照した残差関数を学習することが好ましい、このようにすれば、学習時の層が深くても勾配爆発や勾配消失の問題を防ぐことができる。したがって、より正確な時空間情報を反映したノード特徴量の算出が可能となる。 The spatiotemporal feature amount calculation unit 110 can calculate the spatiotemporal feature amount of each graph data using GNN by the processing of each block described above, and can update the node feature amount by reflecting it on the node feature amount. .. In the GNN learning in the spatiotemporal feature calculation unit 110, it is preferable to learn the residual function with reference to the input of any layer. In this way, even if the layer at the time of learning is deep, a gradient explosion or a gradient explosion occurs. The problem of vanishing gradient can be prevented. Therefore, it is possible to calculate the node features that reflect more accurate spatiotemporal information.
 図12は、空間畳み込み演算処理部1110における演算処理を表す数式の一例を示す図である。空間畳み込み演算処理部1110は、例えば図12に示すような行列演算式をそれぞれ計算することにより、空間畳み込み演算を行う。そして、得られたN×D×P(Nはノード数、Dはノード特徴量の長さ、Pは行列演算のチャンネル数=エッジ特徴量の長さ)のテンソルに対して連結または平均プーリングを行い、これをGNNの層数に応じて設けられた残余畳み込み演算ブロック111の個数分だけ繰り返し実施することで、空間畳み込み後の特徴量を算出し、さらに時間畳み込み演算を行い、時空間特徴量を算出し、ノード特徴量に反映させる。 FIG. 12 is a diagram showing an example of a mathematical formula representing arithmetic processing in the space convolution arithmetic processing unit 1110. The space convolution calculation processing unit 1110 performs a space convolution calculation by, for example, calculating each of the matrix calculation formulas as shown in FIG. 12. Then, concatenation or average pooling is performed on the obtained tensor of N × D × P (N is the number of nodes, D is the length of the node features, and P is the number of channels of the matrix operation = the length of the edge features). By repeating this for the number of residual convolution calculation blocks 111 provided according to the number of layers of GNN, the feature amount after spatial convolution is calculated, and the time convolution calculation is further performed to perform the spatiotemporal feature amount. Is calculated and reflected in the node features.
 ここで、空間畳み込み演算処理部1110により実施される畳み込み演算と、時間畳み込み演算処理部1111により実施される畳み込み演算とは、それぞれ以下の数式(1)、(2)で表される。 Here, the convolution operation performed by the space convolution calculation processing unit 1110 and the convolution calculation performed by the time convolution calculation processing unit 1111 are expressed by the following mathematical formulas (1) and (2), respectively.
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
 数式(1)において、Oは連結または平均プーリング、φは非直線活性化関数、lは空間畳み込み演算処理部1110が対応するGNNの層番号をそれぞれ表している。また、数式(2)において、kは時間畳み込み演算処理部1111が対応するGNNの層番号を表している。 In the formula (1), O represents the concatenation or average pooling, φ represents the non-linear activation function, and l represents the layer number of the GNN corresponding to the space convolution arithmetic processing unit 1110. Further, in the mathematical formula (2), k represents the layer number of the GNN corresponding to the time convolution calculation processing unit 1111.
 また、図12および数式(1)、(2)において、HN×Dは空間ノード特徴量の行列を表し、Nはグラフデータ内のノード数、Dはノード特徴量の長さ(次数)をそれぞれ表している。M L×Dはi番目のノードに対する時間ノード特徴量の行列を表し、Lは時間の長さを表している。EN×N×Pはエッジ特徴量の行列を表し、Eijはi番目のノードとj番目のノードを繋ぐエッジの特徴量(次数P)を表している。ここで、i番目のノードとj番目のノードを繋ぐエッジが存在しない場合はEij=0である。 Further, in FIG. 12 and mathematical formulas (1) and (2), H N × D represents a matrix of spatial node features, N represents the number of nodes in the graph data, and D represents the length (order) of the node features. Each is represented. M i L × D represents a matrix of time node features for the i-th node, and L represents the length of time. EN × N × P represents a matrix of edge features, and Eij represents an edge feature (order P) connecting the i-th node and the j-th node. Here, if there is no edge connecting the i-th node and the j-th node, Eij = 0.
 また、図12および数式(1)、(2)において、F 1×Dはi番目のノードに対する時間ノード特徴量の行列を表している。Fijはj番目のグラフデータにおけるj番目のノードの存否を表している。ここで、j番目のグラフデータにj番目のノードが存在しない場合はFij=0、存在する場合はFij=1である。 Further, in FIG. 12 and mathematical formulas (1) and (2), Fi 1 × D represents a matrix of time node features for the i -th node. Fij represents the existence or nonexistence of the jth node in the jth graph data. Here, if the j-th node does not exist in the j-th graph data, F ij = 0, and if it exists, F ij = 1.
 さらに、図12および数式(1)、(2)において、Q1×Lは時間方向のノード間の関係性に対する重み付けのための畳み込みカーネルを表し、W は空間方向のノード特徴量に関するD×Dサイズの重み付け行列を表し、W は時間方向のノード特徴量に関するD×Dサイズの重み付け行列を表している。 Further, in FIG. 12 and the equations (1) and (2), Q1 × L represents a convolutional kernel for weighting the relationship between the nodes in the time direction, and WS l is D regarding the node features in the spatial direction. It represents a weighting matrix of × D size, and WTk represents a weighting matrix of D × D size regarding the node features in the time direction.
 図13は、時空間特徴量算出部110が行う処理の概要を示す図である。図13において、点線は空間畳み込み演算処理部1110による空間畳み込み演算を表し、破線は時間畳み込み演算処理部1111による時間畳み込み演算を表している。図13に示すように、例えばt番目のグラフデータにおけるノード3には、隣接するノード1、4の特徴量と、これらの隣接ノードとの間に設定されたエッジの特徴量とに応じた空間特徴量が、空間畳み込み演算によって加えられる。また、直前のt-1番目のグラフデータにおけるノード3の特徴量と、直後のt+1番目のグラフデータにおけるノード3の特徴量とに応じた時間特徴量が、時間畳み込み演算によって加えられる。これにより、ノード3に対するt番目のグラフデータの時空間特徴量が算出されてノード3の特徴量に反映される。 FIG. 13 is a diagram showing an outline of the processing performed by the spatiotemporal feature amount calculation unit 110. In FIG. 13, the dotted line represents the space convolution calculation by the space convolution calculation processing unit 1110, and the broken line represents the time convolution calculation by the time convolution calculation processing unit 1111. As shown in FIG. 13, for example, node 3 in the t-th graph data has a space corresponding to the feature amounts of adjacent nodes 1 and 4 and the feature amount of the edge set between these adjacent nodes. Features are added by spatial convolution. Further, the time feature amount corresponding to the feature amount of the node 3 in the immediately preceding t-1st graph data and the feature amount of the node 3 in the immediately following t + 1st graph data is added by the time convolution operation. As a result, the spatiotemporal feature amount of the t-th graph data for the node 3 is calculated and reflected in the feature amount of the node 3.
 図14は、異常検知部130の構成を示すブロック図である。図14に示すように、異常検知部130は、特徴量分布クラスタリング部131、中心点距離計算部132および異常判定部133を備えて構成される。 FIG. 14 is a block diagram showing the configuration of the abnormality detection unit 130. As shown in FIG. 14, the abnormality detection unit 130 includes a feature quantity distribution clustering unit 131, a center point distance calculation unit 132, and an abnormality determination unit 133.
 特徴量分布クラスタリング部131は、ノード特徴量取得部120によりノード特徴量蓄積部90から取得された各ノードの特徴量のクラスタリング処理を行い、ノード特徴量の分布を求める。ここでは、例えば各ノードの特徴量を2次元マップ上にそれぞれプロットすることで、ノード特徴量の分布を求める。 The feature amount distribution clustering unit 131 performs clustering processing of the feature amount of each node acquired from the node feature amount storage unit 90 by the node feature amount acquisition unit 120, and obtains the distribution of the node feature amount. Here, for example, the distribution of the node features is obtained by plotting the features of each node on a two-dimensional map.
 中心点距離計算部132は、特徴量分布クラスタリング部131により求められたノード特徴量の分布において、各ノード特徴量の中心点からの距離を計算する。これにより、時空間特徴量が反映された各ノードの特徴量を互いに比較する。中心点距離計算部132により計算された各ノード特徴量の中心点からの距離は、各ノードに対応する要素の脅威の度合いを示す脅威予兆度として、脅威予兆度保存部140に格納される。 The center point distance calculation unit 132 calculates the distance from the center point of each node feature amount in the distribution of the node feature amount obtained by the feature amount distribution clustering unit 131. As a result, the features of each node reflecting the spatiotemporal features are compared with each other. The distance from the center point of each node feature amount calculated by the center point distance calculation unit 132 is stored in the threat sign degree storage unit 140 as a threat sign degree indicating the degree of threat of the element corresponding to each node.
 異常判定部133は、中心点距離計算部132により計算された距離に基づいて、各ノードの脅威予兆度を判定する。その結果、脅威予兆度が所定値以上のノードが存在する場合は、そのノードに対応する要素を不審者または不審物と判断して、監視対象場所の異常を検知し、ユーザへの報知を行う。ユーザへの報知は、例えば不図示の警報装置を用いて行われる。このとき、監視カメラの映像中で不審者や不審物と判断された要素の位置を強調表示してもよい。異常判定部133による異常検知結果は、脅威予兆度と対応付けて脅威予兆度保存部140に格納される。 The abnormality determination unit 133 determines the threat sign degree of each node based on the distance calculated by the center point distance calculation unit 132. As a result, if there is a node with a threat sign of more than a predetermined value, the element corresponding to that node is determined to be a suspicious person or a suspicious object, an abnormality in the monitored location is detected, and the user is notified. .. Notification to the user is performed using, for example, an alarm device (not shown). At this time, the position of an element determined to be a suspicious person or a suspicious object may be highlighted in the image of the surveillance camera. The abnormality detection result by the abnormality determination unit 133 is stored in the threat sign degree storage unit 140 in association with the threat sign degree.
 異常検知部130では、以上説明した各ブロックの処理により、時空間特徴量算出部110によって算出された時空間特徴量に基づいて、監視対象場所における異常を検知するとともに、要素ごとの時空間特徴量を互いに比較し、その比較結果に基づいて要素ごとの脅威予兆度を求めることができる。 The anomaly detection unit 130 detects an abnormality in the monitored location based on the spatiotemporal feature amount calculated by the spatiotemporal feature amount calculation unit 110 by the processing of each block described above, and also detects the spatiotemporal feature for each element. The quantities can be compared with each other, and the threat predictiveness for each element can be obtained based on the comparison result.
 図15は、異常検知部130が行う処理の概要を示す図である。図15に示すように、異常検知部130は、ノードP3、P6、O2を含むグラフデータの各ノードについて、時空間特徴量が判定されたノード特徴量を2次元マップ上にそれぞれプロットすることでノード特徴量の分布を求める。そして、求められたノード特徴量の分布の中心点を求め、この中心点から各ノード特徴量までの距離を計算することで、各ノードの脅威予兆度を求める。その結果、脅威予兆度が所定値以上であるノードに対応する要素、例えばノード特徴量が分布図上で分布円4の外側にあるノードP6に対応する人物が、不審者または不審物であると判断され、異常が検知される。 FIG. 15 is a diagram showing an outline of the processing performed by the abnormality detection unit 130. As shown in FIG. 15, the abnormality detection unit 130 plots the node features for which the spatiotemporal features have been determined for each node of the graph data including the nodes P3, P6, and O2 on a two-dimensional map. Find the distribution of node features. Then, the central point of the distribution of the obtained node feature amount is obtained, and the distance from this center point to each node feature amount is calculated to obtain the threat sign degree of each node. As a result, it is determined that the element corresponding to the node whose threat sign degree is equal to or higher than the predetermined value, for example, the person corresponding to the node P6 whose node feature amount is outside the distribution circle 4 on the distribution map is a suspicious person or a suspicious object. Judgment is made and an abnormality is detected.
 図16は、判定根拠提示部150の構成を示すブロック図である。図16に示すように、判定根拠提示部150は、根拠確認対象選択部151、サブグラフ抽出処理部152、人物属性脅威寄与度提示部153、物体属性脅威寄与度提示部154、行動履歴寄与度提示部155および言語化サマリ生成部156を備えて構成される。 FIG. 16 is a block diagram showing the configuration of the determination basis presentation unit 150. As shown in FIG. 16, the judgment basis presentation unit 150 includes a basis confirmation target selection unit 151, a subgraph extraction processing unit 152, a person attribute threat contribution presentation unit 153, an object attribute threat contribution presentation unit 154, and an action history contribution presentation. It is configured to include a unit 155 and a verbalization summary generation unit 156.
 根拠確認対象選択部151は、脅威予兆度保存部140に格納された脅威予兆度を取得し、取得した各ノードの脅威予兆度に基づいて、異常検知部130により異常を検知されたノードを含むグラフデータのいずれかの部分を、異常検知の根拠確認の対象として選択する。ここでは、例えば最も脅威予兆度が高いノードに関連する部分を自動的に選択してもよいし、ユーザの操作に応じて任意のノードを指定し、そのノードに関連する部分を選択してもよい。 The basis confirmation target selection unit 151 acquires the threat sign degree stored in the threat sign degree storage unit 140, and includes the node in which the abnormality is detected by the abnormality detection unit 130 based on the acquired threat sign degree of each node. Select any part of the graph data as the target for confirming the basis for abnormality detection. Here, for example, the part related to the node with the highest threat sign may be automatically selected, or an arbitrary node may be specified according to the user's operation and the part related to that node may be selected. good.
 サブグラフ抽出処理部152は、グラフデータベース30に格納されたグラフデータを取得し、取得したグラフデータにおいて根拠確認対象選択部151により選択された部分を、異常検知の根拠確認の対象を示すサブグラフとして抽出する。例えば、最も脅威予兆度が高いノードやユーザに指定されたノードと、そのノードに接続されている各ノードおよび各エッジとを、サブグラフとして抽出する。 The subgraph extraction processing unit 152 acquires the graph data stored in the graph database 30, and extracts the portion selected by the basis confirmation target selection unit 151 in the acquired graph data as a subgraph indicating the basis confirmation target for abnormality detection. do. For example, the node with the highest threat sign or the node specified by the user, each node connected to the node, and each edge are extracted as a subgraph.
 人物属性脅威寄与度提示部153は、サブグラフ抽出処理部152により抽出されたサブグラフに含まれるノードが人物を表す場合に、その人物の属性による脅威予兆度への寄与度を計算し、可視化してユーザに提示する。例えば、当該ノードのノード情報に含まれる属性情報が表す様々な属性項目(性別、年齢、服装、マスク着用の有無、滞在時間など)について、要素寄与度保存部160に保存された要素寄与度、すなわちノード特徴量に対する属性項目ごとの重みに基づき、各属性項目の寄与度を計算する。そして、計算された寄与度が高い方から所定数の属性項目を選択し、各属性項目の内容と寄与度を異常検知画面において所定のレイアウトで提示する。 When the node included in the subgraph extracted by the subgraph extraction processing unit 152 represents a person, the person attribute threat contribution presentation unit 153 calculates and visualizes the contribution of the person's attribute to the threat sign degree. Present to the user. For example, the element contribution degree stored in the element contribution degree storage unit 160 for various attribute items (gender, age, clothes, whether or not a mask is worn, staying time, etc.) represented by the attribute information included in the node information of the node. That is, the contribution of each attribute item is calculated based on the weight of each attribute item with respect to the node feature amount. Then, a predetermined number of attribute items are selected from the one with the highest calculated contribution, and the content and contribution of each attribute item are presented in a predetermined layout on the abnormality detection screen.
 物体属性脅威寄与度提示部154は、サブグラフ抽出処理部152により抽出されたサブグラフに含まれるノードが物体を表す場合に、その物体の属性による脅威予兆度への寄与度を計算し、可視化してユーザに提示する。例えば、当該ノードのノード情報に含まれる属性情報が表す様々な属性項目(大きさ、色、滞在時間など)について、要素寄与度保存部160に保存された要素寄与度、すなわちノード特徴量に対する属性項目ごとの重みに基づき、各属性項目の寄与度を計算する。そして、計算された寄与度が高い方から所定数の属性項目を選択し、各属性項目の内容と寄与度を異常検知画面において所定のレイアウトで提示する。 When the node included in the subgraph extracted by the subgraph extraction processing unit 152 represents an object, the object attribute threat contribution presentation unit 154 calculates and visualizes the contribution of the object attribute to the threat sign degree. Present to the user. For example, for various attribute items (size, color, staying time, etc.) represented by the attribute information included in the node information of the node, the element contribution degree stored in the element contribution degree storage unit 160, that is, the attribute for the node feature amount. Calculate the contribution of each attribute item based on the weight of each item. Then, a predetermined number of attribute items are selected from the one with the highest calculated contribution, and the content and contribution of each attribute item are presented in a predetermined layout on the abnormality detection screen.
 行動履歴寄与度提示部155は、サブグラフ抽出処理部152により抽出されたサブグラフに含まれるノードが人物または物体を表す場合に、その人物または物体と他の人物または物体との間で行われた行動による脅威予兆度への寄与度を計算し、可視化してユーザに提示する。例えば、当該ノードに接続された各エッジについて、要素寄与度保存部160に保存された要素寄与度、すなわちエッジ特徴量に対する重みに基づき、各エッジの寄与度を計算する。そして、計算された寄与度が高い方から所定数のエッジを選択し、各エッジが表す行動内容と寄与度を異常検知画面において所定のレイアウトで提示する。 The action history contribution presentation unit 155 is an action performed between the person or object and another person or object when the node included in the subgraph extracted by the subgraph extraction processing unit 152 represents a person or an object. Calculates the degree of contribution to the threat sign by, visualizes it, and presents it to the user. For example, for each edge connected to the node, the contribution of each edge is calculated based on the element contribution stored in the element contribution storage unit 160, that is, the weight for the edge feature amount. Then, a predetermined number of edges are selected from the one with the highest calculated contribution, and the action content and contribution represented by each edge are presented in a predetermined layout on the abnormality detection screen.
 言語化サマリ生成部156は、人物属性脅威寄与度提示部153、物体属性脅威寄与度提示部154および行動履歴寄与度提示部155によりそれぞれ提示された内容を言語化することで、異常検知の根拠を簡潔に表現したテキスト(サマリ)を生成する。そして、生成したサマリを異常検知画面内の所定の位置に表示する。 The verbalization summary generation unit 156 verbalizes the contents presented by the person attribute threat contribution presentation unit 153, the object attribute threat contribution presentation unit 154, and the action history contribution presentation unit 155, respectively, and is the basis for abnormality detection. Generate a text (summary) that expresses concisely. Then, the generated summary is displayed at a predetermined position on the abnormality detection screen.
 判定根拠提示部150では、以上説明した各ブロックの処理により、異常検知部130によって異常が検知された人物や物体などの要素について、当該要素に対して算出された脅威予兆度と、脅威予兆度への寄与度が高い当該要素の特徴または行動の情報と、を少なくとも含む異常検知画面を、異常検知部130の判定根拠を示す画面としてユーザに提示することができる。 In the determination basis presentation unit 150, the threat predictive degree and the threat predictive degree calculated for the element such as a person or an object in which the abnormality is detected by the abnormality detection unit 130 by the processing of each block described above An abnormality detection screen including at least information on the characteristics or behaviors of the element having a high degree of contribution to the above can be presented to the user as a screen showing the determination basis of the abnormality detection unit 130.
 図17は、根拠確認対象選択部151およびサブグラフ抽出処理部152が行う処理の概要を示す図である。図17において、(a)はサブグラフ抽出前のグラフデータを可視化した例を、(b)はサブグラフ抽出後のグラフデータを可視化した例をそれぞれ示している。 FIG. 17 is a diagram showing an outline of the processing performed by the evidence confirmation target selection unit 151 and the subgraph extraction processing unit 152. In FIG. 17, (a) shows an example of visualizing the graph data before the subgraph extraction, and (b) shows an example of visualizing the graph data after the subgraph extraction.
 図17(a)に示すグラフデータにおいてユーザが所定の操作(例えばマウスのクリック等)によりいずれかのノードを指定すると、根拠確認対象選択部151は、指定されたノードと、そのノードに接続されている各ノードおよび各エッジとを、異常検知の根拠確認の対象として選択する。このときサブグラフ抽出処理部152は、根拠確認対象選択部151により選択されたノードとエッジをサブグラフとして抽出し、抽出されたサブグラフを強調表示するとともに、グラフデータのサブグラフ以外の部分をグレーアウトして表示することで、サブグラフを可視化する。 When the user designates any node by a predetermined operation (for example, clicking a mouse) in the graph data shown in FIG. 17A, the basis confirmation target selection unit 151 is connected to the designated node and the node. Select each node and each edge as the target for checking the basis for abnormality detection. At this time, the subgraph extraction processing unit 152 extracts the nodes and edges selected by the basis confirmation target selection unit 151 as subgraphs, highlights the extracted subgraphs, and grays out and displays the parts other than the subgraphs of the graph data. By doing so, the subgraph is visualized.
 例えば、図17(a)のグラフデータにおいて、ユーザがノードO2を指定した場合を考える。この場合、指定されたノードO2と、ノードO2に隣接するノードP2、P4と、ノードO2、P2、P4間にそれぞれ設定されたエッジとを含む部分が、根拠確認対象選択部151により選択され、サブグラフ抽出処理部152よりサブグラフとして抽出される。そして図17(b)に示すように、抽出されたこれらのノードおよびエッジがそれぞれ強調表示され、他の部分がグレーアウトして表示されることで、サブグラフが可視化される。 For example, consider the case where the user specifies the node O2 in the graph data of FIG. 17 (a). In this case, the portion including the designated node O2, the nodes P2 and P4 adjacent to the node O2, and the edges set between the nodes O2, P2, and P4 are selected by the grounds confirmation target selection unit 151. It is extracted as a subgraph from the subgraph extraction processing unit 152. Then, as shown in FIG. 17 (b), these extracted nodes and edges are highlighted, and the other parts are grayed out, so that the subgraph is visualized.
 図18は、判定根拠提示部150により表示される異常検知画面の例を示す図である。図18に示す異常検知画面180では、異常を検知された人物と物体のそれぞれについて、その脅威予兆度が脅威レベルとして示されるとともに、脅威予兆度に対する特徴や行動ごとの寄与度が示されている。具体的には、カメラ2で撮影された人物については「マスク」、「滞在時間」、「上半身色」の各項目に対する寄与度が示され、カメラ1で撮影された物体については「置き去り」、「滞在時間」、「受け渡し」の各項目に対する寄与度が示されている。また、これらの人物や物体に関する不審点として、言語化サマリ生成部156によって生成されたサマリが表示されている。さらに、人物がとった不審な行動を示す映像とその撮影時刻とが、行動タイムラインとして表示されている。 FIG. 18 is a diagram showing an example of an abnormality detection screen displayed by the determination basis presentation unit 150. On the anomaly detection screen 180 shown in FIG. 18, for each of the person and the object in which the abnormality is detected, the threat sign degree is shown as a threat level, and the characteristics and the contribution degree of each action to the threat sign degree are shown. .. Specifically, the contribution to each item of "mask", "stay time", and "upper body color" is shown for the person photographed by the camera 2, and "left behind" for the object photographed by the camera 1. The degree of contribution to each item of "stay time" and "delivery" is shown. Further, as a suspicious point regarding these persons or objects, a summary generated by the verbalization summary generation unit 156 is displayed. Further, a video showing a suspicious action taken by a person and the shooting time thereof are displayed as an action timeline.
 なお、図18に示した異常検知画面180は一例であり、異常検知部130による異常検知結果およびその根拠をユーザにとって分かりやすく提示できれば、これ以外の内容や画面レイアウトで異常検知画面を表示してもよい。 The abnormality detection screen 180 shown in FIG. 18 is an example, and if the abnormality detection result by the abnormality detection unit 130 and its basis can be presented in an easy-to-understand manner for the user, the abnormality detection screen is displayed with other contents and screen layout. May be good.
 以上説明した本発明の第1の実施形態によれば、以下の作用効果を奏する。 According to the first embodiment of the present invention described above, the following effects are exhibited.
(1)異常検知システム1は、所定の監視対象場所を撮影して得られた映像または画像に基づいて、監視対象場所における複数の要素を検知し、要素ごとの属性および要素間の関係性を表すグラフデータを生成するグラフデータ生成部20と、グラフデータ生成部20により生成されたグラフデータの時空間特徴量を算出する時空間特徴量算出部110と、時空間特徴量算出部110により算出された時空間特徴量に基づいて、監視対象場所における異常を検知する異常検知部130とを備える。このようにしたので、様々な人物や物体を撮影した映像または画像から、不審な行動や異常な行動を正確に発見して異常を検知することができる。 (1) The abnormality detection system 1 detects a plurality of elements in a monitored place based on an image or an image obtained by photographing a predetermined monitored place, and determines the attributes of each element and the relationship between the elements. Calculated by the graph data generation unit 20 that generates the graph data to be represented, the spatiotemporal feature amount calculation unit 110 that calculates the spatiotemporal feature amount of the graph data generated by the graph data generation unit 20, and the spatiotemporal feature amount calculation unit 110. An abnormality detecting unit 130 for detecting an abnormality in a monitored place is provided based on the spatiotemporal feature amount. By doing so, it is possible to accurately detect suspicious behavior or abnormal behavior from images or images of various people or objects and detect the abnormality.
(2)グラフデータ生成部20は、エンティティ検知処理部21により、映像または画像に映り込んだ人物または物体を要素として検知し、検知した人物または物体の特徴を要素ごとの属性として取得するとともに、関係性検知処理部23により、人物が他の人物または物体に対して行う行動を要素間の関係性として取得する。このようにしたので、グラフデータを生成するために必要な情報を映像または画像から確実に取得して、グラフデータの生成を行うことができる。 (2) The graph data generation unit 20 detects a person or object reflected in a video or image as an element by the entity detection processing unit 21, acquires the characteristics of the detected person or object as an element, and acquires the characteristics of the detected person or object as an attribute for each element. The relationship detection processing unit 23 acquires actions taken by a person with respect to another person or object as relationships between elements. Since this is done, the information necessary for generating the graph data can be reliably acquired from the video or the image, and the graph data can be generated.
(3)グラフデータ生成部20は、要素ごとの属性を表す複数のノードと、要素間の関係性を表す複数のエッジとを組み合わせて、例えば図3に示すようなグラフデータを生成する。このようにしたので、要素ごとの属性および要素間の関係性を分かりやすく表現したグラフデータを生成することができる。 (3) The graph data generation unit 20 combines a plurality of nodes representing the attributes of each element and a plurality of edges representing the relationships between the elements to generate graph data as shown in FIG. 3, for example. By doing so, it is possible to generate graph data that clearly expresses the attributes of each element and the relationships between the elements.
(4)グラフデータ生成部20は、時系列で取得された映像または画像を所定の時刻区間Δtごとに区切った複数の時間範囲について、その時間範囲ごとにグラフデータを生成する。このようにしたので、時系列で変化する映像または画像に対して、データ量と異常検知精度のバランスを考慮した適切なグラフデータを生成することができる。 (4) The graph data generation unit 20 generates graph data for each of a plurality of time ranges obtained by dividing the video or image acquired in time series into predetermined time intervals Δt. Since this is done, it is possible to generate appropriate graph data in consideration of the balance between the amount of data and the accuracy of abnormality detection for a video or image that changes over time.
(5)異常検知システム1は、所定の時間範囲ごとに生成された各グラフデータについて、要素ごとの属性の特徴量を表すノード特徴量を抽出するノード特徴量抽出部70と、時間範囲ごとに生成された各グラフデータについて、要素間の関係性の特徴量を表すエッジ特徴量を抽出するエッジ特徴量抽出部80とをさらに備える。時空間特徴量算出部110は、ノード特徴量抽出部70およびエッジ特徴量抽出部80によりそれぞれ抽出された各グラフデータのノード特徴量およびエッジ特徴量に基づいて、時空間特徴量を算出する。このようにしたので、各ノードの特徴量に隣接ノードとの潜在的な関係性を反映した時空間特徴量を算出することができる。 (5) The abnormality detection system 1 has a node feature amount extraction unit 70 that extracts a node feature amount representing an attribute feature amount for each element for each graph data generated for each predetermined time range, and a node feature amount extraction unit 70 for each time range. Each of the generated graph data is further provided with an edge feature amount extraction unit 80 for extracting an edge feature amount representing the feature amount of the relationship between the elements. The spatiotemporal feature amount calculation unit 110 calculates the spatiotemporal feature amount based on the node feature amount and the edge feature amount of each graph data extracted by the node feature amount extraction unit 70 and the edge feature amount extraction unit 80, respectively. Since this is done, it is possible to calculate the spatiotemporal feature amount that reflects the potential relationship between the feature amount of each node and the adjacent node.
(6)異常検知システム1は、グラフデータを可視化してユーザに提示し、ユーザによるグラフデータの編集を受け付けるグラフデータ可視化編集部60をさらに備える。このようにしたので、不適切なグラフデータが誤って生成された場合でも、ユーザは正しいグラフデータに修正することが可能となる。 (6) The abnormality detection system 1 further includes a graph data visualization editing unit 60 that visualizes graph data, presents it to the user, and accepts the user to edit the graph data. By doing so, even if inappropriate graph data is erroneously generated, the user can correct it to the correct graph data.
(7)時空間特徴量算出部110は、要素ごとに時空間特徴量を算出する。異常検知部130は、特徴量分布クラスタリング部131および中心点距離計算部132により、時空間特徴量算出部110により算出された要素ごとの時空間特徴量を互いに比較し、異常判定部133により、その比較結果に基づいて要素ごとの脅威の度合いを算出することで、異常を検知する。このようにしたので、不審な行動をとる人物や不審物が存在する場合に、これを確実に判断して異常を検知することができる。 (7) The spatiotemporal feature amount calculation unit 110 calculates the spatiotemporal feature amount for each element. The anomaly detection unit 130 compares the spatiotemporal feature amounts of each element calculated by the spatiotemporal feature amount calculation unit 110 with each other by the feature amount distribution clustering unit 131 and the center point distance calculation unit 132, and the abnormality determination unit 133 uses the abnormality determination unit 133. Anomalies are detected by calculating the degree of threat for each element based on the comparison result. Since this is done, when there is a person or a suspicious object that behaves suspiciously, it can be reliably determined and the abnormality can be detected.
(8)異常検知システム1は、判定根拠提示部150により、異常検知部130により異常が検知された要素について、当該要素に対して算出された脅威の度合いと、脅威の度合いへの寄与度が高い当該要素の特徴または行動の情報と、を少なくとも含む異常検知画面180を、異常検知部130の判定根拠を示す画面としてユーザに提示する。このようにしたので、映像に映り込んだ様々な人物や物体のうち、どの人物または物体がどのような理由で不審者または不審物として検知されたのかを、ユーザに分かりやすく確認させることができる。 (8) In the abnormality detection system 1, for an element in which an abnormality is detected by the abnormality detection unit 130 by the determination basis presentation unit 150, the degree of threat calculated for the element and the degree of contribution to the degree of threat are determined. An abnormality detection screen 180 including at least high information on the characteristics or behavior of the element is presented to the user as a screen showing the determination basis of the abnormality detection unit 130. By doing so, it is possible to make the user understand in an easy-to-understand manner which person or object is detected as a suspicious person or suspicious object among various people or objects reflected in the image. ..
(9)異常検知画面180は、当該要素の特徴または行動ごとの脅威の度合いへの寄与度の情報をさらに含む。このようにしたので、不審者または不審物として検知された人物や物体のどのような特徴や行動が脅威の判断において重視されたのかを、ユーザに分かりやすく確認させることができる。 (9) The abnormality detection screen 180 further includes information on the degree of contribution to the degree of threat for each feature or action of the element. By doing so, it is possible to make the user understand in an easy-to-understand manner what characteristics and behaviors of the person or object detected as a suspicious person or a suspicious object are emphasized in the judgment of the threat.
(10)異常検知システム1を構成するコンピュータは、所定の監視対象場所を撮影して得られた映像または画像に基づいて、監視対象場所における複数の要素を検知する処理(エンティティ検知処理部21が実施するエンティティ検知処理)と、要素ごとの属性および要素間の関係性を表すグラフデータを生成する処理(グラフデータ生成部20の処理)と、グラフデータの時空間特徴量を算出する処理(時空間特徴量算出部110の処理)と、時空間特徴量に基づいて、監視対象場所における異常を検知する処理(異常検知部130の処理)とを実行する。このようにしたので、コンピュータを用いた処理により、様々な人物や物体を撮影した映像または画像から、不審な行動や異常な行動を正確に発見して異常を検知することができる。 (10) The computer constituting the abnormality detection system 1 detects a plurality of elements in the monitored place based on the video or image obtained by photographing the predetermined monitored place (the entity detection processing unit 21 performs the process). Entity detection processing to be executed), processing to generate graph data showing attributes of each element and relationships between elements (processing of graph data generation unit 20), and processing to calculate spatiotemporal feature amount of graph data (time). The processing of the spatial feature amount calculation unit 110) and the processing of detecting an abnormality in the monitored place based on the spatiotemporal feature amount (processing of the abnormality detecting unit 130) are executed. Since this is done, it is possible to accurately detect suspicious behaviors and abnormal behaviors and detect abnormalities from images or images of various people and objects photographed by processing using a computer.
 なお、本発明は上記実施形態に限定されるものではなく、その要旨を逸脱しない範囲内で、任意の構成要素を用いて実施可能である。以上説明した実施形態や変形例はあくまで一例であり、発明の特徴が損なわれない限り、本発明はこれらの内容に限定されるものではない。また、上記では種々の実施形態や変形例を説明したが、本発明はこれらの内容に限定されるものではない。本発明の技術的思想の範囲内で考えられるその他の態様も本発明の範囲内に含まれる。 It should be noted that the present invention is not limited to the above embodiment, and can be carried out by using any component within the range not deviating from the gist thereof. The embodiments and modifications described above are merely examples, and the present invention is not limited to these contents as long as the features of the invention are not impaired. Further, although various embodiments and modifications have been described above, the present invention is not limited to these contents. Other aspects considered within the scope of the technical idea of the present invention are also included within the scope of the present invention.
 1…異常検知システム、10…カメラ動画像入力部、20…グラフデータ生成部、30…グラフデータベース、40…ノードデータベース、50…エッジデータベース、60…グラフデータ可視化編集部、70…ノード特徴量抽出部、80…エッジ特徴量抽出部、90…ノード特徴量蓄積部、100…エッジ特徴量蓄積部、110…時空間特徴量算出部、120…ノード特徴量取得部、130…異常検知部、140…脅威予兆度保存部、150…判定根拠提示部、160…要素寄与度保存部 1 ... Abnormality detection system, 10 ... Camera moving image input unit, 20 ... Graph data generation unit, 30 ... Graph database, 40 ... Node database, 50 ... Edge database, 60 ... Graph data visualization editing unit, 70 ... Node feature amount extraction Unit, 80 ... Edge feature amount extraction unit, 90 ... Node feature amount storage unit, 100 ... Edge feature amount storage unit, 110 ... Spatio-temporal feature amount calculation unit, 120 ... Node feature amount acquisition unit, 130 ... Abnormality detection unit, 140 ... Threat sign storage unit, 150 ... Judgment basis presentation unit, 160 ... Element contribution storage unit

Claims (11)

  1.  所定の監視対象場所を撮影して得られた映像または画像に基づいて、前記監視対象場所における複数の要素を検知し、前記要素ごとの属性および前記要素間の関係性を表すグラフデータを生成するグラフデータ生成部と、
     前記グラフデータ生成部により生成された前記グラフデータの時空間特徴量を算出する時空間特徴量算出部と、
     前記時空間特徴量算出部により算出された前記時空間特徴量に基づいて、前記監視対象場所における異常を検知する異常検知部と、を備える異常検知システム。
    Based on the video or image obtained by shooting a predetermined monitoring target location, a plurality of elements in the monitoring target location are detected, and graph data showing the attributes of each element and the relationship between the elements is generated. Graph data generator and
    A spatiotemporal feature amount calculation unit for calculating the spatiotemporal feature amount of the graph data generated by the graph data generation unit, and a spatiotemporal feature amount calculation unit.
    An abnormality detection system including an abnormality detection unit that detects an abnormality in the monitoring target location based on the space-time feature amount calculated by the space-time feature amount calculation unit.
  2.  請求項1に記載の異常検知システムにおいて、
     前記グラフデータ生成部は、前記映像または前記画像に映り込んだ人物または物体を前記要素として検知し、検知した前記人物または前記物体の特徴を前記要素ごとの属性として取得するとともに、前記人物が他の前記人物または前記物体に対して行う行動を前記要素間の関係性として取得する異常検知システム。
    In the abnormality detection system according to claim 1,
    The graph data generation unit detects the image or the person or object reflected in the image as the element, acquires the detected feature of the person or the object as an attribute for each element, and the person is another person. An abnormality detection system that acquires an action performed on the person or the object as a relationship between the elements.
  3.  請求項1に記載の異常検知システムにおいて、
     前記グラフデータ生成部は、前記要素ごとの属性を表す複数のノードと、前記要素間の関係性を表す複数のエッジとを組み合わせて、前記グラフデータを生成する異常検知システム。
    In the abnormality detection system according to claim 1,
    The graph data generation unit is an abnormality detection system that generates graph data by combining a plurality of nodes representing attributes for each element and a plurality of edges representing relationships between the elements.
  4.  請求項1に記載の異常検知システムにおいて、
     前記グラフデータ生成部は、時系列で取得された前記映像または前記画像を所定の時刻区間ごとに区切った複数の時間範囲について、前記時間範囲ごとに前記グラフデータを生成する異常検知システム。
    In the abnormality detection system according to claim 1,
    The graph data generation unit is an abnormality detection system that generates graph data for each time range of a plurality of time ranges obtained by dividing the video or the image acquired in time series into predetermined time intervals.
  5.  請求項4に記載の異常検知システムにおいて、
     前記時間範囲ごとに生成された各グラフデータについて、前記要素ごとの属性の特徴量を表すノード特徴量を抽出するノード特徴量抽出部と、
     前記時間範囲ごとに生成された各グラフデータについて、前記要素間の関係性の特徴量を表すエッジ特徴量を抽出するエッジ特徴量抽出部と、をさらに備え、
     前記時空間特徴量算出部は、前記ノード特徴量抽出部および前記エッジ特徴量抽出部によりそれぞれ抽出された前記各グラフデータの前記ノード特徴量および前記エッジ特徴量に基づいて、前記時空間特徴量を算出する異常検知システム。
    In the abnormality detection system according to claim 4,
    For each graph data generated for each time range, a node feature amount extraction unit that extracts a node feature amount representing the feature amount of the attribute for each element, and a node feature amount extraction unit.
    For each graph data generated for each time range, an edge feature amount extraction unit for extracting an edge feature amount representing the feature amount of the relationship between the elements is further provided.
    The spatiotemporal feature amount calculation unit is based on the node feature amount and the edge feature amount of each graph data extracted by the node feature amount extraction unit and the edge feature amount extraction unit, respectively. Anomaly detection system that calculates.
  6.  請求項1に記載の異常検知システムにおいて、
     前記グラフデータを可視化してユーザに提示し、前記ユーザによる前記グラフデータの編集を受け付けるグラフデータ可視化編集部をさらに備える異常検知システム。
    In the abnormality detection system according to claim 1,
    An abnormality detection system further comprising a graph data visualization editing unit that visualizes the graph data, presents the graph data to the user, and accepts the user to edit the graph data.
  7.  請求項1に記載の異常検知システムにおいて、
     前記時空間特徴量算出部は、前記要素ごとに前記時空間特徴量を算出し、
     前記異常検知部は、前記時空間特徴量算出部により算出された前記要素ごとの前記時空間特徴量を互いに比較し、その比較結果に基づいて前記要素ごとの脅威の度合いを算出することで、前記異常を検知する異常検知システム。
    In the abnormality detection system according to claim 1,
    The spatiotemporal feature amount calculation unit calculates the spatiotemporal feature amount for each of the elements.
    The anomaly detection unit compares the spatiotemporal feature amount for each element calculated by the spatiotemporal feature amount calculation unit with each other, and calculates the degree of threat for each element based on the comparison result. An abnormality detection system that detects the abnormality.
  8.  請求項7に記載の異常検知システムにおいて、
     前記異常検知部により前記異常が検知された前記要素について、当該要素に対して算出された前記脅威の度合いと、前記脅威の度合いへの寄与度が高い当該要素の特徴または行動の情報と、を少なくとも含む異常検知画面を、前記異常検知部の判定根拠を示す画面としてユーザに提示する異常検知システム。
    In the abnormality detection system according to claim 7,
    For the element in which the abnormality is detected by the abnormality detection unit, the degree of the threat calculated for the element and information on the characteristics or behavior of the element having a high degree of contribution to the degree of the threat are provided. An abnormality detection system that presents an abnormality detection screen including at least to the user as a screen showing the determination basis of the abnormality detection unit.
  9.  請求項8に記載の異常検知システムにおいて、
     前記異常検知画面は、当該要素の特徴または行動ごとの前記寄与度の情報をさらに含む異常検知システム。
    In the abnormality detection system according to claim 8,
    The anomaly detection screen is an anomaly detection system that further includes information on the degree of contribution for each feature or action of the element.
  10.  所定の場所を撮影して得られた映像または画像に含まれる複数の要素のうち、脅威の度合いが高い要素について、当該要素の前記脅威の度合いと、前記脅威の度合いへの寄与度が高い当該要素の特徴または行動の情報と、を少なくとも含む異常検知画面を、ユーザに提示する異常検知システム。 Of the plurality of elements included in the video or image obtained by shooting a predetermined place, the element having a high degree of threat has a high degree of threat and a high degree of contribution to the degree of threat. An anomaly detection system that presents an anomaly detection screen to the user, including at least information on the characteristics or behavior of the element.
  11.  コンピュータにより、
     所定の監視対象場所を撮影して得られた映像または画像に基づいて、前記監視対象場所における複数の要素を検知する処理と、
     前記要素ごとの属性および前記要素間の関係性を表すグラフデータを生成する処理と、
     前記グラフデータの時空間特徴量を算出する処理と、
     前記時空間特徴量に基づいて、前記監視対象場所における異常を検知する処理と、を実行する異常検知方法。
    By computer
    A process of detecting a plurality of elements in the monitored place based on an image or an image obtained by shooting a predetermined monitored place, and a process of detecting a plurality of elements in the monitored place.
    Processing to generate graph data showing the attributes of each element and the relationships between the elements,
    The process of calculating the spatiotemporal features of the graph data and
    An abnormality detection method for executing a process of detecting an abnormality in the monitored location based on the spatiotemporal feature amount.
PCT/JP2021/031589 2020-11-30 2021-08-27 Abnormality detection device and abnormality detection method WO2022113453A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2020-198749 2020-11-30
JP2020198749A JP2022086627A (en) 2020-11-30 2020-11-30 Abnormality detection device and abnormality detection method

Publications (1)

Publication Number Publication Date
WO2022113453A1 true WO2022113453A1 (en) 2022-06-02

Family

ID=81754513

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/031589 WO2022113453A1 (en) 2020-11-30 2021-08-27 Abnormality detection device and abnormality detection method

Country Status (2)

Country Link
JP (1) JP2022086627A (en)
WO (1) WO2022113453A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014186497A (en) * 2013-03-22 2014-10-02 Sumitomo Electric Ind Ltd Image analysis device, image analysis method, and image analysis program
JP2016081355A (en) * 2014-10-17 2016-05-16 キヤノン株式会社 Irregularity detection method, irregularity detection device and program
JP2019152927A (en) * 2018-02-28 2019-09-12 株式会社エクォス・リサーチ Image data generation device, image recognition device, image data generation program and image recognition program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014186497A (en) * 2013-03-22 2014-10-02 Sumitomo Electric Ind Ltd Image analysis device, image analysis method, and image analysis program
JP2016081355A (en) * 2014-10-17 2016-05-16 キヤノン株式会社 Irregularity detection method, irregularity detection device and program
JP2019152927A (en) * 2018-02-28 2019-09-12 株式会社エクォス・リサーチ Image data generation device, image recognition device, image data generation program and image recognition program

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
IDEUCHI, MASAO; MURAKAMI, TOMOKAZU; TAMARU, SYUHEI; TAKASHIO, KAZUNORI; TOKUDA, HIDEYUKI: "SARADA: Support for Application of Recognizing environment by Auto Detection of Anomalies", IPSJ SIG TECHNICAL REPORT, vol. 2003, no. 115 (2003-UBI-002), 8 November 2003 (2003-11-08), JP , pages 207 - 212, XP009536943, ISSN: 0919-6072 *

Also Published As

Publication number Publication date
JP2022086627A (en) 2022-06-09

Similar Documents

Publication Publication Date Title
US10470510B1 (en) Systems and methods for full body measurements extraction using multiple deep learning networks for body feature measurements
JP5647627B2 (en) Unusual pattern discovery
CN104919794B (en) For extracting the method and system of metadata from master-slave mode camera tracking system
WO2022113439A1 (en) Data analysis device and data analysis method
JP2006079272A (en) Abnormal behavior detection apparatus and abnormal behavior detection method
WO2017212813A1 (en) Image search device, image search system, and image search method
JP2023015989A (en) Item identification and tracking system
US20230342389A1 (en) Image Search in Walkthrough Videos
CN108229289B (en) Target retrieval method and device and electronic equipment
US20170053172A1 (en) Image processing apparatus, and image processing method
Nemmour et al. Fuzzy neural network architecture for change detection in remotely sensed imagery
CN116264004A (en) Vision-based monitoring of field safety compliance based on worker re-identification and personal protective equipment classification
JP2019070934A (en) Video processing apparatus, video processing method and program
CN112784966B (en) Anomaly detection method and system in defined spatial region
JP5493747B2 (en) Tracking method, monitoring system and program
Bazo et al. Baptizo: A sensor fusion based model for tracking the identity of human poses
Sree et al. An evolutionary computing approach to solve object identification problem for fall detection in computer vision-based video surveillance applications
WO2022113453A1 (en) Abnormality detection device and abnormality detection method
Bouma et al. WPSS: Watching people security services
Chen et al. Urban damage estimation using statistical processing of satellite images
Singh et al. Chaotic whale-atom search optimization-based deep stacked auto encoder for crowd behaviour recognition
JP2018142137A (en) Information processing device, information processing method and program
US10872235B2 (en) Tracking shoppers and employees
Irvine et al. Context and quality estimation in video for enhanced event detection
Zhao et al. Estimating the size of crowds through deep learning

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21897431

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21897431

Country of ref document: EP

Kind code of ref document: A1