CN111553305B - System and method for identifying illegal videos - Google Patents

System and method for identifying illegal videos Download PDF

Info

Publication number
CN111553305B
CN111553305B CN202010387707.8A CN202010387707A CN111553305B CN 111553305 B CN111553305 B CN 111553305B CN 202010387707 A CN202010387707 A CN 202010387707A CN 111553305 B CN111553305 B CN 111553305B
Authority
CN
China
Prior art keywords
scene
target
violation
data
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010387707.8A
Other languages
Chinese (zh)
Other versions
CN111553305A (en
Inventor
徐非凡
王勇
高原
刘希宏
韩红卫
苏兴华
李小成
徐智锋
王洁飞
刘思远
李富平
谭宁军
苏彦平
陈胜伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China National Petroleum Corp
CNPC Chuanqing Drilling Engineering Co Ltd
Original Assignee
China National Petroleum Corp
CNPC Chuanqing Drilling Engineering Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China National Petroleum Corp, CNPC Chuanqing Drilling Engineering Co Ltd filed Critical China National Petroleum Corp
Priority to CN202010387707.8A priority Critical patent/CN111553305B/en
Publication of CN111553305A publication Critical patent/CN111553305A/en
Application granted granted Critical
Publication of CN111553305B publication Critical patent/CN111553305B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/48Matching video sequences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a system and a method for identifying a violation video, which belong to the field of industrial and civil security monitoring. The system and the method for identifying the illegal video enhance the working efficiency of personnel, enable monitoring personnel to monitor a plurality of cameras at the same time, reduce the working intensity of monitoring personnel, comprehensively cover all construction sites and make up the defects of single, one-sided and unstable existing illegal identification methods.

Description

System and method for identifying illegal videos
Technical Field
The invention belongs to the field of security monitoring, and particularly relates to a system and a method for identifying illegal videos.
Background
Video monitoring technology applications are mainly focused on government departments and special departments and industries of finance, public security, traffic, power, and the like. Among them, government and financial sectors occupy 20.9% and 20.6% of market share, respectively. However, with the progress of social informatization, the demands of video monitoring in more and more industries and fields are greatly increased, and video monitoring technology starts to extend from individual fields of banks, traffic and the like to multiple fields, and the traditional security monitoring is developed to management monitoring and production management monitoring.
The traditional method relies on site management personnel to monitor the behavior safety of operators and check and treat the hidden object trouble, but the hidden object trouble can not be found out in time, the requirement on the monitoring personnel is high, and the actual effect is poor. The existing monitoring mode is that a full-time monitoring person observes an image video to check and correct staff's illegal behaviors and check and correct physical hidden dangers, and generally only can observe images of individual construction sites and twelve cameras simultaneously, so that the monitoring coverage area is narrow and the monitoring efficiency is low.
The existing intelligent monitoring system only stays at the discrimination level of static violations, and for continuous action violations and process violations, discrimination is only carried out from the characteristics of motion tracks, and the system has no mode universality and environmental integrity. The same behavior pattern may not hold when certain factors in the environment change. Therefore, the reliability of the recognition cannot be ensured.
The invention patent application with the application number of 201710408664.5 discloses a security monitoring system, which is one of the security systems and comprises a monitoring server, an alarm module and a video monitoring module, wherein the monitoring server is connected with a local area network, and the local area network is connected with a control module and is accessed to the Internet; the video monitoring module comprises: an infrared camera and a spherical camera for shooting monitoring pictures. The monitoring system architecture can detect abnormal conditions such as fire, illegal invasion and the like and alarm in time. The technical scheme is that the early intelligent monitoring system realizes the identification of certain scene modes, such as flame identification, intrusion identification and the like, by manually customizing parameters, and the intelligent monitoring system can only identify the abnormality of a fixed scene, has weak universality, and has technical defects of insufficient robustness and the like.
Disclosure of Invention
The embodiment of the invention aims to provide a system and a method for identifying illegal videos, which overcome the technical defects.
In order to solve the technical problems, the invention provides a violation video recognition system, which comprises:
the violation template construction module is used for carrying out data labeling and model training according to the original data acquired by the network camera to obtain a target detection model;
the intelligent processing module consists of an image acquisition sub-module and a target tracking sub-module, wherein the image acquisition sub-module acquires original data, converts photons into the original image data through a CMOS sensor chip, compresses the image and inputs the image into the target tracking sub-module; the target tracking sub-module performs feature extraction and feature screening on the collected original image data by loading a target detection model to obtain a target, and records the position of the target in the image by a target record in the previous frame data after the target is obtained to obtain a target track so as to realize target tracking;
the violation judging module consists of a tracking data describing layer, a scene describing layer and a scene identifying layer, wherein the data describing layer carries out merging and arrangement on the output result of the intelligent processing module and then transmits the merged result to the scene describing layer; the scene description layer generates three types of data, namely a time relation data structure, a space relation data structure and a scene element data structure, according to the data provided by the data description layer, and the three types of data are transmitted to the scene recognition layer for violation discrimination after being generated;
the alarm output module consists of a hidden danger judging module, an alarm prompting module and an alarm parameter configuration module; the hidden danger judging module performs redundancy screening output according to the rule breaking judging result and the redundancy screening rule of the rule breaking judging module; the alarm prompting module performs visual display on the judged result; the alarm parameter configuration module is mainly used for setting redundancy screening rules of the judging result.
Further, the composition of the time relation data structure is the sequence of key targets on a time axis and the movement speed thereof; the composition of the spatial relationship data structure is the up-down left-right relationship between key targets on a two-dimensional image; the scene element data structure comprises the types of targets, the number of targets of various types and the positions of the targets in the scene.
Preferably, the scene recognition layer consists of four sub-modules of a scene template, a trigger, a scene recorder and a violation template;
the scene template templates the time, place and character basic composition information contained in a certain specific procedure and action;
the trigger matches the scene information constructed by the data description layer at each moment with the scene template, and when the current scene is successfully matched with the elements in the scene template, the scene recorder submodule is triggered to start recording;
the scene recorder module records interaction information among each key element contained in the scene and matches the interaction information with the violation template; when the matching is successful, outputting a matching result, inquiring a corresponding violation item according to the matching result, and sending the violation item to an alarm output module;
the violation templates template the relationship among time, place, person, cause, pass, result basic composition information, time of occurrence of the target set, composition of the target set in a certain time period and target set elements in a certain time period contained in a certain specific process and action.
Further, the scene recorder module uses cosine similarity to describe similarity of the template vector to the state description.
Preferably, the interaction information comprises the time of occurrence of the target set, the composition of the target set in a certain time period and the relationship among the target set elements in a certain time period.
The invention also provides a method for identifying the illegal video, which comprises the following steps:
step one: violation template construction
According to the original data acquired by the network camera, performing data labeling and model training to obtain a target detection model;
step two: intelligent processing
Collecting original data, converting photons into original image data through a CMOS sensor chip, performing image compression, performing feature extraction and feature screening on the collected original image data to obtain a target, recording the position of the target in an image through target recording in previous frame data after the target is obtained, and obtaining a target track to realize target tracking;
step three: violation identification
Combining and sorting the intelligent processing results, generating three types of data, namely a time relationship data structure, a space relationship data structure and a scene element data structure, and then carrying out violation discrimination;
step four: alarm output
And performing redundancy screening output according to the rule violation judgment result and the redundancy screening rule, and performing visual display on the judgment result.
Further, the rule violation discrimination in the third step comprises the following steps:
templating time, place and character basic composition information contained in a specific process and action;
templating the relationship among time, place, character, cause, pass, result basic composition information, time of occurrence of target set, composition of target set in a certain time period and target set elements in a certain time period contained in a certain specific process and action;
matching scene information constructed by the data description layer at each moment with a scene template, and triggering a scene recorder submodule to start recording when the current scene is successfully matched with elements in the scene template;
the scene recorder module records interaction information among each key element contained in the scene and matches the interaction information with the violation template; and when the matching is successful, outputting a matching result, and inquiring and outputting corresponding violation entries according to the matching result.
The beneficial effects of the invention are as follows:
(1) The violation video recognition system enhances the working efficiency of personnel, so that monitoring personnel can monitor a plurality of cameras at the same time, the working intensity of monitoring professional personnel is reduced, and all construction sites are comprehensively covered.
(2) Compared with the traditional violation identification modeling method, a more comprehensive and basic scene element description and scene recovery mode is provided, and the problems of singleness, unilateralness, instability and the like of the existing violation identification mode are solved. The method can model and identify most of operation procedures and actions, and has universality.
In order to make the above-mentioned objects of the present invention more comprehensible, preferred embodiments accompanied with figures are described in detail below.
Drawings
FIG. 1 is a frame diagram of a violation identification system;
FIG. 2 is a schematic diagram of a violation determination method;
FIG. 3 is a flow chart of a violation method;
fig. 4 is a target element information screen;
FIG. 5 is a schematic diagram one of scenario 1;
FIG. 6 is a schematic diagram II of scenario 1;
FIG. 7 is a schematic diagram of scenario 2;
fig. 8 is a schematic diagram of scenario 3.
Detailed Description
Further advantages and effects of the present invention will become apparent to those skilled in the art from the disclosure of the present specification, by describing the embodiments of the present invention with specific examples.
In the present invention, the up, down, left, and right in the drawings are regarded as the up, down, left, and right of the offending video identification system described in the present specification.
The exemplary embodiments of the present invention will now be described with reference to the accompanying drawings, however, the present invention may be embodied in many different forms and is not limited to the examples described herein, which are provided to fully and completely disclose the present invention and fully convey the scope of the invention to those skilled in the art. The terminology used in the exemplary embodiments illustrated in the accompanying drawings is not intended to be limiting of the invention. In the drawings, like elements/components are referred to by like reference numerals.
Unless otherwise indicated, terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art. In addition, it will be understood that terms defined in commonly used dictionaries should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and will not be interpreted in an idealized or overly formal sense.
First embodiment:
a first embodiment of the present invention relates to a violation video recognition system, see fig. 1, comprising:
the violation template construction module is used for carrying out data labeling and model training according to the original data acquired by the network camera to obtain a target detection model;
the intelligent processing module consists of an image acquisition sub-module and a target tracking sub-module, wherein the image acquisition sub-module acquires original data, converts photons into the original image data through the CMOS sensor chip, compresses the image and inputs the image into the target tracking sub-module; the target tracking sub-module performs feature extraction and feature screening on the collected original image data by loading a target detection model to obtain a target, and records the position of the target in the image by a target record in the previous frame data after the target is obtained to obtain a target track so as to realize target tracking;
the violation judging module consists of a tracking data describing layer, a scene describing layer and a scene identifying layer, wherein the data describing layer carries out merging and arrangement on the output result of the intelligent processing module and then transmits the merged result to the scene describing layer; the scene description layer generates three types of data, namely a time relation data structure, a space relation data structure and a scene element data structure, according to the data provided by the data description layer, and the three types of data are transmitted to the scene recognition layer for violation discrimination after being generated;
the alarm output module consists of a hidden danger judging module, an alarm prompting module and an alarm parameter configuration module; the hidden danger judging module performs redundancy screening output according to the rule breaking judging result and the redundancy screening rule of the rule breaking judging module; the alarm prompting module performs visual display on the judged result; the alarm parameter configuration module is mainly used for setting redundancy screening rules of the judging result.
Specifically, the violation template construction module performs data annotation and model training by using an existing data annotation tool LabelImg, a model training frame Mxnet, tensorFlow and the like according to the original data acquired by the network camera to obtain a model for target detection;
the intelligent processing module mainly comprises two sub-modules of image acquisition and target tracking. The image acquisition sub-module is responsible for acquiring the original data, converts photons into the original image data through the CMOS sensor chip, compresses the image, and inputs the image into the target tracking sub-module. The target tracking sub-module performs feature extraction on the collected original image data by loading a target detection model, and further performs feature screening on features in the image data to obtain a target. After the target is acquired, the position of the target in the image is recorded through the target record in the previous frame data, and the target track is acquired through updating the position of the previous frame data target, so that the target tracking is realized.
The key point of the violation discrimination is that 6 elements of time, place, person, cause and pass in the scene are restored. Whether a certain working procedure or a specific limb action, the environment must be built around, and the construction is meaningful in a specific scene.
The alarm output module mainly comprises a hidden danger judging module, an alarm prompting module and an alarm parameter configuration module. And the hidden danger judging sub-module performs redundancy screening output on the judging result of the operation procedure and the operation action according to the rule violation judging module. And the alarm prompting sub-module performs visual display on the judged result. The alarm parameter configuration sub-module is mainly used for setting redundancy screening rules of the judging result.
Second embodiment:
the present embodiment relates to a system for identifying a violation video, referring to fig. 1, including:
the violation template construction module is used for carrying out data labeling and model training according to the original data acquired by the network camera to obtain a target detection model;
the intelligent processing module consists of an image acquisition sub-module and a target tracking sub-module, wherein the image acquisition sub-module acquires original data, converts photons into the original image data through the CMOS sensor chip, compresses the image and inputs the image into the target tracking sub-module; the target tracking sub-module performs feature extraction and feature screening on the collected original image data by loading a target detection model to obtain a target, and records the position of the target in the image by a target record in the previous frame data after the target is obtained to obtain a target track so as to realize target tracking;
the violation judging module consists of a tracking data describing layer, a scene describing layer and a scene identifying layer, wherein the data describing layer carries out merging and arrangement on the output result of the intelligent processing module and then transmits the merged result to the scene describing layer; the scene description layer generates three types of data, namely a time relation data structure, a space relation data structure and a scene element data structure, according to the data provided by the data description layer, and the three types of data are transmitted to the scene recognition layer for violation discrimination after being generated;
specifically, the trace data description layer classifies output results of the intelligent processing module into two types of information: a target front and rear frame position information list, a target category information list. Each element in the position information list of the front and rear frames of the targets is composed of two pieces of coordinate information, and the space positions of each target in the front and rear frames of images are recorded; each element in the target category information list is composed of a name describing the target category, and category information of each target in the current state is recorded.
The scene description layer generates three types of data, namely a time relation data structure, a space relation data structure and a scene element data structure, based on the data provided by the tracking data description layer, and the three types of data are transmitted to the next scene recognition layer for discrimination after being generated.
The spatial relationship data structure is composed of the upper, lower, left and right relationship between key targets on the two-dimensional image.
The time relation data structure is composed of the sequence of key targets on a time axis and the movement speed of the key targets.
The scene element data structure is composed of the types of targets appearing in the scene (i.e., all targets that may appear in the scene to be detected that may be separated into individual individuals), the number of targets of each type, and the target location.
The specific generation method of the three types of data comprises the following steps:
1. a spatial relationship description list is constructed, wherein the list consists of spatial relationship description elements among targets, and each spatial relationship description element consists of category names of two targets and spatial distances of the two targets.
The class names of the two targets are taken from class names corresponding to the two targets to be described in the target class information list, and the space distance between the two targets is generated by calculating Euclidean distance by using post-frame coordinates of two elements to be described in the target front-rear frame position information list.
2. A time relation description list is constructed, and elements in the list are used for describing the movement trend of the same target. The target motion trend is generated by calculating the euclidean distance of two coordinates by using the spatial coordinates of each target element in the target front and rear frame position information list on the front and rear frame images.
3. A scene element description list is constructed, each element in the list consists of the category and id of the object appearing in the scene, and the id is the number of the objects in the current list.
The scene recognition layer consists of four sub-modules of a scene template, a trigger, a scene recorder and a violation template.
The scene template submodule templates time, place and character basic composition information contained in a certain specific procedure and action to generate a scene template.
The trigger submodule matches the scene information such as time, place, person and the like constructed by the data description layer at each moment with the scene template. And triggering a scene recorder submodule to start recording after the current scene is successfully matched with the elements in the scene template.
The scene recorder module uses cosine similarity to describe the similarity of the template vector to the state description. The scene recorder module is used for acquiring target elements simultaneously existing in the scene description layer and the violation template by inquiring the scene description layer, recording interaction information between each key element (refer to elements in a target list in the violation template) contained in the scene, recording three types of information, namely the time when a target set appears, the composition of the target set in a certain time period and the relationship among the target set elements in a certain time period, and matching with a time sequence dictionary in the violation template. And when the matching is successful, outputting a matching result. And inquiring the corresponding violation entry according to the matching result, and sending the violation entry to an alarm output module.
The rule breaking template sub-module is used for templating the relationship among target set elements in a certain period of time, a certain place, a person, a cause, a pass and a result which are included in a certain procedure and action and are basically composed of information, namely the occurrence time of a target set. The construction method of the violation template is to generate a target list, a target distance list and a single target time domain displacement list which are contained in the violation operation. The violation templates are accumulated in time domain based on the scene templates and consist of a higher-level time sequence dictionary. The index in the time sequence dictionary is a time stamp, and the elements in the time sequence dictionary are three lists corresponding to the time stamp, wherein:
each element of the target list is composed of a target class name and a corresponding sequence number in the scene;
the distance between the targets and each element of the list is formed by the Euclidean distance between each target in the target list and all other targets on the image;
each element in the single-target time domain displacement list is configured as a spatial distance of each target on the front and rear frame images.
The composition of the time relation data structure is the sequence of key targets on a time axis and the movement speed of the key targets; the composition of the spatial relationship data structure is the relationship between key targets on the two-dimensional image; the composition of the scene element data structure is the kind of targets, the number of targets of each kind and the positions of the targets appearing in the scene.
The alarm output module consists of a hidden danger judging module, an alarm prompting module and an alarm parameter configuration module; the hidden danger judging module performs redundancy screening output according to the rule breaking judging result and the redundancy screening rule of the rule breaking judging module; the alarm prompting module performs visual display on the judged result; the alarm parameter configuration module is mainly used for setting redundancy screening rules of the judging result.
Third embodiment:
the embodiment provides a method for identifying a violation video, which comprises the following steps:
step one: violation template construction
According to the original data acquired by the network camera, performing data labeling and model training to obtain a target detection model;
step two: intelligent processing
Collecting original data, converting photons into original image data through a CMOS sensor chip, performing image compression, performing feature extraction and feature screening on the collected original image data to obtain a target, recording the position of the target in an image through target recording in previous frame data after the target is obtained, and obtaining a target track to realize target tracking;
step three: violation identification
Combining and sorting the intelligent processing results, generating three types of data, namely a time relationship data structure, a space relationship data structure and a scene element data structure, and then carrying out violation discrimination;
step four: alarm output
And performing redundancy screening output according to the rule violation judgment result and the redundancy screening rule, and performing visual display on the judgment result.
The method for identifying the violation videos comprises 4 steps of constructing a violation template, intelligently processing, judging the violation and outputting an alarm. Original data conversion, information extraction, violation discrimination and alarm output are completed through the combination of the parts. Detecting flow of the violation video identification system:
the system firstly collects the original data through the network camera, performs data labeling and model training, and obtains a model for target detection. On the basis, the original image acquired after model deployment is subjected to target detection processing, a target detection result is used for target tracking, and the violation discrimination of the working procedure and the behavior action is carried out through the output information of the target tracking. After the intelligent processing process is finished, the violation hidden danger judging module judges the intelligent processing result according to the item configuration and hidden danger description content of the violation hidden danger, and finally inputs the behavior data judged as the violation into the alarm processing to alarm. In the alarm processing module, the alarm content and the historical violation data can be managed.
Fourth embodiment:
the embodiment provides a method for identifying a violation video, referring to fig. 2 and fig. 3, including the following steps:
step one: violation template construction
According to the original data acquired by the network camera, performing data labeling and model training to obtain a target detection model;
step two: intelligent processing
Collecting original data, converting photons into original image data through a CMOS sensor chip, performing image compression, performing feature extraction and feature screening on the collected original image data to obtain a target, recording the position of the target in an image through target recording in previous frame data after the target is obtained, and obtaining a target track to realize target tracking;
step three: violation identification
Combining and sorting the intelligent processing results, generating three types of data, namely a time relationship data structure, a space relationship data structure and a scene element data structure, and then carrying out violation discrimination;
step four: alarm output
And performing redundancy screening output according to the rule violation judgment result and the redundancy screening rule, and performing visual display on the judgment result.
Specifically, the rule violation discrimination in the third step comprises the following steps:
templating time, place and character basic composition information contained in a specific process and action;
templating the relationship among time, place, character, cause, pass, result basic composition information, time of occurrence of target set, composition of target set in a certain time period and target set elements in a certain time period contained in a certain specific process and action;
matching scene information constructed by the data description layer at each moment with a scene template, and triggering a scene recorder submodule to start recording when the current scene is successfully matched with elements in the scene template;
the scene recorder module records interaction information among each key element contained in the scene and matches the interaction information with the violation template; and when the matching is successful, outputting a matching result, and inquiring and outputting corresponding violation entries according to the matching result.
The following is a general description of the implementation of the system and method for violations of oil drilling single wellhead operations:
step one: constructing a violation template
(1) Template assembly
Spatial relationship data structure:
the composition is the relationship of the key targets on the two-dimensional image.
In the current violation scene, four targets are included, namely a person, an elevator pin and a drill rod. Its spatial relationship is only used as a counterexample verification and not as a rule-breaking criterion. If the central coordinates of the elevator pin and the central coordinates of the drilling tool are in a certain range, otherwise, the space relation forming the violation is invalid.
Time relation data structure
The composition is the sequence of key targets on a time axis and the movement speed of the key targets.
The lower left corner is assumed to be the origin of the image in the coordinate system describing the image. The longitudinal coordinates of the elevator pin are firstly changed from large to small and then from small to large. When the data describing the elevator pin in the logger conforms to such a changing relationship, the time relationship data template may be successfully matched.
Scene element data structure
In the current violation scene, four targets are included, namely a person, an elevator pin and a drill rod.
Template recording duration
The video is 25 frames per second, and for this violation, the overall job duration is about 1 minute, thus setting the template recording duration to 1500 frames.
Threshold value of violation
And the method is used for judging the similarity between the violation template and the actual scene recorder, and judging the violation when the similarity exceeds a threshold. Here set to 80%
Step two: intelligent processing
The system collects the original data and detects the target to obtain scene information. The data acquisition is carried out by using a special SDK of a camera, and target detection is carried out by selecting a main stream target detection method, such as SSD, fast-rcnn, yolo and the like, according to the complexity of a detection object.
And obtaining the target element information in the picture shown in fig. 4 through an intelligent processing module.
Object name Scene(s) Target quantity
Human body Drilling platform 1
Elevator Drilling platform 1
Elevator pin Drilling platform 1
Drill rod Drilling platform 1
Target name paraphrasing:
drilling tool: a cylindrical pipe for use in the drilling industry for delivering liquids into the subsurface.
And (3) hanging an elevator: connecting piece for fixing drilling tool to hoist
Elevator pin: another state of the elevator indicates that the elevator establishes a temporary stable connection with the drill pipe.
Step three: violation identification
Scenario 1-the elevator pin is in the air (FIG. 5), the elevator pin is landed (FIG. 6)
Element (b): elevator, elevator pin, man, drill rod
After the intelligent processing module inputs the original image data, four targets including an elevator, an elevator pin, a person and a drill rod are detected in the scene.
Spatial relationship: the spatial relationship of the scenario is that the elevator pin moves down relative to the position of the person and the elevator, and the elevator is stationary on the rig floor.
Time relationship: according to the composition of the targets and the positions among the targets in the front and back frame image data, the movement track of the key target (the elevator pin moves downwards) can be obtained. Here, to illustrate the implementation, the state description of the intermediate frames of data is skipped.
The current operation scene can be preliminarily judged to be the tripping operation scene by comparing the current information with the template and the violation template information according to the trigger module. Thus, the scope of the violation entry to be triggered can be narrowed. At this time, the scene recorder is triggered to start recording the current element space state and the element time state.
Scene 2-remove Pin replacement Elevator (see FIG. 7)
Element (b): elevator, elevator pin, man, drill rod
After the intelligent processing module inputs the original image data, four targets including an elevator, an elevator pin, a person and a drill rod are detected in the scene.
Spatial relationship: the spatial relationship of the scene is that the positions of the elevator pin and the elevator are interchanged, and according to the target position coordinate obtained by target detection, the position of the original elevator pin is the elevator, the position of the original elevator is changed into the elevator pin.
Time relationship: from the composition of the objects in the previous image frame data and the locations between the objects, the movement pattern of the key objects (the elevator pin objects disappear and then reappear at the locations of the previous side elevators) can be obtained.
Scene 3-lifting pin lifting (see FIG. 8)
Element (b): elevator, elevator pin, man, drill rod
After the intelligent processing module inputs the original image data, four targets including an elevator, an elevator pin, a person and a drill rod are detected in the scene.
Spatial relationship: the spatial relationship of the scenario is that the elevator pin is moving up relative to both the elevator and the person, and the elevator is stationary.
Time relationship: from the composition of the objects in the previous image frame data and the locations between the objects, the movement pattern of the key objects (the elevator pin objects disappear and then reappear at the locations of the previous side elevators) can be obtained.
And (3) integrating the element contents, the spatial relationship and the scene description information constructed by the time relationship of the three scenes, comparing the scene description information with the corresponding contents in the violation template continuously by the scene recorder, and describing the similarity (the vector length is 1500 frames) of the time sequence dictionary in the violation template and the time sequence dictionary dynamically generated by the scene description layer by using cosine similarity. And dynamically outputting the similarity, comparing the similarity with a threshold value, and when the similarity is larger than a judging threshold value of 80%, indicating that the scene meets the template requirement, and further carrying out logic judgment.
In the current scenario, in a normal working procedure, the number of people in the target set should be at least 2. Since the entire process is always operated by a single person, the operation procedure is determined as a violation in this scenario.
Step four: alarm output
And searching the content of the violation item according to the ID number corresponding to the violation template, printing the violation content on the display, and alarming.
It will be understood by those of ordinary skill in the art that the foregoing embodiments are specific examples of carrying out the invention and that various changes in form and details may be made therein without departing from the spirit and scope of the invention.

Claims (6)

1. A violation video recognition system, comprising:
the violation template construction module is used for carrying out data labeling and model training according to the original data acquired by the network camera to obtain a target detection model;
the intelligent processing module consists of an image acquisition sub-module and a target tracking sub-module, wherein the image acquisition sub-module acquires original data, converts photons into the original image data through a CMOS sensor chip, compresses the image and inputs the image into the target tracking sub-module; the target tracking sub-module performs feature extraction and feature screening on the collected original image data by loading a target detection model to obtain a target, and records the position of the target in the image by a target record in the previous frame data after the target is obtained to obtain a target track so as to realize target tracking;
the violation judging module consists of a tracking data describing layer, a scene describing layer and a scene identifying layer, wherein the tracking data describing layer carries out merging and arrangement on the output result of the intelligent processing module and then transmits the merged and arranged result to the scene describing layer; the scene description layer generates three types of data, namely a time relation data structure, a space relation data structure and a scene element data structure, according to the data provided by the tracking data description layer, and the three types of data are transmitted to the scene recognition layer for violation discrimination after being generated;
the composition of the time relation data structure is the sequence of key targets on a time axis and the movement speed of the key targets; the composition of the spatial relationship data structure is the up-down left-right relationship between key targets on a two-dimensional image; the composition of the scene element data structure is the types of targets, the number of targets of various types and the positions of the targets in the scene;
the alarm output module consists of a hidden danger judging module, an alarm prompting module and an alarm parameter configuration module; the hidden danger judging module performs redundancy screening output according to the rule breaking judging result and the redundancy screening rule of the rule breaking judging module; the alarm prompting module performs visual display on the judged result; the alarm parameter configuration module is used for setting redundancy screening rules of the judging result.
2. The offending video identification system of claim 1, wherein: the scene recognition layer consists of four sub-modules of a scene template, a trigger, a scene recorder and a violation template;
the scene template templates the time, place and character basic composition information contained in specific procedures and actions;
the trigger matches the scene information constructed by the tracking data description layer at each moment with a scene template, and when the current scene is successfully matched with elements in the scene template, the scene recorder submodule is triggered to start recording;
the scene recorder module records interaction information among each key element contained in the scene and matches the interaction information with the violation template; when the matching is successful, outputting a matching result, inquiring a corresponding violation item according to the matching result, and sending the violation item to an alarm output module;
the violation templates template the relationship among time, place, person, cause, pass, result basic composition information, time of occurrence of the target set, composition of the target set in a certain time period and target set elements in a certain time period contained in the specific process and action.
3. The offending video identification system of claim 2, wherein: the scene recorder module uses cosine similarity to describe similarity of the template vector to the state description.
4. The offending video identification system of claim 2, wherein: the interaction information comprises the time of the occurrence of the target set, the composition of the target set in a certain time period and the relationship among the target set elements in a certain time period.
5. A method of identifying a violation video of a violation video identification system according to any of claims 1 to 4, comprising the steps of:
step one: violation template construction
According to the original data acquired by the network camera, performing data labeling and model training to obtain a target detection model;
step two: intelligent processing
Collecting original data, converting photons into original image data through a CMOS sensor chip, performing image compression, performing feature extraction and feature screening on the collected original image data to obtain a target, recording the position of the target in an image through target recording in previous frame data after the target is obtained, and obtaining a target track to realize target tracking;
step three: violation identification
Combining and sorting the intelligent processing results, generating three types of data, namely a time relationship data structure, a space relationship data structure and a scene element data structure, and then carrying out violation discrimination;
step four: alarm output
And performing redundancy screening output according to the rule violation judgment result and the redundancy screening rule, and performing visual display on the judgment result.
6. The method of claim 5, wherein the identifying of the violation in step three comprises the steps of:
templating the basic composition information of the time, place and person contained in the specific process and action;
templating the relationship among time, place, character, cause, pass, result basic composition information, time of occurrence of the target set, composition of the target set in a certain time period and target set elements in a certain time period contained in a specific process and action;
matching scene information constructed by the tracking data description layer at each moment with a scene template, and triggering a scene recorder submodule to start recording when the current scene is successfully matched with elements in the scene template;
the scene recorder module records interaction information among each key element contained in the scene and matches the interaction information with the violation template; and when the matching is successful, outputting a matching result, and inquiring and outputting corresponding violation entries according to the matching result.
CN202010387707.8A 2020-05-09 2020-05-09 System and method for identifying illegal videos Active CN111553305B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010387707.8A CN111553305B (en) 2020-05-09 2020-05-09 System and method for identifying illegal videos

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010387707.8A CN111553305B (en) 2020-05-09 2020-05-09 System and method for identifying illegal videos

Publications (2)

Publication Number Publication Date
CN111553305A CN111553305A (en) 2020-08-18
CN111553305B true CN111553305B (en) 2023-05-05

Family

ID=72004540

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010387707.8A Active CN111553305B (en) 2020-05-09 2020-05-09 System and method for identifying illegal videos

Country Status (1)

Country Link
CN (1) CN111553305B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112711994A (en) * 2020-12-21 2021-04-27 航天信息股份有限公司 Method and system for detecting illegal operation behaviors based on scene recognition
CN112837531A (en) * 2020-12-25 2021-05-25 朗坤智慧科技股份有限公司 Group-level violation behavior video identification method and device based on 5G network
CN116883952B (en) * 2023-09-07 2023-11-17 吉林同益光电科技有限公司 Electric power construction site violation identification method and system based on artificial intelligence algorithm

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101266710A (en) * 2007-03-14 2008-09-17 中国科学院自动化研究所 An all-weather intelligent video analysis monitoring method based on a rule
CN108052900A (en) * 2017-12-12 2018-05-18 成都睿码科技有限责任公司 A kind of method by monitor video automatic decision dressing specification
CN109711320A (en) * 2018-12-24 2019-05-03 兴唐通信科技有限公司 A kind of operator on duty's unlawful practice detection method and system
CN110119656A (en) * 2018-02-07 2019-08-13 中国石油化工股份有限公司 Intelligent monitor system and the scene monitoring method violating the regulations of operation field personnel violating the regulations
CN110163143A (en) * 2019-05-17 2019-08-23 国网河北省电力有限公司沧州供电分公司 Unlawful practice recognition methods, device and terminal device
CN110309735A (en) * 2019-06-14 2019-10-08 平安科技(深圳)有限公司 Exception detecting method, device, server and storage medium
CN110795989A (en) * 2019-08-28 2020-02-14 广东电网有限责任公司 Intelligent safety monitoring system for electric power operation and monitoring method thereof
CN110826514A (en) * 2019-11-13 2020-02-21 国网青海省电力公司海东供电公司 Construction site violation intelligent identification method based on deep learning

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101266710A (en) * 2007-03-14 2008-09-17 中国科学院自动化研究所 An all-weather intelligent video analysis monitoring method based on a rule
CN108052900A (en) * 2017-12-12 2018-05-18 成都睿码科技有限责任公司 A kind of method by monitor video automatic decision dressing specification
CN110119656A (en) * 2018-02-07 2019-08-13 中国石油化工股份有限公司 Intelligent monitor system and the scene monitoring method violating the regulations of operation field personnel violating the regulations
CN109711320A (en) * 2018-12-24 2019-05-03 兴唐通信科技有限公司 A kind of operator on duty's unlawful practice detection method and system
CN110163143A (en) * 2019-05-17 2019-08-23 国网河北省电力有限公司沧州供电分公司 Unlawful practice recognition methods, device and terminal device
CN110309735A (en) * 2019-06-14 2019-10-08 平安科技(深圳)有限公司 Exception detecting method, device, server and storage medium
CN110795989A (en) * 2019-08-28 2020-02-14 广东电网有限责任公司 Intelligent safety monitoring system for electric power operation and monitoring method thereof
CN110826514A (en) * 2019-11-13 2020-02-21 国网青海省电力公司海东供电公司 Construction site violation intelligent identification method based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
李千登等.《钻井作业典型违章行为视频监控智能分析技术研究》.《工业安全与环保》.2019,第45卷(第12期),全文. *

Also Published As

Publication number Publication date
CN111553305A (en) 2020-08-18

Similar Documents

Publication Publication Date Title
CN111553305B (en) System and method for identifying illegal videos
US10276019B2 (en) Surveillance system and method for predicting patient falls using motion feature patterns
JP6905850B2 (en) Image processing system, imaging device, learning model creation method, information processing device
JP6867153B2 (en) Abnormality monitoring system
Cameron et al. Emergency situation awareness from twitter for crisis management
CN101751744B (en) Detection and early warning method of smoke
CN105913559A (en) Motion sensing technique based bank ATM intelligent monitoring method
US20160350583A1 (en) Image search system and image search method
KR102149832B1 (en) Automated Violence Detecting System based on Deep Learning
CN109410497B (en) Bridge opening space safety monitoring and alarming system based on deep learning
CN110245603A (en) A kind of group abnormality behavior real-time detection method
WO2020167155A1 (en) Method and system for detecting troubling events during interaction with a self-service device
Abd et al. Human fall down recognition using coordinates key points skeleton
CN115311735A (en) Intelligent recognition early warning method for abnormal behaviors
CN111753587A (en) Method and device for detecting falling to ground
Mwedzi et al. Machine learning applications for fire detection in a residential building
Limbasiya et al. COVID-19 face mask and social distancing detector using machine learning
CN112330742A (en) Method and device for recording activity routes of key personnel in public area
CN116993265A (en) Intelligent warehouse safety management system based on Internet of things
CN116682162A (en) Robot detection algorithm based on real-time video stream
KR20230097854A (en) Method and system for recognizing dangerous behavior of workers in power plant
CN114429677A (en) Coal mine scene operation behavior safety identification and assessment method and system
Terdal et al. YOLO-Based Video Processing for CCTV Surveillance
CN117351405B (en) Crowd behavior analysis system and method
Nithesh et al. Anomaly Detection in Surveillance Videos Using Deep Learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant