CN113469150A - Method and system for identifying risk behaviors - Google Patents

Method and system for identifying risk behaviors Download PDF

Info

Publication number
CN113469150A
CN113469150A CN202111029514.6A CN202111029514A CN113469150A CN 113469150 A CN113469150 A CN 113469150A CN 202111029514 A CN202111029514 A CN 202111029514A CN 113469150 A CN113469150 A CN 113469150A
Authority
CN
China
Prior art keywords
target object
behavior
neural network
ultra
base station
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111029514.6A
Other languages
Chinese (zh)
Other versions
CN113469150B (en
Inventor
金淼
张军
雷民
周峰
殷小东
陈习文
卢冰
王斯琪
陈卓
周玮
汪泉
付济良
聂高宁
王旭
齐聪
郭子娟
余雪芹
刘俊
朱赤丹
郭鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Electric Power Research Institute Co Ltd CEPRI
Original Assignee
China Electric Power Research Institute Co Ltd CEPRI
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Electric Power Research Institute Co Ltd CEPRI filed Critical China Electric Power Research Institute Co Ltd CEPRI
Priority to CN202111029514.6A priority Critical patent/CN113469150B/en
Publication of CN113469150A publication Critical patent/CN113469150A/en
Application granted granted Critical
Publication of CN113469150B publication Critical patent/CN113469150B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Signal Processing (AREA)
  • Probability & Statistics with Applications (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method and a system for identifying risk behaviors, wherein the method comprises the following steps: selecting acceleration signals of a target object in a continuous time period when the target object performs a target behavior, and acquiring acceleration characteristic vectors of the acceleration signals in the continuous time period through a deep neural network model; selecting a depth image of a target object at the end moment of a target behavior, and acquiring a target object feature vector of the target object in the depth image through a depth neural network model; selecting a two-dimensional image of a target object at the end moment of a target behavior, and acquiring background element characteristics of the target object in the two-dimensional image through a deep neural network model; inputting the acceleration characteristic vector and the target object characteristic vector into the neural network discrimination model to obtain the behavior action of the target object; and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the background element characteristics.

Description

Method and system for identifying risk behaviors
Technical Field
The invention relates to the technical field of human behavior recognition, in particular to a method and a system for recognizing risk behaviors.
Background
The intelligent identification technology for the safety risk of the field operation is a comprehensive technology consisting of machine vision, behavior identification, mode identification and the like, is an auxiliary management and control means for personal risk in the security field, is related to the fields of public safety, smart cities, medical rehabilitation and the like at present, and is also a hotspot development field under the new foundation construction background of China. The core human behavior recognition technology is that a computer is used for analyzing collected behavior information of a person to judge what the person is doing, and classification is carried out according to implementation modes, so that human behavior recognition can be divided into a recognition mode based on a sensor and a recognition mode based on vision, and the combination of the two modes is also included. The human body behavior identification method based on vision realizes human body abnormal behavior identification based on visual information through key technical steps of video processing, motion detection and tracking, feature extraction, mode classification and identification and the like; the wearable sensing information-based human behavior recognition technology is characterized in that inertial motion signal data of body movement is acquired through a plurality of sensor nodes bound or worn on the body, operations such as feature extraction and feature selection are performed on the basis of preprocessing such as denoising of the data, a reasonable classifier is selected for training, and finally a recognition algorithm is verified through action signals acquired actually. The inevitable trend in the field of risk and behavior monitoring and analysis of a multi-sensor combined monitoring mode is that a researcher can simultaneously obtain multi-source data including a color video, a depth image sequence and a three-dimensional human body skeleton sequence by distributing and controlling various visual sensors, wearable motion sensors and the like on an electric power field; under the drive of rich information sources and efficient intelligent algorithms, the field application of the abnormal behavior identification algorithm of the operators is more feasible.
In recent years, the integration of information technology and safety production is greatly promoted in the power industry, and advanced technologies such as intelligent safety tools, video monitoring and image analysis are actively adopted, so that the field safety and technical prevention level is improved.
Prior art 1 (publication No. CN 112149761A) provides a power intelligent construction site violation detection method based on YOLOv4 improved algorithm. Prior art 1 discloses: collecting images of a power construction site for training a model; carrying out image enhancement on the acquired image of the electric power construction site; marking the target area obtained after image and data enhancement by using a rectangular frame, and acquiring the coordinates and the included types of the rectangular frame; training an improved model based on YOLOv4 according to the acquired image of the construction site and the image obtained after data enhancement, the acquired coordinates of the rectangular frame and the acquired types; and acquiring images of the electric power construction site in real time, acquiring images of the electric power construction site to be detected in real time according to the trained model, and outputting violation images. Prior art 1 is applicable to the detection of the personnel violation and the construction vehicle violation of job site, realizes that electric power wisdom building site is visual, intelligent management, can effectively promote engineering scene management level and reduce the safety risk.
However, the electric power intelligent construction site violation detection method based on the YOLOv4 improved algorithm disclosed in prior art 1 is too single, the detectable categories are limited, the algorithm recognition efficiency can only detect categories with static characteristics, such as helmet wearing, vehicle loading and the like, while dynamic behaviors of workers cannot be judged, and the violation detection category and accuracy rate have relatively large limitations.
In the prior art 2, (publication number: CN 111814601A) proposes a video analysis method combining target detection and human body posture estimation, which obtains a live camera video stream through a streaming media server, and captures a video frame from the live camera video stream to obtain a decoded picture; performing rapid target detection on each obtained picture by adopting a deep learning algorithm to find out target objects of all identifiable categories; judging whether a person exists, and sequentially analyzing dangerous behaviors and dangerous scenes of the pictures; in the process of carrying out dangerous behavior analysis and dangerous scene analysis, if a dangerous event is found, an alarm is reported and pushed to a display platform in time; and continuously analyzing the video stream in real time. According to the method, the target detection and human body posture estimation algorithm are combined, the method is introduced into the field of intelligent monitoring and identification of the infrastructure video, dangerous behaviors or scenes can be identified, the labor force of manually staring at the screen in the traditional monitoring can be liberated, and the efficiency of infrastructure safety management is effectively improved.
However, in the video analysis method combining target detection and human body posture estimation disclosed in prior art 2, a two-dimensional image is used for alarming a dangerous scene, but due to the problem that objects overlap and are shielded in the two-dimensional image, false alarm is easily generated when a worker and a dangerous area appear in the same image. And the image data can only be used for detecting simple behaviors, and more semantic information is needed to be used for danger analysis on complex behaviors.
Therefore, a technique is needed to enable intelligent identification of field work risks.
Disclosure of Invention
The technical scheme of the invention provides a method and a system for identifying risk behaviors, which are used for solving the problem of how to intelligently identify the risk of field operation.
In order to solve the above problem, the present invention provides a method for identifying a risk behavior, the method comprising:
selecting acceleration signals of a target object in a continuous time period when the target object performs a target behavior, and acquiring acceleration characteristic vectors of the acceleration signals in the continuous time period through a deep neural network model;
selecting a depth image of a target object at the end moment of a target behavior, and acquiring a target object feature vector of the target object in the depth image through a depth neural network model;
selecting a two-dimensional image of a target object at the end moment of a target behavior, and acquiring background element characteristics of the target object in the two-dimensional image through a deep neural network model;
inputting the acceleration characteristic vector and the target object characteristic vector into the neural network discrimination model to obtain the behavior action of the target object;
and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the background element characteristics.
Preferably, the method further comprises the following steps: determining a depth and a width of the deep neural network model;
the depth of the deep neural network model is as follows:
Figure 599836DEST_PATH_IMAGE001
the width of the deep neural network model is as follows:
Figure 241032DEST_PATH_IMAGE002
wherein
Figure 255125DEST_PATH_IMAGE003
And
Figure 196536DEST_PATH_IMAGE004
to find the constants by the method of searching the optimum values,
Figure 309986DEST_PATH_IMAGE005
is a manually adjusted parameter;
when the constraint condition is satisfied
Figure 579424DEST_PATH_IMAGE006
Adaptively obtaining parameters of the deep neural network model, wherein
Figure 6994DEST_PATH_IMAGE007
Is a set parameter.
Preferably, the method further comprises the following steps: determining attention weights for the deep neural network model:
determining an attention feature pyramid structure, wherein each layer of the attention feature pyramid structure is provided with different attention weights, and the attention feature weights respectively correspond to feature contribution ratios of the two-dimensional images with different resolutions;
the attention weight calculation formula is as follows:
Figure 927546DEST_PATH_IMAGE008
wherein
Figure 680738DEST_PATH_IMAGE009
A vector of scalar or two-dimensional image channels that are features; where i is the serial number of the current input feature, j is the serial number of the non-output feature, ϵ = 0.0001.
Preferably, the method further comprises the following steps: determining a position of the target object:
acquiring the time consumed by data transmission between a preset ultra-wideband base station and an ultra-wideband tag of the target object, and measuring the distance between the ultra-wideband base station and the tag by a two-way time flight method based on the time consumed by data transmission;
determining a relative position of the target object and the ultra-wideband base station based on a distance between the ultra-wideband base station and the tag;
positioning the ultra-wideband base station, and determining the absolute position of the ultra-wideband base station;
determining an absolute position of the target object based on the absolute position of the ultra-wideband base station and a relative position of the target object and the ultra-wideband base station;
and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the absolute position of the target object and the background element characteristics.
Based on another aspect of the invention, the invention provides a system for identifying risk behaviors, the system comprising:
the first acquisition unit is used for selecting acceleration signals of a target object in a continuous time period when the target object performs a target behavior, and acquiring acceleration characteristic vectors of the acceleration signals in the continuous time period through a deep neural network model;
the second acquisition unit is used for selecting a depth image of a target object at the end moment of a target behavior and acquiring a target object feature vector of the target object in the depth image through a depth neural network model;
the third acquisition unit is used for selecting a two-dimensional image of a target object at the end moment of a target behavior and acquiring background element characteristics of the target object in the two-dimensional image through a deep neural network model;
a fourth obtaining unit, configured to input the acceleration feature vector and the target object feature vector into the neural network discrimination model, and obtain a behavior action of the target object;
and the result unit is used for judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the background element characteristics.
Preferably, the device further comprises a fifth obtaining unit, configured to determine a depth and a width of the deep neural network model;
the depth of the deep neural network model is as follows:
Figure 358844DEST_PATH_IMAGE010
the width of the deep neural network model is as follows:
Figure 465472DEST_PATH_IMAGE011
wherein
Figure 850316DEST_PATH_IMAGE012
And
Figure 899044DEST_PATH_IMAGE013
to find the constants by the method of searching the optimum values,
Figure 267708DEST_PATH_IMAGE005
is a manually adjusted parameter;
when the constraint condition is satisfied
Figure 99398DEST_PATH_IMAGE014
Adaptively obtaining parameters of the deep neural network model, wherein
Figure 476765DEST_PATH_IMAGE015
Is a set parameter.
Preferably, the fifth obtaining unit is further configured to: determining attention weights for the deep neural network model:
determining an attention feature pyramid structure, wherein each layer of the attention feature pyramid structure is provided with different attention weights, and the attention feature weights respectively correspond to feature contribution ratios of the two-dimensional images with different resolutions;
the attention weight calculation formula is as follows:
Figure 571760DEST_PATH_IMAGE008
wherein
Figure 286775DEST_PATH_IMAGE009
A vector of scalar or two-dimensional image channels that are features; where i is the serial number of the current input feature, j is the serial number of the non-output feature, ϵ = 0.0001.
Preferably, the system further comprises a sixth acquisition unit, configured to determine the position of the target object:
acquiring the time consumed by data transmission between a preset ultra-wideband base station and an ultra-wideband tag of the target object, and measuring the distance between the ultra-wideband base station and the tag by a two-way time flight method based on the time consumed by data transmission;
determining a relative position of the target object and the ultra-wideband base station based on a distance between the ultra-wideband base station and the tag;
positioning the ultra-wideband base station, and determining the absolute position of the ultra-wideband base station;
determining an absolute position of the target object based on the absolute position of the ultra-wideband base station and a relative position of the target object and the ultra-wideband base station;
and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the absolute position of the target object and the background element characteristics.
The invention provides a computer-readable storage medium, characterized in that the computer-readable storage medium stores a computer program for executing the above-mentioned method for identifying a risk behaviour.
The present invention provides an electronic device, characterized in that the electronic device includes:
a processor;
a memory for storing the processor-executable instructions;
the processor is used for reading the executable instructions from the memory and executing the method for identifying the risk behaviors.
The technical scheme of the invention provides a method and a system for identifying risk behaviors, wherein the method comprises the following steps: selecting acceleration signals of a target object in a continuous time period when the target object performs a target behavior, and extracting acceleration characteristic vectors of the acceleration signals in the continuous time period through a deep neural network model; selecting a depth image of a target object at the end moment of a target behavior, and extracting a target object feature vector of the target object in the depth image through a depth neural network model; selecting a two-dimensional image of a target object at the end moment of a target behavior, and extracting background element characteristics of the target object in the two-dimensional image through a deep neural network model; inputting the acceleration characteristic vector and the target object characteristic vector into a neural network discrimination model to obtain the behavior action of the target object; and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the background element characteristics. The intelligent identification scheme for the risk of the field operation, which is provided by the technical scheme of the invention, changes the manual supervision of passive field operation into active intelligent monitoring, and provides possibility for the prior early warning of the safety risk of the field operation. According to the technical scheme, the work mode that the computer does not stop for 24 hours is used, the safety supervision task is completed by assistance of manpower, the problem that the energy of safety management and control personnel is limited at present is solved, meanwhile, the labor cost can be reduced, the power field safety management appeal is written into an algorithm, the safety management and control are performed on the operation field, and the objectivity of safety supervision is effectively guaranteed.
Drawings
A more complete understanding of exemplary embodiments of the present invention may be had by reference to the following drawings in which:
FIG. 1 is a flow chart of a method for identifying risk behaviors in accordance with a preferred embodiment of the present invention;
FIG. 2 is a schematic diagram of an embodiment of a cross-modal power job site risk behavior identification system according to a preferred embodiment of the present invention;
FIG. 3 is a schematic diagram of an example of a cross-modal power job site risk behavior identification method according to a preferred embodiment of the present invention;
FIG. 4 is a schematic diagram of one example of a network structure of an image target recognition algorithm in accordance with a preferred embodiment of the present invention;
FIG. 5 is a schematic diagram of one example of a network structure of a cross-modal behavior recognition algorithm in accordance with a preferred embodiment of the present invention;
FIG. 6 is a schematic diagram of an example of an indoor and outdoor integrated precise positioning algorithm according to a preferred embodiment of the present invention;
FIG. 7 is a schematic diagram of an embodiment of a cross-modal power job site risk behavior identification device in accordance with a preferred embodiment of the present invention; and
fig. 8 is a diagram of a system for identifying risk behaviors in accordance with a preferred embodiment of the present invention.
Detailed Description
The exemplary embodiments of the present invention will now be described with reference to the accompanying drawings, however, the present invention may be embodied in many different forms and is not limited to the embodiments described herein, which are provided for complete and complete disclosure of the present invention and to fully convey the scope of the present invention to those skilled in the art. The terminology used in the exemplary embodiments illustrated in the accompanying drawings is not intended to be limiting of the invention. In the drawings, the same units/elements are denoted by the same reference numerals.
Unless otherwise defined, terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. Further, it will be understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and will not be interpreted in an idealized or overly formal sense.
FIG. 1 is a flow chart of a method for identifying risk behaviors in accordance with a preferred embodiment of the present invention. Aiming at the requirement of safety control of the existing electric power field operation personnel, the invention provides a cross-modal electric power operation site risk behavior identification method, the implementation mode of the invention can realize the analysis of multi-modal information of the electric power field operation, and the intelligent sensing and judging mode is used for detecting whether the behavior of the working personnel meets the safety standard.
As shown in fig. 1, the present invention provides a method for identifying risk behaviors, the method comprising:
step 101: selecting acceleration signals of a target object in a continuous time period when the target object performs a target behavior, and acquiring acceleration characteristic vectors of the acceleration signals in the continuous time period through a deep neural network model;
step 102: selecting a depth image of a target object at the end moment of a target behavior, and acquiring a target object feature vector of the target object in the depth image through a depth neural network model;
step 103: selecting a two-dimensional image of a target object at the end moment of a target behavior, and acquiring background element characteristics of the target object in the two-dimensional image through a deep neural network model;
step 104: inputting the acceleration characteristic vector and the target object characteristic vector into a neural network discrimination model to obtain the behavior action of the target object;
step 105: and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the background element characteristics.
The invention provides a cross-modal electric power operation site risk behavior identification method, which comprises the following steps: the system comprises an image target identification algorithm, a risk behavior identification algorithm, a cross-modal behavior identification algorithm and an indoor and outdoor integrated accurate positioning algorithm.
The input data of the cross-modal behavior recognition algorithm provided by the invention are data of a plurality of different modalities, including a two-dimensional image acquired by a common camera, a depth image acquired by a depth camera and an acceleration signal acquired by an acceleration sensor. The cross-modal behavior recognition algorithm comprises the following steps:
step one, carrying out data processing on the acceleration signals in a period of continuous time and extracting features through a neural network model. Firstly, data processing is needed, and the specific method is that the data processing is carried out in three axial directions
Figure 859839DEST_PATH_IMAGE016
Single dimensional data overlay formation
Figure 281593DEST_PATH_IMAGE017
A two-dimensional matrix. And then inputting the data into a convolutional neural network model, and extracting the characteristic vector of the acceleration data through superposition of a plurality of convolutional and pooling layers. Finally, obtaining a one-dimensional characteristic vector;
and step two, extracting the human body characteristics of the staff from the depth image at the finishing moment through a depth neural network. The deep neural network first preprocesses the data through the ResNet-50 network and combines the output characteristics of each stage of the ResNet-50 network. Then respectively extracting human key points and partial associated vectors as human characteristic vectors of the working personnel through two multilayer convolutional neural networks;
and step three, extracting background element characteristics of the two-dimensional image at the ending moment through a deep neural network. The deep neural network adopts a selective search method to extract a plurality of candidate areas on a test image. And then adjusting the size of the candidate region to a fixed size so that the candidate region can be matched with the input size of the features, training a classifier for each class by obtaining the feature vector through the model, and scoring the output feature vector by using the classifier. For each class, calculating an index of a coincidence region, adopting non-maximum value inhibition, and removing the regions of the overlapping positions on the basis of the regions of the highest score; finally, training a linear regression model to generate a stricter bounding box for each identified object in the image;
inputting all the characteristics into a neural network discrimination model, judging the behavior of the worker through the acceleration signal characteristics and the depth image characteristics, and judging whether the behavior of the worker meets the safety standard or not by combining the background characteristics of the two-dimensional image. Such as: in the high-voltage electroscope operation, when the worker does not lift the high-voltage electroscope rod or the worker only has the hand-lifting action but does not have the high-voltage electroscope rod in the hand, the worker judges that the action is illegal.
The image target recognition algorithm is used for recognizing and extracting various static characteristic elements in a power operation working scene, for example, whether protective equipment of workers is completely worn is detected through images shot by a camera; the risk behavior recognition algorithm is used for risk behavior recognition which cannot be completed by a single image, and comprises a cross-modal behavior recognition algorithm and an indoor and outdoor integrated precise positioning algorithm. The cross-modal behavior recognition algorithm is used for recognizing the behavior and motion states of the workers and analyzing whether the behavior of the workers meets the safety standard or not; the indoor and outdoor integrated precise positioning algorithm is used for determining and tracking the real-time position of the worker, and gives a warning when the worker enters a dangerous working area or leaves a working post according to a preset dangerous working area and a preset safe working area.
Preferably, before identifying the risk behavior, the method further comprises: and acquiring a two-dimensional image and a depth image of the target behavior and an acceleration signal in the motion process of the target behavior.
Preferably, the method further comprises the following steps: determining a depth and a width of the deep neural network model;
the depth of the deep neural network model is as follows:
Figure 422856DEST_PATH_IMAGE018
the width of the deep neural network model is:
Figure 234954DEST_PATH_IMAGE019
wherein
Figure 1922DEST_PATH_IMAGE020
And
Figure 481445DEST_PATH_IMAGE021
to find the constants by the method of searching the optimum values,
Figure 714980DEST_PATH_IMAGE022
is a manually adjusted parameter;
when the constraint condition is satisfied
Figure 155320DEST_PATH_IMAGE023
Adaptively obtaining parameters of the deep neural network model, wherein
Figure 804607DEST_PATH_IMAGE024
Is a set parameter. Wherein the parameter is a learning rate, so that the objective function converges to a local minimum.
Preferably, the method further comprises the following steps: determining attention weights for the deep neural network model:
determining an attention characteristic pyramid structure, wherein each layer of the attention characteristic pyramid structure is provided with different attention weights which respectively correspond to characteristic contribution ratios of two-dimensional images with different resolutions;
the attention weight calculation formula is:
Figure 528849DEST_PATH_IMAGE025
wherein
Figure 136548DEST_PATH_IMAGE026
A vector of scalar or two-dimensional image channels that are features; where i is the serial number of the current input feature, j is the serial number of the non-output feature, ϵ = 0.0001. The features are semantic features of the two-dimensional image, including color, texture, shape, spatial relationship, attribute features, and the like.
Preferably, the image target recognition algorithm can automatically search for the depth and width of the optimal demodulation whole neural network through a group of composite coefficients, and different from the traditional method which depends on manually adjusting the structure of the network, the method can search out a group of optimal parameters which best accord with the network. The depth of the network model can be expressed as
Figure 985555DEST_PATH_IMAGE027
The width of the network model can be expressed as
Figure 579479DEST_PATH_IMAGE028
Wherein
Figure 768015DEST_PATH_IMAGE029
And
Figure 671249DEST_PATH_IMAGE030
as constants that can be found by the search for optimum,
Figure 7552DEST_PATH_IMAGE031
are manually adjustable parameters. When it is satisfied with
Figure 529800DEST_PATH_IMAGE032
Can adaptively obtain the parameters of the network, wherein
Figure 445279DEST_PATH_IMAGE033
Is a set parameter. The belt being constrained mostThe optimal parameter solving method can be realized by adjusting parameters
Figure 394781DEST_PATH_IMAGE034
A control model when
Figure 952801DEST_PATH_IMAGE035
A minimum optimal basic model is obtained; when increasing
Figure 403374DEST_PATH_IMAGE034
When two dimensions of the model are simultaneously enlarged, the model is enlarged, the performance is improved, and the consumed resources are enlarged.
The image target detection algorithm adopts a multi-scale data fusion method, and the input of the algorithm is not limited to fixed resolution, and can be data with any resolution. The multi-scale data fusion method adopts an attention feature pyramid structure, and different attention weights are introduced into each layer of the feature pyramid, so that feature contribution ratios of different layers can be adjusted while input data of different scales are fused. The attention weight is calculated by the formula
Figure 300923DEST_PATH_IMAGE036
Wherein
Figure 296692DEST_PATH_IMAGE037
A scalar quantity of the features or a vector quantity of a certain channel, and also can be a multi-dimensional tensor; where i is the serial number of the current input feature, j is the serial number of the non-output feature, ϵ = 0.0001. To ensure that the weight is greater than 0, a ReLU activation function is used before the weight calculation step.
Before judging whether the behavior action of the target object meets the preset safety behavior standard based on the background element characteristics, the method further comprises the following steps: determining the position of the target object:
setting an ultra-wideband base station, determining the time consumed by data transmission between the ultra-wideband base station and an ultra-wideband tag of a target object, and measuring the distance between the ultra-wideband base station and the tag by a two-way time flight method based on the time consumed by data transmission;
determining the relative position of the target object and the ultra-wideband base station based on the distance between the ultra-wideband base station and the label;
positioning the ultra-wideband base station, and determining the absolute position of the ultra-wideband base station;
determining the absolute position of the target object based on the absolute position of the ultra-wideband base station and the relative position of the target object and the ultra-wideband base station;
and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the absolute position and the background element characteristics of the target object.
The indoor and outdoor integrated precise positioning algorithm provided by the invention provides a high-accuracy positioning scheme by combining the Beidou satellite and the ultra-wideband technology. The Beidou satellite provides large-range positioning in an outdoor scene, and the ultra-wideband technology provides outdoor small-range or indoor high-accuracy positioning. The two positioning techniques are combined as follows:
step one, a worker wears the equipment provided with the ultra-wideband tag and enters a working scene provided with a plurality of ultra-wideband base stations. And measuring the distance between the base station and the tag by using a bidirectional time flight method according to the time consumed by data transmission between the ultra-wideband base station and the tag. Determining the relative position of a worker and a base station through a neural network training model according to the distance relationship between the base station and the label;
and step two, measuring the absolute position of the ultra-wideband base station by a Beidou satellite positioning technology. By using a real-time dynamic algorithm of multiple base stations, each fixed reference station does not directly send any correction information to a mobile user, but all original data is sent to a data processing center through a data communication link, a mobile station sends the approximate coordinates of the mobile station to a control center, the data management center broadcasts a group of fixed optimal reference stations after receiving position information, and a virtual reference station is simulated near the position of the user by using the data of the fixed optimal reference stations to form an ultra-short baseline with the mobile station. And performing auxiliary fusion analysis on all the data to determine the absolute position of the worker in the map.
The invention provides a cross-modal electric power operation site risk behavior recognition system, which comprises: the device comprises a sensing unit, a transmission unit, an edge computing unit, a cloud platform unit and a terminal equipment unit.
The sensing unit is used for acquiring and capturing original data of multiple modes of an electric power operation site and performing all-dimensional data acquisition on the risk behavior information of the electric power operation site. The method comprises the following steps: the method comprises the steps of collecting a two-dimensional image of a working site through a common camera, collecting a depth image of the working site through a depth camera, collecting an acceleration signal through an acceleration sensor, and collecting a signal for positioning through a Beidou chip and an ultra-wideband chip.
The transmission unit is used for transmitting the data acquired by the sensing unit and transmitting the data acquired by the sensing unit to the edge calculation unit;
the edge computing unit is used for receiving the captured data, processing and analyzing the captured data, and judging whether the behavior of the worker has risks or not through the cross-modal power operation site risk behavior identification method. And issuing a warning when the safety risk is judged to exist.
The cloud platform unit is used for summarizing and storing risk behaviors of all work places in the area, and then visual display and data analysis can be carried out on summarized data. The cloud platform has the main functions of remote visualization of operation, violation warning, violation statistics and the like. The remote operation visualization function can be used for remotely viewing all the real-time video images of the power working site accessed to the cloud platform. The violation warning function can receive the warning sent by the edge calculation unit and visually display the warning about the time, the place and the violation image. The violation behavior statistical function can be used for counting and analyzing all violation operations, for example, the number of violations in each region is checked, so that safety supervision is enhanced, and safety education should be developed when frequent violations occur.
The terminal equipment unit can be used for recording the construction work ticket flow and recording the risks in the operation process. The terminal equipment unit can perform data transmission with the cloud platform, when the staff violate the operation, the cloud platform can also send an alarm to the terminal equipment, the business process is suspended, and the subsequent steps cannot be performed until the safety supervisor eliminates the risk to confirm the construction safety.
The data of multiple modes in the sensing unit of the invention needs to be captured by different devices respectively, and the arrangement position of the collected devices needs to meet the data collection specification. For example, the position of the camera is not suitable to be too close to make the work site shoot incompletely, and is not suitable to be too far to cause the staff and the equipment to be too small in the image; the arrangement position of the sensors is to avoid signal shielding and interference; the Beidou chip is arranged in a relatively open area, and satellite signals can be received.
The transmission modes of the transmission unit comprise a wired transmission mode and a wireless transmission mode, and the optimal transmission mode is selected according to different data transmission requirements. Fixed equipment such as a visual camera, an ultra-wideband base station and the like adopts wired transmission to reduce mutual interference of signals, and equipment needing to be moved such as an acceleration sensor, an ultra-wideband tag and the like adopts a wireless transmission mode to ensure portability.
The wired transmission in the transmission unit of the invention mainly adopts a universal serial bus to carry out data transmission. The serial bus has the advantages of simple connection, hot plug support, high transmission speed, strong error correction capability, no need of extra power supply, low cost and the like. The universal serial bus interface is consistent and simple in connection, the system can automatically detect and configure the equipment, and the system supporting hot plugging of newly added equipment does not need to be restarted; the theoretical rate of the universal serial bus interface is 10Gbps, and as partial bandwidth is reserved to support other functions, the actual effective bandwidth is about 7.2Gbps, and the theoretical transmission speed can reach 900 MB/s; the transmission protocol comprises functions of transmission error management, error recovery and the like, and simultaneously, transmission errors are processed according to different transmission types; the bus can provide 5V voltage power supply for the devices connected to the bus, and the small devices do not need additional power supply; the interface circuit is simple and easy to realize, and the cost is lower than that of a serial port and a parallel port.
In the transmission unit of the invention, ZigBee is mainly adopted for data transmission in wireless transmission. ZigBee transmission has the advantages of low power consumption, low cost, short time delay, high capacity, high safety, unlicensed frequency band and the like. The power consumption of the ZigBee standby is lower than that of common wireless transmission modes, such as Bluetooth, WiFi and the like; the requirement on the communication controller is reduced by greatly simplifying the protocol, so that the cost is low; the response speed of Zigbee is high, generally, the sleep state is only 15ms, and the node connection and network entry only needs 30 ms; a three-level security mode is provided, including no security setting, using an access control list to prevent illegal acquisition of data, and using a symmetric password of a high-level encryption standard to flexibly determine security attributes thereof; the common frequency bands of ZigBee are 2.4GHZ, 868MHz and 915MHz, which are all unlicensed frequency bands.
According to the edge computing unit, the data processing and analyzing method adopts the cross-mode electric power operation site risk behavior identification method, and whether the operation behavior of the worker has a safety risk or not is judged. And repeatedly and circularly detecting until the operation is finished, giving out a warning when the risk is detected, and sending the risk record data to the cloud platform.
The invention provides a cross-modal electric power operation site risk behavior recognition device, which comprises: the system comprises an equipment box, an intelligent safety helmet and a positioning base station;
the equipment box comprises an algorithm operation platform, a camera, a depth camera, a wireless communication equipment receiving end and alarm equipment. The camera is in wired connection with the algorithm operation platform through a universal serial bus, and the sensor is in wireless connection with the algorithm operation platform through ZigBee. Acquiring image information through a common camera and a depth camera, acquiring sensor information through a receiving end of wireless communication equipment, analyzing data through an algorithm operating platform to judge whether the operation of a worker meets the standard, and sending out a warning by warning equipment when dangerous behaviors are detected;
the intelligent safety helmet is provided with an acceleration sensor, a Beidou positioning chip, an ultra-wideband positioning label chip, a power supply module and a wireless transmission module sending end. The wireless transmission module sending end is used for sending the acquired sensing information to the equipment box in real time for data analysis;
the positioning base station comprises an ultra-wideband positioning base station module, a Beidou positioning module and a power supply module. The positioning base station needs to be dispersedly arranged in a working site, and the periphery of the positioning base station cannot be shielded by metal or concrete buildings.
Referring to fig. 2, a schematic diagram of an embodiment of a method for identifying risk behaviors of a cross-modal power operation site provided in the present invention is shown. The method comprises the following steps:
the method comprises the following steps that firstly, a sensing unit collects and captures original data of multiple modes of an electric power operation site, and omnibearing data collection is carried out on the electric power operation site. The concrete expression is as follows: the method comprises the steps of collecting a two-dimensional image of a working site through a common camera, collecting a depth image of the working site through a depth camera, collecting an acceleration signal through an acceleration sensor, and collecting a signal for positioning through a Beidou chip and an ultra-wideband chip. The acquired equipment arrangement position needs to meet the data acquisition specification, for example, the camera arrangement position is not suitable for being too close to ensure that the working site cannot shoot completely, and is not suitable for being too far away to ensure that the working personnel and the equipment are too small in the image; the arrangement position of the sensor is to avoid the interference of signal metal and a concrete shelter on signal transmission; the position that big dipper chip was arranged should be in comparatively spacious region, guarantees to receive satellite signal.
And step two, the transmission unit is used for transmitting the data acquired by the sensing unit and transmitting the data acquired by the sensing unit to the edge calculation unit. The transmission modes of the transmission unit comprise a wired transmission mode and a wireless transmission mode, and the optimal transmission mode is selected according to different data transmission requirements. Fixed equipment such as a camera and an ultra-wideband base station adopts wired transmission to reduce mutual interference of signals, and equipment needing to be moved such as an acceleration sensor and an ultra-wideband tag adopts a wireless transmission mode to ensure portability.
The wire transmission mainly adopts a universal serial bus for data transmission, and the serial bus has the advantages of simple connection, hot plug support, high transmission speed, strong error correction capability, no need of extra power supply, low cost and the like. The universal serial bus interface is consistent and simple in connection, the system can automatically detect and configure the equipment, and the system supporting hot plugging of newly added equipment does not need to be restarted; the theoretical rate of the universal serial bus interface is 10Gbps, and as partial bandwidth is reserved to support other functions, the actual effective bandwidth is about 7.2Gbps, and the theoretical transmission speed can reach 900 MB/s; the transmission protocol comprises functions of transmission error management, error recovery and the like, and simultaneously, transmission errors are processed according to different transmission types; the bus can provide 5V voltage power supply for the devices connected to the bus, and the small devices do not need additional power supply; the interface circuit is simple and easy to realize, and the cost is lower than that of a serial port and a parallel port.
The wireless transmission mainly adopts ZigBee for data transmission, the ZigBee transmission has the advantages of low power consumption, low cost, short time delay, high capacity, high safety, unlicensed frequency band and the like, and a ZigBee chip in a 2.4GHz frequency band is selected as a chip for wireless transmission in the embodiment.
And thirdly, the edge computing unit is used for receiving, processing and analyzing the captured data, repeatedly circulating the detection of the power field operation condition until the operation is finished through a cross-mode power operation field risk behavior identification method, sending a warning when the safety risk of field operation personnel is judged, and sending the behavior with the risk and the image data to the cloud platform.
And fourthly, the cloud platform unit is used for summarizing and storing the risk behaviors of all the working places in the area, and then visual display and data analysis can be carried out on the summarized data. The cloud platform has the main functions of remote visualization of operation, violation warning, violation statistics and the like. The operation remote visualization function can remotely view all real-time video images accessed to the cloud platform. The violation warning function can receive the warning sent by the edge calculation unit and visually display the time, the place and the violation image. The violation behavior statistical function can count and analyze all violation operations, for example, the violation operations with the largest number of occurrences are analyzed to facilitate safety education, and the number of violations in each region is checked to facilitate strengthening safety supervision.
And step five, the terminal equipment unit is used for recording the construction work ticket flow and preventing the illegal work ticket from continuing when the operation has risks. The terminal equipment unit can carry out data transmission with the cloud platform, and when the staff breaks rules and regulations, the cloud platform can also send an alarm to the terminal equipment, suspend the work ticket flow, and cannot carry out subsequent steps until the safety supervisor eliminates risks to confirm construction safety.
Referring to fig. 3, a schematic diagram of an embodiment of a method for identifying risk behaviors of a cross-modal power operation site provided in the present invention is shown. The method comprises an image target recognition algorithm and a risk behavior recognition algorithm:
the image target recognition algorithm is used for recognizing and extracting various static characteristic elements in the power operation working scene. For example, images taken by a camera detect whether the protective equipment of the staff is fully worn.
The risk behavior recognition algorithm is used for recognizing risk behaviors which cannot be completed by a single image. The algorithm comprises a cross-modal behavior recognition algorithm and an indoor and outdoor integrated precise positioning algorithm. The cross-modal behavior recognition algorithm is used for recognizing the behavior and motion states of the workers and analyzing whether the behavior of the workers meets the safety standard or not; the indoor and outdoor integrated precise positioning algorithm is used for determining and tracking the real-time position of the worker, and analyzing whether the position of the worker meets the safety standard. And according to the pre-defined dangerous area and the working area, when the staff enters the dangerous area or leaves the working post, a warning is given.
Referring to fig. 4, a schematic diagram of an embodiment of a network structure of an image target recognition algorithm provided in the present invention is shown. The method comprises the following steps:
convolutional network
Figure 607587DEST_PATH_IMAGE038
To (1) a
Figure 471638DEST_PATH_IMAGE039
A convolutional layer can be represented as
Figure 348327DEST_PATH_IMAGE040
Figure 905210DEST_PATH_IMAGE041
Representing the input tensor in its entirety,
Figure 437823DEST_PATH_IMAGE042
representsThe output tensor i is the number of the convolutional layer. The whole network is composed of
Figure 715352DEST_PATH_IMAGE043
A convolutional layer composition, which can be expressed as
Figure 587493DEST_PATH_IMAGE044
Can be simplified into
Figure 439911DEST_PATH_IMAGE045
And j is the number of the convolutional layer. In practical applications, a plurality of convolutional layers with the same structure will be generally referred to as a phase, for example, ResNet has 5 phases, each convolutional layer in a phase has the same structure, and a convolutional network can be expressed by taking a phase as a unit as follows:
Figure 194240DEST_PATH_IMAGE046
wherein
Figure 665673DEST_PATH_IMAGE047
Is shown as
Figure 267687DEST_PATH_IMAGE039
The dimensions of the layer tensor are such that,
Figure 900793DEST_PATH_IMAGE048
is shown as
Figure 407998DEST_PATH_IMAGE039
A stage consisting of a convolutional layer
Figure 542176DEST_PATH_IMAGE049
Repetition of
Figure 123330DEST_PATH_IMAGE050
And (5) secondary composition. The design of networks in general focuses on finding the best network layer
Figure 65354DEST_PATH_IMAGE049
Larger networks have larger model widths and depths, often yielding higher accuracy, butThe accuracy gain saturates rapidly after reaching a certain value, which illustrates the limitation of only expanding a single dimension. The dimensions of the model expansion are not completely independent, which means that the expansion dimensions need to be balanced.
The image target recognition algorithm can automatically search the depth and width of the optimal demodulation whole neural network through a group of composite coefficients, and is different from the traditional method that the structure of the network is manually adjusted, and the method can search a group of optimal parameters which best accord with the network. The depth of the network model can be expressed as
Figure 997538DEST_PATH_IMAGE051
The width of the network model can be expressed as
Figure 607511DEST_PATH_IMAGE052
Wherein
Figure 698964DEST_PATH_IMAGE053
And
Figure 408294DEST_PATH_IMAGE054
as constants that can be found by the search for optimum,
Figure 968719DEST_PATH_IMAGE055
are manually adjustable parameters. When it is satisfied with
Figure 116804DEST_PATH_IMAGE056
Can adaptively obtain the parameters of the network, wherein
Figure 938129DEST_PATH_IMAGE057
Is a set parameter. The method for solving the optimal parameters with the constraints can be realized by adjusting the parameters
Figure 942994DEST_PATH_IMAGE058
A control model when
Figure 849770DEST_PATH_IMAGE059
When obtain oneA minimum optimal base model; when increasing
Figure 880174DEST_PATH_IMAGE058
When two dimensions of the model are simultaneously enlarged, the model is enlarged, the performance is improved, and the consumed resources are enlarged.
In fusing different input features, previous research on FPN and some improvements to FPN mostly simply adds features. In practice these different input features have different resolutions and their contributions to the fused output features are often found to be unequal by experimental observation. In order to solve the problem, the image target detection algorithm adopts a multi-scale data fusion method. Where the algorithm input is not limited to a fixed resolution, it may be data of any resolution. The multi-scale data fusion method adopts an attention feature pyramid structure, and different attention weights are introduced into each layer of the feature pyramid, so that feature contribution ratios of different layers can be adjusted while input data of different scales are fused. The attention weight is calculated by the formula
Figure 352744DEST_PATH_IMAGE060
Wherein
Figure 935035DEST_PATH_IMAGE061
A scalar quantity of the features or a vector quantity of a certain channel, and also can be a multi-dimensional tensor; where i is the serial number of the current input feature, j is the serial number of the non-output feature, ϵ = 0.0001. To ensure that the weight is greater than 0, a ReLU activation function is used before the weight calculation step.
Referring to fig. 5, a schematic diagram of an embodiment of a network structure of a cross-modal behavior recognition algorithm provided in the present invention is shown. The input data of the cross-modal behavior recognition algorithm are data of a plurality of different modalities, including a two-dimensional image acquired by a common camera, a depth image acquired by a depth camera, and an acceleration signal acquired by an acceleration sensor. The cross-modal behavior recognition algorithm comprises the following steps:
step one, carrying out data processing on the acceleration signals in a period of continuous time and extracting features through a neural network model. And extracting the human body characteristics of the staff from the depth image at the ending moment through a depth neural network. And extracting background element characteristics of the two-dimensional image at the ending moment through a deep neural network. Inputting all the characteristics into a neural network discrimination model, judging the behavior of the worker through the acceleration signal characteristics and the depth image characteristics, and judging whether the behavior of the worker meets the safety standard or not by combining the background characteristics of the two-dimensional image.
For example, in the high-voltage electroscope operation, when the worker does not lift the high-voltage electroscope rod, or the worker only has the hand-lifting action but does not have the high-voltage electroscope rod in the hand, the worker determines the illegal action.
Fig. 6 is a schematic diagram of an embodiment of an indoor and outdoor integrated precise positioning algorithm provided in the present invention. The method comprises the following steps:
the indoor and outdoor integrated precise positioning algorithm provides a high-accuracy positioning scheme through combination of a Beidou satellite and an ultra-wideband technology. The Beidou satellite provides large-range positioning in an outdoor scene, and the ultra-wideband technology provides outdoor small-range or indoor high-accuracy positioning. The two positioning techniques are combined as follows:
step one, a worker wears a device provided with an ultra-wideband tag and enters a working scene provided with a plurality of ultra-wideband base stations, and the scheme of 4 base stations is adopted in the embodiment of the invention. And measuring the distance between the base station and the tag by using a bidirectional time flight method according to the time consumed by data transmission between the ultra-wideband base station and the tag. Determining the relative position of a worker and a base station through a neural network training model according to the distance relationship between the base station and the label;
and step two, measuring the absolute position of the ultra-wideband base station by a Beidou satellite positioning technology. By using a real-time dynamic algorithm of multiple base stations, each fixed reference station does not directly send any correction information to a mobile user, but all original data is sent to a data processing center through a data communication link, a mobile station sends the approximate coordinates of the mobile station to a control center, the data management center broadcasts a group of fixed optimal reference stations after receiving position information, and a virtual reference station is simulated near the position of the user by using the data of the fixed optimal reference stations to form an ultra-short baseline with the mobile station. And performing auxiliary fusion analysis on all the data to determine the absolute position of the worker in the map.
Referring to fig. 7, the schematic diagram of an embodiment of the cross-modal power operation site risk behavior identification device provided in the present invention includes an equipment box, an intelligent safety helmet, and a positioning base station:
the equipment box comprises an algorithm operation platform, a camera, a depth camera, a wireless communication equipment receiving end and alarm equipment. The camera is in wired connection with the algorithm operation platform through a universal serial bus, and the sensor is in wireless connection with the algorithm operation platform through ZigBee. Acquiring image information through a common camera and a depth camera, acquiring sensor information through a receiving end of wireless communication equipment, analyzing data through an algorithm operating platform to judge whether the operation of a worker meets the standard, and sending out a warning by warning equipment when dangerous behaviors are detected;
the intelligent safety helmet is provided with an acceleration sensor, a Beidou positioning chip, an ultra-wideband positioning label chip, a power supply module and a wireless transmission module sending end. The wireless transmission module sending end is used for sending the acquired sensing information to the equipment box in real time for data analysis;
the positioning base station is composed of an ultra-wideband positioning base station module, a Beidou positioning module and a power supply module. The positioning base station needs to be dispersedly arranged in a working site, and the periphery of the positioning base station cannot be shielded by metal or concrete buildings. In order to ensure the positioning accuracy, the arrangement position of the positioning base station is open enough.
The method for identifying the risk behaviors of the cross-modal power operation site in the power production can accurately monitor the positions and behavior states of all workers on the site in real time, and can give an alarm in time when abnormal behaviors such as approaching a dangerous area, leaving a working post, contacting a charged body illegally and the like occur. Compared with the existing production mode of manual observation, the efficiency is greatly improved, the safety of electric power production is improved, and the method has high practicability.
Fig. 8 is a diagram of a system for identifying risk behaviors in accordance with a preferred embodiment of the present invention. As shown in fig. 8, the present invention provides a system for identifying risk behaviors, the system comprising:
the first obtaining unit 801 is configured to select acceleration signals of a target object in a continuous time period when the target object performs a target behavior, and obtain an acceleration feature vector of the acceleration signals in the continuous time period through a deep neural network model.
The second obtaining unit 802 is configured to select a depth image of the target object at the target behavior end time, and obtain a target object feature vector of the target object in the depth image through the depth neural network model.
A third obtaining unit 803, configured to select a two-dimensional image of the target object at the end time of the target behavior, and obtain, through the deep neural network model, a background element feature of the target object in the two-dimensional image.
A fourth obtaining unit 804, configured to input the acceleration feature vector and the target object feature vector into the neural network discrimination model, and obtain a behavior action of the target object.
And the result unit 805 is configured to determine whether the behavior action of the target object meets a preset safety behavior standard based on the background element characteristics.
Preferably, the system further comprises a fifth obtaining unit for determining the depth and width of the deep neural network model;
the depth of the deep neural network model is as follows:
Figure 188162DEST_PATH_IMAGE062
the width of the deep neural network model is:
Figure 881311DEST_PATH_IMAGE063
wherein
Figure 473967DEST_PATH_IMAGE064
And
Figure 102525DEST_PATH_IMAGE065
to find the constants by the method of searching the optimum values,
Figure 452735DEST_PATH_IMAGE066
is a manually adjusted parameter;
when the constraint condition is satisfied
Figure 74209DEST_PATH_IMAGE067
Adaptively obtaining parameters of the deep neural network model, wherein
Figure 724634DEST_PATH_IMAGE068
Is a set parameter. Wherein the parameter is a learning rate, so that the objective function converges to a local minimum.
Preferably, a fifth obtaining unit is further included, configured to determine attention weights of the deep neural network model:
determining that different attention weights are set for each layer of the attention feature pyramid structure and respectively correspond to feature contribution ratios of two-dimensional images with different resolutions;
the attention weight calculation formula is:
Figure 445465DEST_PATH_IMAGE069
wherein
Figure 710003DEST_PATH_IMAGE070
A vector of scalar or two-dimensional image channels that are features; where i is the serial number of the current input feature, j is the serial number of the non-output feature, ϵ = 0.0001. The feature is a semantic feature of the two-dimensional image, and comprises color, texture, shape, spatial relationship, attribute feature and the like.
Preferably, the system further comprises a sixth acquisition unit for determining the position of the target object: acquiring the time consumed by data transmission between a preset ultra-wideband base station and an ultra-wideband tag of a target object, and measuring the distance between the ultra-wideband base station and the tag by a two-way time flight method based on the time consumed by the data transmission;
determining the relative position of the target object and the ultra-wideband base station based on the distance between the ultra-wideband base station and the label;
positioning the ultra-wideband base station, and determining the absolute position of the ultra-wideband base station;
determining the absolute position of the target object based on the absolute position of the ultra-wideband base station and the relative position of the target object and the ultra-wideband base station;
and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the absolute position and the background element characteristics of the target object.
The system 800 for identifying a risk behavior according to the preferred embodiment of the present invention corresponds to the method 100 for identifying a risk behavior according to the preferred embodiment of the present invention, and will not be described herein again.
The invention provides a computer-readable storage medium, characterized in that the computer-readable storage medium stores a computer program for executing the above-mentioned method for identifying a risk behaviour.
The present invention provides an electronic device, characterized in that the electronic device includes:
a processor;
a memory for storing the processor-executable instructions;
the processor is used for reading the executable instructions from the memory and executing the method for identifying the risk behaviors.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, system, or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
It should be noted that the above-mentioned embodiments are only for illustrating the technical solutions of the present invention and not for limiting the protection scope thereof, and although the present invention has been described in detail with reference to the above-mentioned embodiments, those skilled in the art should understand that after reading the present invention, they can make various changes, modifications or equivalents to the specific embodiments of the present invention, but these changes, modifications or equivalents are within the protection scope of the appended claims.
The invention has been described with reference to a few embodiments. However, other embodiments of the invention than the one disclosed above are equally possible within the scope of the invention, as would be apparent to a person skilled in the art from the appended patent claims.
Generally, all terms used in the claims are to be interpreted according to their ordinary meaning in the technical field, unless explicitly defined otherwise herein. All references to "a/an/the [ device, component, etc ]" are to be interpreted openly as referring to at least one instance of said device, component, etc., unless explicitly stated otherwise. The steps of any method disclosed herein do not have to be performed in the exact order disclosed, unless explicitly stated.

Claims (10)

1. A method for identifying risk behaviors, the method comprising:
selecting acceleration signals of a target object in a continuous time period when the target object performs a target behavior, and acquiring acceleration characteristic vectors of the acceleration signals in the continuous time period through a deep neural network model;
selecting a depth image of a target object at the end moment of a target behavior, and acquiring a target object feature vector of the target object in the depth image through a depth neural network model;
selecting a two-dimensional image of a target object at the end moment of a target behavior, and acquiring background element characteristics of the target object in the two-dimensional image through a deep neural network model;
inputting the acceleration characteristic vector and the target object characteristic vector into the neural network discrimination model to obtain the behavior action of the target object;
and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the background element characteristics.
2. The method of claim 1, further comprising determining a depth and a width of the deep neural network model:
the depth of the deep neural network model is as follows:
Figure 354376DEST_PATH_IMAGE001
the width of the deep neural network model is as follows:
Figure 373148DEST_PATH_IMAGE002
wherein
Figure 520095DEST_PATH_IMAGE003
And
Figure 515471DEST_PATH_IMAGE004
to find the constants by the method of searching the optimum values,
Figure 431474DEST_PATH_IMAGE005
is a manually adjusted parameter;
when the constraint condition is satisfied
Figure 89989DEST_PATH_IMAGE006
Adaptively obtaining parameters of the deep neural network model, wherein
Figure 724232DEST_PATH_IMAGE007
Is a set parameter.
3. The method of claim 1, further comprising: determining attention weights for the deep neural network model:
determining an attention feature pyramid structure, wherein each layer of the attention feature pyramid structure is provided with different attention weights, and the attention feature weights respectively correspond to feature contribution ratios of the two-dimensional images with different resolutions;
the attention weight calculation formula is as follows:
Figure 759184DEST_PATH_IMAGE008
wherein
Figure 795273DEST_PATH_IMAGE009
A vector of scalar or two-dimensional image channels that are features; wherein
Figure 890268DEST_PATH_IMAGE010
4. The method of claim 1, before determining whether the behavior action of the target object meets a preset safety behavior standard based on the background element feature, further comprising: determining a position of the target object:
acquiring the time consumed by data transmission between a preset ultra-wideband base station and an ultra-wideband tag of the target object, and measuring the distance between the ultra-wideband base station and the tag by a two-way time flight method based on the time consumed by data transmission;
determining a relative position of the target object and the ultra-wideband base station based on a distance between the ultra-wideband base station and the tag;
positioning the ultra-wideband base station, and determining the absolute position of the ultra-wideband base station;
determining an absolute position of the target object based on the absolute position of the ultra-wideband base station and a relative position of the target object and the ultra-wideband base station;
based on the background element characteristics, judging whether the behavior action of the target object meets a preset safety behavior standard or not, including:
and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the absolute position of the target object and the background element characteristics.
5. A system for identifying risk behaviors, the system comprising:
the first acquisition unit is used for selecting acceleration signals of a target object in a continuous time period when the target object performs a target behavior, and acquiring acceleration characteristic vectors of the acceleration signals in the continuous time period through a deep neural network model;
the second acquisition unit is used for selecting a depth image of a target object at the end moment of a target behavior and acquiring a target object feature vector of the target object in the depth image through a depth neural network model;
the third acquisition unit is used for selecting a two-dimensional image of a target object at the end moment of a target behavior and acquiring background element characteristics of the target object in the two-dimensional image through a deep neural network model;
a fourth obtaining unit, configured to input the acceleration feature vector and the target object feature vector into the neural network discrimination model, and obtain a behavior action of the target object;
and the result unit is used for judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the background element characteristics.
6. The system of claim 5, further comprising a fifth acquisition unit for determining a depth and a width of the deep neural network model;
the depth of the deep neural network model is as follows:
Figure 746229DEST_PATH_IMAGE011
the width of the deep neural network model is as follows:
Figure 850451DEST_PATH_IMAGE002
wherein
Figure 209888DEST_PATH_IMAGE012
And
Figure 741364DEST_PATH_IMAGE004
to find the constants by the method of searching the optimum values,
Figure 317576DEST_PATH_IMAGE005
is a manually adjusted parameter;
when the constraint condition is satisfied
Figure 225490DEST_PATH_IMAGE006
Adaptively obtaining parameters of the deep neural network model, wherein
Figure 705013DEST_PATH_IMAGE007
Is a set parameter.
7. The system of claim 5, further comprising a fifth obtaining unit for determining attention weights of the deep neural network model:
determining an attention feature pyramid structure, wherein each layer of the attention feature pyramid structure is provided with different attention weights, and the attention feature weights respectively correspond to feature contribution ratios of the two-dimensional images with different resolutions;
the attention weight calculation formula is as follows:
Figure 407389DEST_PATH_IMAGE013
wherein
Figure 972363DEST_PATH_IMAGE009
A vector of scalar or two-dimensional image channels that are features; wherein
Figure 152808DEST_PATH_IMAGE010
8. The system of claim 5, further comprising a sixth acquisition unit for determining the position of the target object:
acquiring the time consumed by data transmission between a preset ultra-wideband base station and an ultra-wideband tag of the target object, and measuring the distance between the ultra-wideband base station and the tag by a two-way time flight method based on the time consumed by data transmission;
determining a relative position of the target object and the ultra-wideband base station based on a distance between the ultra-wideband base station and the tag;
positioning the ultra-wideband base station, and determining the absolute position of the ultra-wideband base station;
determining an absolute position of the target object based on the absolute position of the ultra-wideband base station and a relative position of the target object and the ultra-wideband base station;
and judging whether the behavior action of the target object meets a preset safety behavior standard or not based on the absolute position of the target object and the background element characteristics.
9. A computer-readable storage medium, characterized in that it stores a computer program for performing the method of any of the preceding claims 1-4.
10. An electronic device, characterized in that the electronic device comprises:
a processor;
a memory for storing the processor-executable instructions;
the processor is used for reading the executable instructions from the memory and executing the instructions to realize the method of any one of the claims 1 to 4.
CN202111029514.6A 2021-09-03 2021-09-03 Method and system for identifying risk behaviors Active CN113469150B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111029514.6A CN113469150B (en) 2021-09-03 2021-09-03 Method and system for identifying risk behaviors

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111029514.6A CN113469150B (en) 2021-09-03 2021-09-03 Method and system for identifying risk behaviors

Publications (2)

Publication Number Publication Date
CN113469150A true CN113469150A (en) 2021-10-01
CN113469150B CN113469150B (en) 2021-11-12

Family

ID=77867222

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111029514.6A Active CN113469150B (en) 2021-09-03 2021-09-03 Method and system for identifying risk behaviors

Country Status (1)

Country Link
CN (1) CN113469150B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114049729A (en) * 2021-10-20 2022-02-15 上海市安装工程集团有限公司 Monitoring and safety alarm system for dangerous protection area of construction site and use method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109171738A (en) * 2018-07-13 2019-01-11 杭州电子科技大学 Fall detection method based on human body acceleration multiple features fusion and KNN
CN110555406A (en) * 2019-08-31 2019-12-10 武汉理工大学 Video moving target identification method based on Haar-like characteristics and CNN matching
CN111144174A (en) * 2018-11-02 2020-05-12 浙江树人学院 System for identifying falling behavior of old people in video by using neural network and traditional algorithm

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109171738A (en) * 2018-07-13 2019-01-11 杭州电子科技大学 Fall detection method based on human body acceleration multiple features fusion and KNN
CN111144174A (en) * 2018-11-02 2020-05-12 浙江树人学院 System for identifying falling behavior of old people in video by using neural network and traditional algorithm
CN110555406A (en) * 2019-08-31 2019-12-10 武汉理工大学 Video moving target identification method based on Haar-like characteristics and CNN matching

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
TAO ZHEN ET AL.: "Walking Gait Phase Detection Based on Acceleration Signals Using LSTM-DNN Algorithm", 《ALGORITHMS》 *
徐川龙 等: "一种基于三维加速度传感器的人体行为识别方法", 《计算机系统应用》 *
金淼 等: "参数自适应去阴影前景目标检测算法", 《华中科技大学学报(自然科学版)》 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114049729A (en) * 2021-10-20 2022-02-15 上海市安装工程集团有限公司 Monitoring and safety alarm system for dangerous protection area of construction site and use method

Also Published As

Publication number Publication date
CN113469150B (en) 2021-11-12

Similar Documents

Publication Publication Date Title
Fang et al. Computer vision aided inspection on falling prevention measures for steeplejacks in an aerial environment
Chen et al. Towards on-site hazards identification of improper use of personal protective equipment using deep learning-based geometric relationships and hierarchical scene graph
CN107358146B (en) Method for processing video frequency, device and storage medium
CN109766755B (en) Face recognition method and related product
CN110458895A (en) Conversion method, device, equipment and the storage medium of image coordinate system
Jin et al. DWCA‐YOLOv5: An Improve Single Shot Detector for Safety Helmet Detection
CN107341443B (en) Method for processing video frequency, device and storage medium
CN110659391A (en) Video detection method and device
CN106878670A (en) A kind of method for processing video frequency and device
CN109740444A (en) Flow of the people information displaying method and Related product
Wang et al. Vision-based hand signal recognition in construction: A feasibility study
Wang et al. A safety helmet and protective clothing detection method based on improved-yolo v 3
CN114677640A (en) Intelligent construction site safety monitoring system and method based on machine vision
CN106971142A (en) A kind of image processing method and device
CN113469150B (en) Method and system for identifying risk behaviors
CN116152863A (en) Personnel information identification method and device, electronic equipment and storage medium
CN117173791A (en) Distribution network constructor violation detection method and system based on action recognition
CN117115728A (en) Risk identification method and system applied to field operation of transformer substation
Xu et al. A novel computer vision‐based approach for monitoring safety harness use in construction
Chang et al. Safety risk assessment of electric power operation site based on variable precision rough set
CN113627321A (en) Image identification method and device based on artificial intelligence and computer equipment
CN207020820U (en) Red-running Monitor System
CN107301373A (en) Data processing method, device and storage medium
CN116524428A (en) Electric power operation safety risk identification method based on target detection and scene fusion
CN114758286B (en) Intelligent edge safety monitoring method and device based on work ticket event

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant