CN113642474A - Hazardous area personnel monitoring method based on YOLOV5 - Google Patents

Hazardous area personnel monitoring method based on YOLOV5 Download PDF

Info

Publication number
CN113642474A
CN113642474A CN202110941137.7A CN202110941137A CN113642474A CN 113642474 A CN113642474 A CN 113642474A CN 202110941137 A CN202110941137 A CN 202110941137A CN 113642474 A CN113642474 A CN 113642474A
Authority
CN
China
Prior art keywords
deep learning
network model
learning network
monitoring
yolov5
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110941137.7A
Other languages
Chinese (zh)
Inventor
窦涛
沈雪松
陆金波
吴昊翰
贺荣鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan Aerospace Electro & Hydraulic Control Co ltd
Original Assignee
Sichuan Aerospace Electro & Hydraulic Control Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan Aerospace Electro & Hydraulic Control Co ltd filed Critical Sichuan Aerospace Electro & Hydraulic Control Co ltd
Priority to CN202110941137.7A priority Critical patent/CN113642474A/en
Publication of CN113642474A publication Critical patent/CN113642474A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The invention discloses a method for monitoring personnel in a dangerous area based on YOLOV5, which comprises the following steps: step 1, constructing a deep learning model training set: acquiring images from an underground monitoring video stream, and carrying out sample annotation on the images; step 2, training a deep learning network model based on YOLOV5 by using the sample marked in the step 1, and step 3, constructing a test data set, transmitting the test data set into the trained deep learning network model for testing, judging whether the deep learning network model meets the engineering requirements, if so, finishing the training, otherwise, modifying the network parameters for training again; and 4, identifying dangerous behaviors of the personnel in the video monitoring under the mine by using the trained deep learning network model, and outputting an identification result. The invention introduces more feature extraction and self-adaptive modules on a stable-based Yolov5 framework, can be effectively applied to the actual personnel behavior monitoring task, and greatly improves the monitoring data analysis capability.

Description

Hazardous area personnel monitoring method based on YOLOV5
Technical Field
The invention belongs to the technical field of image detection and image segmentation, and particularly relates to a method for monitoring personnel in a dangerous area based on YOLOV 5.
Background
The target detection is a basic algorithm in the field of universal identity recognition, is a core part of an intelligent monitoring system and an important branch of image processing and computer vision disciplines, and plays a vital role in subsequent tasks such as face recognition, gait recognition, crowd counting, instance segmentation and the like. Due to the wide application of deep learning, the target detection algorithm is developed rapidly.
The YOLO-based target detection algorithm redefines object detection as a regression problem. It applies a single Convolutional Neural Network (CNN) to the entire image, divides the image into meshes, and predicts class probabilities and bounding boxes for each mesh.
The target detection algorithm applies a method called "Non Max Suppression" (Non Max Suppression) to the objects of each class to filter out the bounding box with "confidence" less than a threshold, which can predict the image.
The bounding box of the target detection can be described by using four descriptors, namely: the center, height, width, value of the bounding box map to the class to which the object belongs.
Dark image enhancement algorithms are used to enhance useful information in an image with the aim of improving the visual impact of the image. The current common algorithms for image enhancement are as follows: histogram equalization, histogram specification, image enhancement algorithm based on physical model, image enhancement algorithm based on partial differential equation and variation, and change domain image enhancement algorithm, etc. with the development of technology, the algorithm is continuously improved.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provide a method for monitoring personnel in a dangerous area based on Yolov5, which introduces more feature extraction and self-adaption modules on a stable-base Yolov5 framework, can be effectively applied to an actual personnel behavior monitoring task and can greatly improve the data analysis capability of monitoring.
The purpose of the invention is realized by the following technical scheme: a hazardous area personnel monitoring method based on YOLOV5 comprises the following steps:
step 1, constructing a deep learning model training set: acquiring images from an underground monitoring video stream, and carrying out sample annotation on the images;
step 2, training a deep learning network model based on YOLOV5 by using the samples marked in the step 1,
step 3, constructing a test data set, introducing the trained deep learning network model for testing, judging whether the deep learning network model meets the engineering requirements, if so, finishing the training, otherwise, modifying the network parameters and training again;
and 4, identifying dangerous behaviors of the personnel in the video monitoring under the mine by using the trained deep learning network model, and outputting an identification result.
Further, the step 1 comprises the following substeps:
step 11, obtaining images from the mine monitoring video stream, and preprocessing the images: enhancing the image with the brightness lower than a set threshold value; then, randomly turning horizontally/vertically, rotating and cutting the image;
and step 12, carrying out sample marking on the processed images, and carrying out pre-marking on the position information of the personnel on each image.
Further, the deep learning network model consists of three main components:
1) a Backbone component: aggregating and forming a convolutional neural network of image features on different image fine granularities;
2) a pack of the Neck component: a series of network layers that blend and combine image features and pass the image features to a prediction layer;
3) prediction component: and predicting the image characteristics, generating a boundary box and predicting the category.
Further, the Backbone component comprises a Focus layer, a CBL module, a CSP _ A module, an AC-Block module, a CBL module and an SPP module which are connected in sequence.
Further, in the step 3, the specific method for judging whether the deep learning network model meets the engineering requirement is as follows: judging whether the deep learning network model can accurately identify the staff in the video picture; judging whether the deep learning network model can accurately identify the activity area of the worker or not; judging whether the deep learning network model can accurately judge whether the worker enters a dangerous area in the picture or not; and judging whether the deep learning network model can give real-time early warning feedback to the violation in time.
The invention has the beneficial effects that: the invention introduces more feature extraction and self-adaptive modules on a stable-based Yolov5 framework, can be effectively applied to actual personnel behavior monitoring tasks, and can greatly improve the monitored data analysis capability. By utilizing a key area monitoring strategy, the personnel activity condition in a key area (dangerous area) can be efficiently and accurately concerned, and early warning feedback is made.
Drawings
FIG. 1 is a flow chart of a hazardous area personnel monitoring method based on YOLOV5 of the present invention;
fig. 2 is a structural diagram of the deep learning network model based on YOLOV5 according to the present invention.
Detailed Description
The technical scheme of the invention is further explained by combining the attached drawings.
As shown in fig. 1, the hazardous area personnel monitoring method based on YOLOV5 of the present invention includes the following steps:
step 1, constructing a deep learning model training set: acquiring images from an underground monitoring video stream, and carrying out sample annotation on the images; the method specifically comprises the following substeps:
step 11, obtaining images from the mine monitoring video stream, and preprocessing the images: the image with the brightness lower than the set threshold value is enhanced, so that the data set can better meet the training requirement, and the high-precision prediction can be made on the video acquired in a dark environment; then, random horizontal/vertical turning, rotation and cutting operations are carried out on the image, the sample capacity of a training set is increased, and the generalization capability and robustness of the model are increased;
and step 12, performing sample labeling on the processed images, performing pre-labeling on the position information of the personnel on each image, and using the pre-labeled position information as label data for model training and parameter optimization.
Step 2, training a deep learning network model based on YOLOV5 by using the samples marked in the step 1, applying an efficient and practical feature extraction module to the deep model, transmitting the processed image data into the training model for training, and adjusting a network optimization strategy to enable the performance index of the model to be optimal; on a stable basis of the YOLOV5 framework, more feature extraction and adaptation modules are introduced, and a specific network model is shown in fig. 2. The deep learning network model of the invention is composed of three main components:
1) a Backbone component: aggregating and forming a convolutional neural network of image features on different image fine granularities;
2) a pack of the Neck component: a series of network layers that blend and combine image features and pass the image features to a prediction layer;
3) prediction component: and predicting the image characteristics, generating a boundary box and predicting the category.
The Backbone component comprises a Focus layer, a CBL module, a CSP _ A module, an AC-Block module, a CBL module and an SPP module which are sequentially connected. According to the invention, an AC-Block module (asymmetric convolution structure) is added behind each CSP _ A module of the backhaul component, so that the feature extraction of the convolution on the central position can be enhanced, the defect of low detection accuracy rate due to poor feature extraction capability in the prior art when the video quality is poor is overcome, and the detection and evidence obtaining accuracy rate is improved. The rest of the network structure is the same as the existing YOLOV5 network model.
The Focus layer is used for copying an input image into four parts and then cutting the four pictures into four slices through a slicing operation;
the CSP _ A module is firstly processed by a convolution structure (Conv + Batch _ norm + Leaky relu, CBL) of 1x1, then added with the initial input by a residual structure, and finally spliced with the initial input by the residual structure.
The Neck component adopts an FPN + PAN structure, the FPN layers transmit strong semantic features from top to bottom, the feature information of the high layers is transmitted and fused in an up-sampling mode, and feature graphs of 76 x 76, 38 x 38 and 19 x 19 are output; next, a PAN structure is connected, strong positioning features are transmitted from bottom to top, feature graphs of 76 x 76, 38 x 38 and 19 x 19 are output in a down-sampling mode, concatee operation is carried out on the feature graphs and the output graphs of the FPN structure, parameter aggregation is carried out on different detection layers from different trunk layers, and the feature extraction capability is further improved;
the FA module is divided into a channel feature attention CFA and a pixel feature attention PFA, wherein the CFA is responsible for distributing different weights to feature maps of different channels; the pixel attention is then responsible for giving different weights of attention to different regions of the image.
The model of the invention can be effectively applied to the actual personnel behavior monitoring task.
Step 3, constructing a test data set, introducing the trained deep learning network model for testing, judging whether the deep learning network model meets the engineering requirements, if so, finishing the training, otherwise, modifying the network parameters and training again;
the specific method for judging whether the deep learning network model meets the engineering requirements comprises the following steps: judging whether the deep learning network model can accurately identify the staff in the video picture; judging whether the deep learning network model can accurately identify the activity area of the worker or not; judging whether the deep learning network model can accurately judge whether the worker enters a dangerous area in the picture or not; and judging whether the deep learning network model can give real-time early warning feedback to the violation in time.
And 4, identifying dangerous behaviors of the personnel in the video monitoring under the mine by using the trained deep learning network model, and outputting an identification result.
And constructing an overall monitoring scheme of the system, formulating a dangerous area personnel monitoring strategy, adopting the deep learning network model to identify personnel dangerous behavior information in the mine video monitoring, and feeding back a corresponding early warning task, thereby realizing the dangerous area personnel monitoring task. The monitoring scheme and strategy of the specific implementation of the invention are as follows: 1) carrying out video monitoring on the position where the dangerous area exists; 2) monitoring the personnel activity area of the returned video pictures by using the computation network model; 3) carrying out accuracy evaluation on the data result detected by the model; 4) judging whether the person in the picture enters an activity area or not by using an algorithm; 5) for violation behaviors, the system automatically implements early warning measures.
Meanwhile, danger feedback signals of the system are fed back through multi-dimensional information, and the danger feedback signals comprise information feedback of specific positions of people, red early warning feedback of abnormal pictures of a system software interface, sound warning feedback of an early warning device and the like. The invention can carry out various violation analyses on dangerous behaviors of personnel in the video pictures, namely system information such as corresponding positions, specific activity areas, corresponding monitoring camera pictures and the like of the violation personnel is obtained through an algorithm, the violation behaviors are correctly classified by comparing the system information with a behavior database, and corresponding early warning feedback is implemented.
It will be appreciated by those of ordinary skill in the art that the embodiments described herein are intended to assist the reader in understanding the principles of the invention and are to be construed as being without limitation to such specifically recited embodiments and examples. Those skilled in the art can make various other specific changes and combinations based on the teachings of the present invention without departing from the spirit of the invention, and these changes and combinations are within the scope of the invention.

Claims (5)

1. A hazardous area personnel monitoring method based on YOLOV5 is characterized by comprising the following steps:
step 1, constructing a deep learning model training set: acquiring images from an underground monitoring video stream, and carrying out sample annotation on the images;
step 2, training a deep learning network model based on YOLOV5 by using the samples marked in the step 1,
step 3, constructing a test data set, introducing the trained deep learning network model for testing, judging whether the deep learning network model meets the engineering requirements, if so, finishing the training, otherwise, modifying the network parameters and training again;
and 4, identifying dangerous behaviors of the personnel in the video monitoring under the mine by using the trained deep learning network model, and outputting an identification result.
2. The YOLOV 5-based hazardous area personnel monitoring method of claim 1, wherein the step 1 comprises the following substeps:
step 11, obtaining images from the mine monitoring video stream, and preprocessing the images: enhancing the image with the brightness lower than a set threshold value; then, randomly turning horizontally/vertically, rotating and cutting the image;
and step 12, carrying out sample marking on the processed images, and carrying out pre-marking on the position information of the personnel on each image.
3. The method for hazardous area personnel monitoring based on YOLOV5 of claim 1, wherein the deep learning network model consists of three main components:
1) a Backbone component: aggregating and forming a convolutional neural network of image features on different image fine granularities;
2) a pack of the Neck component: a series of network layers that blend and combine image features and pass the image features to a prediction layer;
3) prediction component: and predicting the image characteristics, generating a boundary box and predicting the category.
4. The YOLOV 5-based dangerous area personnel monitoring method according to claim 1, wherein the Backbone component comprises a Focus layer, a CBL module, a CSP _ A module, an AC-Block module, a CBL module and an SPP module which are connected in sequence.
5. The method for monitoring personnel in dangerous areas based on YOLOV5 of claim 1, wherein the specific method for determining whether the deep learning network model meets the engineering requirements in step 3 is: judging whether the deep learning network model can accurately identify the staff in the video picture; judging whether the deep learning network model can accurately identify the activity area of the worker or not; judging whether the deep learning network model can accurately judge whether the worker enters a dangerous area in the picture or not; and judging whether the deep learning network model can give real-time early warning feedback to the violation in time.
CN202110941137.7A 2021-08-17 2021-08-17 Hazardous area personnel monitoring method based on YOLOV5 Pending CN113642474A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110941137.7A CN113642474A (en) 2021-08-17 2021-08-17 Hazardous area personnel monitoring method based on YOLOV5

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110941137.7A CN113642474A (en) 2021-08-17 2021-08-17 Hazardous area personnel monitoring method based on YOLOV5

Publications (1)

Publication Number Publication Date
CN113642474A true CN113642474A (en) 2021-11-12

Family

ID=78422233

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110941137.7A Pending CN113642474A (en) 2021-08-17 2021-08-17 Hazardous area personnel monitoring method based on YOLOV5

Country Status (1)

Country Link
CN (1) CN113642474A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114724246A (en) * 2022-04-11 2022-07-08 中国人民解放军东部战区总医院 Dangerous behavior identification method and device
CN115041487A (en) * 2022-05-18 2022-09-13 深圳能源环保股份有限公司 Fly ash landfill comprehensive management method and system based on high-precision outdoor positioning
CN115294661A (en) * 2022-10-10 2022-11-04 青岛浩海网络科技股份有限公司 Pedestrian dangerous behavior identification method based on deep learning
CN116206255A (en) * 2023-01-06 2023-06-02 广州纬纶信息科技有限公司 Dangerous area personnel monitoring method and device based on machine vision
CN116740821A (en) * 2023-08-16 2023-09-12 南京迅集科技有限公司 Intelligent workshop control method and system based on edge calculation
CN117456610A (en) * 2023-12-21 2024-01-26 浪潮软件科技有限公司 Climbing abnormal behavior detection method and system and electronic equipment
CN117456610B (en) * 2023-12-21 2024-04-12 浪潮软件科技有限公司 Climbing abnormal behavior detection method and system and electronic equipment

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110457984A (en) * 2019-05-21 2019-11-15 电子科技大学 Pedestrian's attribute recognition approach under monitoring scene based on ResNet-50
CN112966589A (en) * 2021-03-03 2021-06-15 中润油联天下网络科技有限公司 Behavior identification method in dangerous area

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110457984A (en) * 2019-05-21 2019-11-15 电子科技大学 Pedestrian's attribute recognition approach under monitoring scene based on ResNet-50
CN112966589A (en) * 2021-03-03 2021-06-15 中润油联天下网络科技有限公司 Behavior identification method in dangerous area

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
YNXDB2002: "YOLO V5 模型结构及迁移学习", 《CSDN》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114724246A (en) * 2022-04-11 2022-07-08 中国人民解放军东部战区总医院 Dangerous behavior identification method and device
CN114724246B (en) * 2022-04-11 2024-01-30 中国人民解放军东部战区总医院 Dangerous behavior identification method and device
CN115041487A (en) * 2022-05-18 2022-09-13 深圳能源环保股份有限公司 Fly ash landfill comprehensive management method and system based on high-precision outdoor positioning
CN115294661A (en) * 2022-10-10 2022-11-04 青岛浩海网络科技股份有限公司 Pedestrian dangerous behavior identification method based on deep learning
CN116206255A (en) * 2023-01-06 2023-06-02 广州纬纶信息科技有限公司 Dangerous area personnel monitoring method and device based on machine vision
CN116206255B (en) * 2023-01-06 2024-02-20 广州纬纶信息科技有限公司 Dangerous area personnel monitoring method and device based on machine vision
CN116740821A (en) * 2023-08-16 2023-09-12 南京迅集科技有限公司 Intelligent workshop control method and system based on edge calculation
CN116740821B (en) * 2023-08-16 2023-10-24 南京迅集科技有限公司 Intelligent workshop control method and system based on edge calculation
CN117456610A (en) * 2023-12-21 2024-01-26 浪潮软件科技有限公司 Climbing abnormal behavior detection method and system and electronic equipment
CN117456610B (en) * 2023-12-21 2024-04-12 浪潮软件科技有限公司 Climbing abnormal behavior detection method and system and electronic equipment

Similar Documents

Publication Publication Date Title
CN113642474A (en) Hazardous area personnel monitoring method based on YOLOV5
CN110689054B (en) Worker violation monitoring method
WO2021051601A1 (en) Method and system for selecting detection box using mask r-cnn, and electronic device and storage medium
CN110969130B (en) Driver dangerous action identification method and system based on YOLOV3
CN111368690B (en) Deep learning-based video image ship detection method and system under influence of sea waves
CN111222478A (en) Construction site safety protection detection method and system
CN110298297A (en) Flame identification method and device
CN114648714A (en) YOLO-based workshop normative behavior monitoring method
CN112131951A (en) System for automatically identifying behaviors of illegal ladder use in construction
CN115797736B (en) Training method, device, equipment and medium for target detection model and target detection method, device, equipment and medium
CN109389105A (en) A kind of iris detection and viewpoint classification method based on multitask
CN114373162B (en) Dangerous area personnel intrusion detection method and system for transformer substation video monitoring
CN116977256A (en) Training method, device, equipment and storage medium for defect detection model
CN113642473A (en) Mining coal machine state identification method based on computer vision
CN114694090A (en) Campus abnormal behavior detection method based on improved PBAS algorithm and YOLOv5
CN114913233A (en) Image processing method, apparatus, device, medium, and product
CN115083229A (en) Intelligent recognition and warning system of flight training equipment based on AI visual recognition
CN114241401A (en) Abnormality determination method, apparatus, device, medium, and product
GB2605580A (en) Apparatus and method of image processing to detect a substance spill on a solid surface
CN112967335A (en) Bubble size monitoring method and device
Chatrasi et al. Pedestrian and Object Detection using Image Processing by YOLOv3 and YOLOv2
CN110728310A (en) Target detection model fusion method and system based on hyper-parameter optimization
CN116503406B (en) Hydraulic engineering information management system based on big data
CN117011628A (en) Image classification model training method, image classification method and device
Saini et al. Hybrid Deep Learning Model for The SmartManufacturing

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination