CN112052797A - MaskRCNN-based video fire identification method and system - Google Patents
MaskRCNN-based video fire identification method and system Download PDFInfo
- Publication number
- CN112052797A CN112052797A CN202010931021.0A CN202010931021A CN112052797A CN 112052797 A CN112052797 A CN 112052797A CN 202010931021 A CN202010931021 A CN 202010931021A CN 112052797 A CN112052797 A CN 112052797A
- Authority
- CN
- China
- Prior art keywords
- maskrcnn
- network
- target
- image
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 42
- 239000000779 smoke Substances 0.000 claims abstract description 56
- 238000001514 detection method Methods 0.000 claims abstract description 31
- 238000011176 pooling Methods 0.000 claims abstract description 17
- 230000003068 static effect Effects 0.000 claims abstract description 16
- 238000012549 training Methods 0.000 claims description 78
- 238000004364 calculation method Methods 0.000 claims description 49
- 238000010276 construction Methods 0.000 claims description 13
- 238000002372 labelling Methods 0.000 claims description 11
- 238000012545 processing Methods 0.000 claims description 9
- 238000001914 filtration Methods 0.000 claims description 8
- 238000010586 diagram Methods 0.000 claims description 7
- 230000001629 suppression Effects 0.000 claims description 5
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 4
- 238000005457 optimization Methods 0.000 claims description 4
- 238000013528 artificial neural network Methods 0.000 abstract description 5
- 238000013136 deep learning model Methods 0.000 abstract description 4
- 230000006870 function Effects 0.000 description 13
- 238000013527 convolutional neural network Methods 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 4
- 238000013145 classification model Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000005286 illumination Methods 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000007500 overflow downdraw method Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012706 support-vector machine Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Fire-Detection Mechanisms (AREA)
- Image Analysis (AREA)
Abstract
The invention provides a MaskRCNN-based video fire identification method and system, which are characterized in that a MaskRCNN deep learning model is adopted to detect targets such as smoke, flame and the like in a video image, the image in a video stream is transmitted to a trained MaskRCNN model, and through a series of operations such as convolution, pooling and the like, characteristic information in the image can be fully extracted, and predicted coordinate information and score results of the targets such as smoke, flame and the like can be accurately output. The MaskRCNN model has higher accuracy and reliability, but a small amount of false alarms exist, and in order to further reduce the false alarms, the embodiment uses a frame difference-based dynamic energy detection method to filter the detection result of the MaskRCNN, so that most of the false alarms of static objects can be removed. Finally, the deep neural network is adopted to carry out the final classification decision on the image area corresponding to the detected object, so that the false alarm rate is further reduced.
Description
Technical Field
The invention relates to the technical field of fire identification, in particular to a MaskRCNN-based video fire identification method and system.
Background
Currently, existing fire detection methods can be roughly divided into two categories: a fire detection method based on traditional image processing and a fire detection method based on deep learning. The fire detection method based on the traditional image processing is generally based on the visual characteristics of images, early scholars propose modeling color information in the images to extract a suspected flame area, and the method is high in real-time performance, but only considers the color characteristics, so that the method faces the problem of low accuracy. Then, researchers propose a method of dynamic background modeling to detect dynamic regions in the video, then obtain candidate regions through a color model, and finally complete final screening through shape features such as forms, textures and the like. Compared with a pure color model, the method has much higher performance, and meanwhile, the method for modeling the dynamic background greatly reduces the false alarm phenomenon of a static object. The fire detection methods based on the traditional image processing involve the use of color features more or less, however, in some real scenes, the reliability of the algorithm is greatly reduced due to the influence of factors such as color cast, overexposure and illumination of a camera. Subsequent scholars introduce machine learning algorithms such as a support vector machine and an artificial neural network into fire identification, and classify suspected regions by using the machine learning algorithms after extracting image features of the suspected regions. The method obtains better accuracy under the influence of interference factors such as illumination, shielding and the like, and the performance of the method is superior to that of most fire detection algorithms. However, these algorithms cannot effectively utilize the massive data set to improve the performance of the algorithms, and meanwhile, researchers are required to design features manually, which is cumbersome. The fire detection method based on deep learning generally adopts the existing mature algorithms, such as FasterRCNN, MaskRCNN, SSD, YOLO and the like, directly carries out target detection on flame and smoke in a single static image through a convolutional neural network, and locates out targets such as suspected flame and suspected smoke in the image.
For example, in the video smoke detection method based on the convolutional neural network disclosed as application No. 201911323715.x, a suspected smoke area is extracted, and then the convolutional neural network is used for image classification; the method has the problem of low detection precision.
In addition, the traditional image processing method needs to manually perform complex feature extraction and design, has strong dependency on manually selected features, is generally only suitable for a single specific scene, has poor performance in a real complex scene, and generally has the problems of low detection rate, serious false alarm phenomenon and the like.
Disclosure of Invention
The invention aims to solve the technical problems of how to improve the accuracy of video fire identification and reduce the false alarm rate.
The invention solves the technical problems through the following technical means:
the MaskRCNN-based video fire identification method is characterized by comprising the following steps of: the method comprises the following steps:
s01, training a MaskRCNN network to obtain a target MaskRCNN network;
s02, transmitting the images in the video stream to a target MaskRCNN network, and extracting characteristic information in the images to obtain an image classification data set;
s03, constructing an EfficientNet network, initializing the weight of the network, and training by using the image classification data set constructed in the step S02 to obtain a target EfficientNet network;
s04, detecting smoke and fire targets in the video by using a target MaskRCNN network to obtain coordinates, probability values and categories of target frames of suspected smoke and flame;
s05, judging frame difference energy, namely filtering the false alarm phenomenon of the static object by judging the frame difference energy of the target frame obtained in the step S04;
and S06, making a final decision by a classification network, and classifying the target frames reserved after the frame difference energy is judged by adopting an EfficientNet network.
The method adopts a MaskRCNN deep learning model to detect targets such as smoke, flame and the like in the video image, transmits the image in the video stream to the trained MaskRCNN model, can fully extract the characteristic information in the image through a series of operations such as convolution, pooling and the like, and can accurately output the predicted coordinate information and score results of the targets such as smoke, flame and the like. The MaskRCNN model has higher accuracy and reliability, but a small amount of false alarms exist, and in order to further reduce the false alarms, the embodiment uses a frame difference-based dynamic energy detection method to filter the detection result of the MaskRCNN, so that most of the false alarms of static objects can be removed. Finally, the deep neural network is adopted to carry out the final classification decision on the image area corresponding to the detected object, so that the false alarm rate is further reduced.
Further, the training of the MaskRCNN network in the target MaskRCNN network construction module specifically includes: collecting fire image samples, carrying out sample labeling, constructing a training data set, labeling coordinates and categories of targets such as smoke, flame and the like in the images, and carrying out data enhancement on the data set; constructing a MaskRCNN network, initializing the weight of the MaskRCNN network, and training the MaskRCNN network by using the constructed training set to obtain a target MaskRCNN network; the sample labels are specifically: marking the coordinates of the upper left corner (x1, y1) and the coordinates of the lower right corner (x2, y2) of the smoke and flame objects, and performing data enhancement operations on samples in the image dataset, wherein the data enhancement operations comprise random cropping, random brightness dithering, random saturation dithering, random contrast dithering, random hue dithering and Mixup operations, and the processing formula of Mixup is as follows:
wherein, lambda-Beta (1.5 );
wherein xiRepresenting images i, x to be fusedjRepresenting images j, y to be fusediAnd yjRespectively representing image i and imagej label information;
the training process of the MaskRCNN network comprises the following steps:
step S011, initializing the MaskRCNN network by using the trained network parameters on the COCO data set;
step S012: after image samples in the training data set are scaled to 1024x1024, extracting an overall characteristic diagram of the training sample image by using a ResNet101+ FPN network in MaskRCNN;
step S013: inputting the overall feature map into an RPN network, predicting an ROI (region of interest), and selecting positive and negative samples according to the overlapping ratio of a target frame of a candidate region and a labeled target frame;
step S014: ROIAlign pooling calculation is carried out on the ROI area on the feature map corresponding to the positive and negative samples, so as to obtain a candidate area feature map with a fixed size;
when ROIAlign pooling calculation is carried out, firstly, an ROI area target frame is mapped onto a feature map, then a feature image ROI area is obtained according to a minimum circumscribed rectangle algorithm, the ROI area is divided into m multiplied by m grids, 4 points on the feature map are selected by each grid to carry out bilinear difference, and finally, the feature map with the size of m multiplied by m is obtained;
step S015: classifying the ROI area feature graph after ROIAlign calculation and performing regression calculation on a target frame;
step S016: calculating a loss function of the MaskRCNN, performing gradient calculation on the loss function through a random gradient descent algorithm, and updating the weight of the MaskRCNN network;
step S017: and repeating the steps S012 to S016 until reaching the preset iteration number, stopping training and storing the MaskRCNN network.
Further, the step S02 specifically includes: and detecting a large amount of video data and image data by using a target MaskRCNN network, cutting out targets with suspected smoke and flame, and constructing an image classification data set comprising 3 categories of smoke, flame and probability values.
Further, the step S03 is specifically: and initializing parameters of the EfficientNet network by using the trained EfficientNet classification network parameters on the ImageNet data set, inputting the image classification data set, performing end-to-end training, performing gradient calculation on a classification loss function through an Adam optimization algorithm, updating the parameters of the EfficientNet network, and stopping training after finishing the training of a set round to obtain the target EfficientNet network.
Further, the step S04 is specifically: acquiring images from a video stream, zooming the video images to 1024x1024 sizes, inputting the video images into a target MashRCNN network, and predicting to obtain target frame coordinates, probability values and categories of suspected smoke and flame; and performing NMS non-maximum suppression on all the predicted targets, and filtering out overlapped invalid target boxes.
Further, the step S05 is specifically: cutting out an image area corresponding to a target frame in each frame of the nearest adjacent N frames according to the target frame, wherein N is larger than 2, respectively carrying out frame difference calculation on the adjacent 2 images of the N images acquired by each target frame, carrying out binarization calculation according to a threshold value T to obtain N-1 binary images, counting the number of non-zero values of pixel values in all the binary images, and dividing the number by the area of the target frame to obtain a final energy value; if the energy value is larger than the set energy threshold value, the target frame is effective, otherwise, the target frame is considered as a static object false alarm, and the target frame is discarded.
The invention also provides a MaskRCNN-based video fire identification system, which comprises:
a target MaskRCNN network construction module trains a MaskRCNN network to obtain a target MaskRCNN network;
the image classification data set construction module is used for transmitting the images in the video stream to a target MaskRCNN network, extracting characteristic information in the images and obtaining an image classification data set;
the target EfficientNet network construction module is used for constructing an EfficientNet network, initializing the weight of the network, and training by using the image classification data set constructed in the step S02 to obtain the target EfficientNet network;
the smoke and fire target detection module is used for detecting smoke and fire targets in the video by using a target MaskRCNN network to obtain target frame coordinates, probability values and categories of suspected smoke and flame;
the frame difference energy judgment module is used for filtering the false alarm phenomenon of the static object by frame difference energy judgment aiming at the target frame;
and the classification network final decision module is used for classifying the target frames reserved after the frame difference energy is judged by adopting an EfficientNet network.
Further, the training of the MaskRCNN network in the target MaskRCNN network construction module specifically includes: collecting fire image samples, carrying out sample labeling, constructing a training data set, labeling coordinates and categories of targets such as smoke, flame and the like in the images, and carrying out data enhancement on the data set; constructing a MaskRCNN network, initializing the weight of the MaskRCNN network, and training the MaskRCNN network by using the constructed training set to obtain a target MaskRCNN network; the sample labels are specifically: marking the coordinates of the upper left corner (x1, y1) and the coordinates of the lower right corner (x2, y2) of the smoke and flame objects, and performing data enhancement operations on samples in the image dataset, wherein the data enhancement operations comprise random cropping, random brightness dithering, random saturation dithering, random contrast dithering, random hue dithering and Mixup operations, and the processing formula of Mixup is as follows:
wherein, lambda-Beta (1.5 );
wherein xiRepresenting images i, x to be fusedjRepresenting images j, y to be fusediAnd yjLabel information respectively representing the image i and the image j;
the training process of the MaskRCNN network comprises the following steps:
step S011, initializing the MaskRCNN network by using the trained network parameters on the COCO data set;
step S012: after image samples in the training data set are scaled to 1024x1024, extracting an overall characteristic diagram of the training sample image by using a ResNet101+ FPN network in MaskRCNN;
step S013: inputting the overall feature map into an RPN network, predicting an ROI (region of interest), and selecting positive and negative samples according to the overlapping ratio of a target frame of a candidate region and a labeled target frame;
step S014: ROIAlign pooling calculation is carried out on the ROI area on the feature map corresponding to the positive and negative samples, so as to obtain a candidate area feature map with a fixed size;
when ROIAlign pooling calculation is carried out, firstly, an ROI area target frame is mapped onto a feature map, then a feature image ROI area is obtained according to a minimum circumscribed rectangle algorithm, the ROI area is divided into m multiplied by m grids, 4 points on the feature map are selected by each grid to carry out bilinear difference, and finally, the feature map with the size of m multiplied by m is obtained;
step S015: classifying the ROI area feature graph after ROIAlign calculation and performing regression calculation on a target frame;
step S016: calculating a loss function of the MaskRCNN, performing gradient calculation on the loss function through a random gradient descent algorithm, and updating the weight of the MaskRCNN network;
step S017: and repeating the steps S012 to S016 until reaching the preset iteration number, stopping training and storing the MaskRCNN network.
Further, the specific execution process of the smoke and fire target detection module is as follows: acquiring images from a video stream, zooming the video images to 1024x1024 sizes, inputting the video images into a target MashRCNN network, and predicting to obtain target frame coordinates, probability values and categories of suspected smoke and flame; and performing NMS non-maximum suppression on all the predicted targets, and filtering out overlapped invalid target boxes.
Further, the frame difference energy determination module specifically executes the following process: cutting out an image area corresponding to a target frame in each frame of the nearest adjacent N frames (N >3) according to the target frame, respectively performing frame difference calculation on the adjacent 2 images of the N images acquired by each target frame, performing binarization calculation according to a threshold value T to obtain N-1 binary images, and counting the number of non-zero values of pixel values in all the binary images, wherein the number is divided by the area of the target frame to obtain a final energy value; if the energy value is larger than the set energy threshold value, the target frame is effective, otherwise, the target frame is considered as a static object false alarm, and the target frame is discarded.
The invention has the advantages that:
the method adopts a MaskRCNN deep learning model to detect targets such as smoke, flame and the like in the video image, transmits the image in the video stream to the trained MaskRCNN model, can fully extract the characteristic information in the image through a series of operations such as convolution, pooling and the like, and can accurately output the predicted coordinate information and score results of the targets such as smoke, flame and the like. The MaskRCNN model has higher accuracy and reliability, but a small amount of false alarms exist, and in order to further reduce the false alarms, the embodiment uses a frame difference-based dynamic energy detection method to filter the detection result of the MaskRCNN, so that most of the false alarms of static objects can be removed. Finally, the deep neural network is adopted to carry out the final classification decision on the image area corresponding to the detected object, so that the false alarm rate is further reduced.
Based on the deep learning algorithm, various characteristics of flame and smoke in the image are independently learned through the convolutional neural network, the learned characteristics can be identified and positioned for the flame and smoke in the image, and the method has the advantages of high detection rate, low false alarm rate, high robustness and the like. In actual deployment, the method can basically meet the real-time requirement through simultaneous acceleration of software and hardware.
Drawings
FIG. 1 is a flow chart of MaskRCNN network and EfficientNet network training in the embodiment of the present invention;
FIG. 2 is a flowchart illustrating a method for video fire identification to detect a fire in accordance with an embodiment of the present invention;
FIG. 3 is a flowchart of MaskRCNN network fire detection according to the present invention;
FIG. 4 is a flow chart of dynamic detection in an embodiment of the present invention;
FIG. 5 is a diagram illustrating the effect of the Mixup method in the embodiment of the present invention;
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the embodiments of the present invention, and it is obvious that the described embodiments are some embodiments of the present invention, but not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The embodiment provides a MaskRCNN-based video fire identification method, which comprises the following steps:
step S1: as shown in fig. 1, collecting fire image samples, performing sample labeling, constructing a training data set, labeling coordinates and categories of targets such as smoke, flame and the like in the images, and performing data enhancement on the data set; the data enhancement operations include random clipping, random luminance dithering, random saturation dithering, random contrast dithering, random hue dithering, mixup, and the like. As shown in fig. 1, the embodiment provides an image fusion method named mixup, in which two different images are fused according to a certain ratio, so that the diversity of a training data set can be effectively increased, as shown in fig. 5;
wherein, the coordinates of the upper left corner (x1, y1) and the coordinates of the lower right corner (x2, y2) of the smoke and flame objects need to be marked, and the data enhancement operation on the samples in the image data set comprises the operations of random clipping, random brightness dithering, random saturation dithering, random contrast dithering, random hue dithering, mixup and the like. The processing formula for Mixup is as follows:
wherein, lambda-Beta (1.5 );
wherein xiRepresenting images i, x to be fusedjRepresenting images j, y to be fusediAnd yjThe label information (target frame and category information) of the image i and the image j are respectively represented.
Step S2: as shown in fig. 2, a MaskRCNN network is constructed, the weights of the network are initialized, the MaskRCNN network is trained by using the training set constructed in step S1, and after 100 rounds of training, the training is stopped;
the MaskRCNN model training process is specifically as follows:
s2.1, initializing parameters of the MaskRCNN network by using model parameters trained on a COCO data set;
step S2.2: after image samples in the training data set are scaled to 1024x1024, extracting an overall characteristic diagram of the training sample image by using a ResNet101+ FPN network in MaskRCNN;
step S2.3: inputting the overall feature map into an RPN network, predicting a candidate Region (ROI), and selecting positive and negative samples according to the overlapping ratio of a target frame of the candidate region and a labeled target frame;
step S2.4: ROIAlign pooling calculation is carried out on the ROI area on the feature map corresponding to the positive and negative samples, so as to obtain a candidate area feature map with a fixed size;
when ROIAlign pooling calculation is carried out, firstly, a candidate region target frame is mapped onto a feature map, then a feature image ROI region is obtained according to a minimum circumscribed rectangle algorithm, the ROI region is divided into m multiplied by m grids, 4 points on the feature map are selected by each grid to carry out bilinear difference, and finally, the feature map with the size of m multiplied by m is obtained;
step S2.5: classifying the candidate region feature graph after ROIAlign calculation and performing regression calculation on a target frame;
step S2.6: calculating a loss function of the MaskRCNN, performing gradient calculation on the loss function through a random gradient descent algorithm, and updating the weight of the MaskRCNN network;
step S2.7: repeating the steps S2.2 to S2.6 until the preset iteration times are reached, stopping training, and storing the MaskRCNN network model;
step S3: as shown in fig. 1, a trained MaskRCNN model is used to detect a large amount of video data and image data, and suspected smoke and flame targets are cut out, so as to construct an image classification dataset including 3 categories of smoke, flame and false alarm;
step S4: as shown in fig. 1, an EfficientNet network is constructed, a pre-trained model on an ImageNet dataset is used to perform parameter initialization on the EfficientNet network, the image classification dataset constructed in step S3 is used for training, and after 50 rounds of training, the training is stopped;
initializing parameters of an EfficientNet network by using the trained EfficientNet classification model parameters on the ImageNet data set, inputting a training data image, performing end-to-end training, performing gradient calculation on a classification loss function through an Adam optimization algorithm, updating the parameters of the EfficientNet network, and stopping training after 50 rounds of training are completed;
as shown in fig. 2, after the target MaskRCNN network and the target EfficientNet network are obtained, the method for identifying the video fire specifically includes the following steps:
step S5, as shown in fig. 3, a single frame image obtained by decoding a video stream is preprocessed, the preprocessed image is subjected to overall image feature extraction through a ResNet101+ FPN network in MaskRCNN, a candidate region is obtained by prediction after the feature map is input into an RPN network, a candidate feature map with a fixed size is output after roiign pooling calculation is performed on the feature map of the candidate region, and finally a classification regression network in MaskRCNN classifies the candidate feature map and regresses an object frame. Considering that the prediction result of MaskRCNN has a large target overlap, NMS calculation is required for the prediction of MaskRCNN to suppress targets with an excessively large overlap region.
Step S6, frame difference energy determination: after MaskRCNN network detection and NMS calculation are completed, dynamic detection needs to be performed on a suspected target region, as shown in fig. 4, a target ROI region in consecutive adjacent N frames (N >2) is first obtained, frame difference calculation is performed between adjacent 2 frames, binarization calculation is performed according to a threshold T to obtain a binary image, the number of pixels having non-zero values in all binary images is counted, and the number is divided by the area of the target frame to obtain a final energy value. Setting the pixels with the absolute value larger than the threshold value T in the frame difference image as 1, setting the rest pixels as 0, counting the number of non-zero-value pixels in all the frame difference images, and dividing the number of the pixels by the area of the target area to obtain the required energy value. If the energy value is larger than the set threshold value, the target frame is judged to be dynamic, and the target frame is effective, otherwise, the target frame is judged to be static, and the target frame is discarded.
Step S7, the classification network makes a final decision: if the suspected target area is judged to be dynamic, further judgment is needed, the suspected target area image is input into a target EfficientNet network, and the target EfficientNet classifies the image and judges whether the image is false alarm or smoke or flame. If the smoke or flame is judged, directly alarming, otherwise, directly entering the detection process of the next frame without alarming.
The embodiment provides a video fire detection method based on deep learning, which is characterized in that a MaskRCNN deep learning model is adopted to detect targets such as smoke, flame and the like in a video image, the image in a video stream is transmitted to the trained MaskRCNN model, and through a series of operations such as convolution, pooling and the like, characteristic information in the image can be fully extracted, and predicted coordinate information and score results of the targets such as smoke, flame and the like can be accurately output. The MaskRCNN model has higher accuracy and reliability, but a small amount of false alarms exist, and in order to further reduce the false alarms, the embodiment uses a frame difference-based dynamic energy detection method to filter the detection result of the MaskRCNN, so that most of the false alarms of static objects can be removed. Finally, the deep neural network is adopted to carry out the final classification decision on the image area corresponding to the detected object, so that the false alarm rate is further reduced.
Considering that the MaskRCNN algorithm requires training data to provide mask labeling information, and the embodiment does not need a semantic segmentation function, the embodiment eliminates the semantic segmentation related branch function in the MaskRCNN network.
The embodiment also provides a video fire identification system based on MaskRCNN, including:
the target MaskRCNN network construction module collects fire image samples, carries out sample marking, constructs a training data set, needs to mark coordinates and types of targets such as smoke, flame and the like in the images, and carries out data enhancement on the data set;
wherein, the coordinates of the upper left corner (x1, y1) and the coordinates of the lower right corner (x2, y2) of the smoke and flame objects need to be marked, and the data enhancement operation on the samples in the image data set comprises the operations of random clipping, random brightness dithering, random saturation dithering, random contrast dithering, random hue dithering, mixup and the like. The processing formula for Mixup is as follows:
wherein, lambda-Beta (1.5 );
wherein xiRepresenting images i, x to be fusedjRepresenting images j, y to be fusediAnd yjThe label information (target frame and category information) of the image i and the image j are respectively represented.
Constructing a MaskRCNN network, initializing the weight of the network, training the MaskRCNN network by using the training set constructed in the step S1, and stopping training after 100 rounds of training;
the MaskRCNN model training process is specifically as follows:
s2.1, initializing parameters of the MaskRCNN network by using model parameters trained on a COCO data set;
step S2.2: after image samples in the training data set are scaled to 1024x1024, extracting an overall characteristic diagram of the training sample image by using a ResNet101+ FPN network in MaskRCNN;
step S2.3: inputting the overall feature map into an RPN network, predicting a candidate Region (ROI), and selecting positive and negative samples according to the overlapping ratio of a target frame of the candidate region and a labeled target frame;
step S2.4: ROIAlign pooling calculation is carried out on the ROI area on the feature map corresponding to the positive and negative samples, so as to obtain a candidate area feature map with a fixed size;
when ROIAlign pooling calculation is carried out, firstly, a candidate region target frame is mapped onto a feature map, then a feature image ROI region is obtained according to a minimum circumscribed rectangle algorithm, the ROI region is divided into m x m grids, 4 points on the feature map are selected by each grid to carry out bilinear difference, and finally, the feature map with the size of m x m is obtained;
step S2.5: classifying the candidate region feature graph after ROIAlign calculation and performing regression calculation on a target frame;
step S2.6: calculating a loss function of the MaskRCNN, performing gradient calculation on the loss function through a random gradient descent algorithm, and updating the weight of the MaskRCNN network;
step S2.7: repeating the steps S2.2 to S2.6 until the preset iteration times are reached, stopping training, and storing the MaskRCNN network model;
the image classification data set construction module is used for detecting a large amount of video data and image data by using a trained MaskRCNN model, cutting out a suspected smoke and flame target, and constructing an image classification data set with the 3 categories of smoke, flame and false alarm;
a target EfficientNet network construction module, wherein the weight of the network is initialized, the image classification data set constructed in the step S3 is used for training, and the training is stopped after 50 rounds of training;
initializing parameters of an EfficientNet network by using the trained EfficientNet classification model parameters on the ImageNet data set, inputting a training data image, performing end-to-end training, performing gradient calculation on a classification loss function through an Adam optimization algorithm, updating the parameters of the EfficientNet network, and stopping training after 50 rounds of training are completed;
after the target MaskRCNN network and the target EfficientNet network are obtained, the method for identifying the video fire specifically comprises the following steps:
a smoke and fire target detection module, MaskRCNN detects smoke and fire targets in the video: acquiring images from a video stream, zooming the video images to 1024x1024 sizes, inputting the video images into a MashRCNN network, and predicting to obtain coordinates, probability values and categories of target frames of suspected smoke and flame; NMS non-maximum suppression is carried out on all the predicted targets, and a large number of overlapped invalid target frames can be filtered;
and the frame difference energy judging module cuts out an image area corresponding to a target frame (assuming that the width of the target frame is w and the height of the target frame is h) in each frame in the nearest adjacent N frames (N >3) according to the target frame obtained in the last step, respectively performs frame difference calculation on the adjacent 2 images, performs binarization calculation according to a threshold value T to obtain N-1 binary images, and counts the number of non-zero pixel values in all the binary images, wherein the number is the final energy value by dividing the area of the target frame. If the energy value is larger than the set energy threshold value, the target frame is effective, otherwise, the target frame is considered as a static object false alarm, and the target frame is discarded;
and the classification network final decision module directly cuts the target through a target frame reserved after the frame difference energy judgment, scales the target to 128x128 size, outputs the EfficientNet network for final classification, and gives an alarm if the prediction is smoke or flame, or does not give an alarm.
The above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.
Claims (10)
1. The MaskRCNN-based video fire identification method is characterized by comprising the following steps of: the method comprises the following steps:
s01, training a MaskRCNN network to obtain a target MaskRCNN network;
s02, transmitting the images in the video stream to a target MaskRCNN network, and extracting characteristic information in the images to obtain an image classification data set;
s03, constructing an EfficientNet network, initializing the weight of the network, and training by using the image classification data set constructed in the step S02 to obtain a target EfficientNet network;
s04, detecting smoke and fire targets in the video by using a target MaskRCNN network to obtain coordinates, probability values and categories of target frames of suspected smoke and flame;
s05, judging frame difference energy, namely filtering the false alarm phenomenon of the static object by judging the frame difference energy of the target frame obtained in the step S04;
and S06, making a final decision by a classification network, and classifying the target frames reserved after the frame difference energy is judged by adopting an EfficientNet network.
2. The MaskRCNN-based video fire identification method according to claim 1, wherein: the MaskRCNN network training in the target MaskRCNN network construction module specifically comprises the following steps: collecting fire image samples, carrying out sample labeling, constructing a training data set, labeling coordinates and categories of targets such as smoke, flame and the like in the images, and carrying out data enhancement on the data set; constructing a MaskRCNN network, initializing the weight of the MaskRCNN network, and training the MaskRCNN network by using the constructed training set to obtain a target MaskRCNN network; the sample labels are specifically: marking the coordinates of the upper left corner (x1, y1) and the coordinates of the lower right corner (x2, y2) of the smoke and flame objects, and performing data enhancement operations on samples in the image dataset, wherein the data enhancement operations comprise random cropping, random brightness dithering, random saturation dithering, random contrast dithering, random hue dithering and Mixup operations, and the processing formula of Mixup is as follows:
wherein, lambda-Beta (1.5 );
wherein xiRepresenting images i, x to be fusedjRepresenting images j, y to be fusediAnd yjLabel information respectively representing the image i and the image j;
the training process of the MaskRCNN network comprises the following steps:
step S011, initializing the MaskRCNN network by using the trained network parameters on the COCO data set;
step S012: after image samples in the training data set are scaled to 1024x1024, extracting an overall characteristic diagram of the training sample image by using a ResNet101+ FPN network in MaskRCNN;
step S013: inputting the overall feature map into an RPN network, predicting an ROI (region of interest), and selecting positive and negative samples according to the overlapping ratio of a target frame of a candidate region and a labeled target frame;
step S014: ROIAlign pooling calculation is carried out on the ROI area on the feature map corresponding to the positive and negative samples, so as to obtain a candidate area feature map with a fixed size;
when ROIAlign pooling calculation is carried out, firstly, an ROI area target frame is mapped onto a feature map, then a feature image ROI area is obtained according to a minimum circumscribed rectangle algorithm, the ROI area is divided into m multiplied by m grids, 4 points on the feature map are selected by each grid to carry out bilinear difference, and finally, the feature map with the size of m multiplied by m is obtained;
step S015: classifying the ROI area feature graph after ROIAlign calculation and performing regression calculation on a target frame;
step S016: calculating a loss function of the MaskRCNN, performing gradient calculation on the loss function through a random gradient descent algorithm, and updating the weight of the MaskRCNN network;
step S017: and repeating the steps S012 to S016 until reaching the preset iteration number, stopping training and storing the MaskRCNN network.
3. The MaskRCNN-based video fire identification method according to claim 2, wherein: the step S02 specifically includes: and detecting a large amount of video data and image data by using a target MaskRCNN network, cutting out targets with suspected smoke and flame, and constructing an image classification data set comprising 3 categories of smoke, flame and probability values.
4. The MaskRCNN-based video fire identification method according to claim 1, wherein: the step S03 specifically includes: and initializing parameters of the EfficientNet network by using the trained EfficientNet classification network parameters on the ImageNet data set, inputting the image classification data set, performing end-to-end training, performing gradient calculation on a classification loss function through an Adam optimization algorithm, updating the parameters of the EfficientNet network, and stopping training after finishing the training of a set round to obtain the target EfficientNet network.
5. The MaskRCNN-based video fire identification method according to claim 1, wherein: the step S04 specifically includes: acquiring images from a video stream, zooming the video images to 1024x1024 sizes, inputting the video images into a target MashRCNN network, and predicting to obtain target frame coordinates, probability values and categories of suspected smoke and flame; and performing NMS non-maximum suppression on all the predicted targets, and filtering out overlapped invalid target boxes.
6. The MaskRCNN-based video fire identification method according to claim 1, wherein: the step S05 specifically includes: cutting out an image area corresponding to a target frame in each frame of the nearest adjacent N frames according to the target frame, wherein N is larger than 2, respectively carrying out frame difference calculation on the adjacent 2 images of the N images acquired by each target frame, carrying out binarization calculation according to a threshold value T to obtain N-1 binary images, counting the number of non-zero values of pixel values in all the binary images, and dividing the number by the area of the target frame to obtain a final energy value; if the energy value is larger than the set energy threshold value, the target frame is effective, otherwise, the target frame is considered as a static object false alarm, and the target frame is discarded.
7. Video fire identification system based on maskRCNN, its characterized in that: the method comprises the following steps:
a target MaskRCNN network construction module trains a MaskRCNN network to obtain a target MaskRCNN network;
the image classification data set construction module is used for transmitting the images in the video stream to a target MaskRCNN network, extracting characteristic information in the images and obtaining an image classification data set;
the target EfficientNet network construction module is used for constructing an EfficientNet network, initializing the weight of the network, and training by using the image classification data set constructed in the step S02 to obtain the target EfficientNet network;
the smoke and fire target detection module is used for detecting smoke and fire targets in the video by using a target MaskRCNN network to obtain target frame coordinates, probability values and categories of suspected smoke and flame;
the frame difference energy judgment module is used for filtering the false alarm phenomenon of the static object by frame difference energy judgment aiming at the target frame;
and the classification network final decision module is used for classifying the target frames reserved after the frame difference energy is judged by adopting an EfficientNet network.
8. The MaskRCNN-based video fire identification system according to claim 7, wherein: the MaskRCNN network training in the target MaskRCNN network construction module specifically comprises the following steps: collecting fire image samples, carrying out sample labeling, constructing a training data set, labeling coordinates and categories of targets such as smoke, flame and the like in the images, and carrying out data enhancement on the data set; constructing a MaskRCNN network, initializing the weight of the MaskRCNN network, and training the MaskRCNN network by using the constructed training set to obtain a target MaskRCNN network; the sample labels are specifically: marking the coordinates of the upper left corner (x1, y1) and the coordinates of the lower right corner (x2, y2) of the smoke and flame objects, and performing data enhancement operations on samples in the image dataset, wherein the data enhancement operations comprise random cropping, random brightness dithering, random saturation dithering, random contrast dithering, random hue dithering and Mixup operations, and the processing formula of Mixup is as follows:
wherein, lambda-Beta (1.5 );
wherein xiRepresenting images i, x to be fusedjRepresenting images j, y to be fusediAnd yjLabel information respectively representing the image i and the image j;
the training process of the MaskRCNN network comprises the following steps:
step S011, initializing the MaskRCNN network by using the trained network parameters on the COCO data set;
step S012: after image samples in the training data set are scaled to 1024x1024, extracting an overall characteristic diagram of the training sample image by using a ResNet101+ FPN network in MaskRCNN;
step S013: inputting the overall feature map into an RPN network, predicting an ROI (region of interest), and selecting positive and negative samples according to the overlapping ratio of a target frame of a candidate region and a labeled target frame;
step S014: ROIAlign pooling calculation is carried out on the ROI area on the feature map corresponding to the positive and negative samples, so as to obtain a candidate area feature map with a fixed size;
when ROIAlign pooling calculation is carried out, firstly, an ROI area target frame is mapped onto a feature map, then a feature image ROI area is obtained according to a minimum circumscribed rectangle algorithm, the ROI area is divided into m multiplied by m grids, 4 points on the feature map are selected by each grid to carry out bilinear difference, and finally, the feature map with the size of m multiplied by m is obtained;
step S015: classifying the ROI area feature graph after ROIAlign calculation and performing regression calculation on a target frame;
step S016: calculating a loss function of the MaskRCNN, performing gradient calculation on the loss function through a random gradient descent algorithm, and updating the weight of the MaskRCNN network;
step S017: and repeating the steps S012 to S016 until reaching the preset iteration number, stopping training and storing the MaskRCNN network.
9. The MaskRCNN-based video fire identification system according to claim 7, wherein: the specific execution process of the smoke and fire target detection module is as follows: acquiring images from a video stream, zooming the video images to 1024x1024 sizes, inputting the video images into a target MashRCNN network, and predicting to obtain target frame coordinates, probability values and categories of suspected smoke and flame; and performing NMS non-maximum suppression on all the predicted targets, and filtering out overlapped invalid target boxes.
10. The MaskRCNN-based video fire identification system according to claim 7, wherein: the frame difference energy judging module specifically executes the following processes: cutting out an image area corresponding to a target frame in each frame of the nearest adjacent N frames (N >3) according to the target frame, respectively performing frame difference calculation on the adjacent 2 images of the N images acquired by each target frame, performing binarization calculation according to a threshold value T to obtain N-1 binary images, and counting the number of non-zero values of pixel values in all the binary images, wherein the number is divided by the area of the target frame to obtain a final energy value; if the energy value is larger than the set energy threshold value, the target frame is effective, otherwise, the target frame is considered as a static object false alarm, and the target frame is discarded.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010931021.0A CN112052797B (en) | 2020-09-07 | 2020-09-07 | MaskRCNN-based video fire disaster identification method and MaskRCNN-based video fire disaster identification system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010931021.0A CN112052797B (en) | 2020-09-07 | 2020-09-07 | MaskRCNN-based video fire disaster identification method and MaskRCNN-based video fire disaster identification system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112052797A true CN112052797A (en) | 2020-12-08 |
CN112052797B CN112052797B (en) | 2024-07-16 |
Family
ID=73609865
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010931021.0A Active CN112052797B (en) | 2020-09-07 | 2020-09-07 | MaskRCNN-based video fire disaster identification method and MaskRCNN-based video fire disaster identification system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112052797B (en) |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112633231A (en) * | 2020-12-30 | 2021-04-09 | 珠海大横琴科技发展有限公司 | Fire disaster identification method and device |
CN112686190A (en) * | 2021-01-05 | 2021-04-20 | 北京林业大学 | Forest fire smoke automatic identification method based on self-adaptive target detection |
CN112699801A (en) * | 2020-12-30 | 2021-04-23 | 上海船舶电子设备研究所(中国船舶重工集团公司第七二六研究所) | Fire identification method and system based on video image |
CN112800929A (en) * | 2021-01-25 | 2021-05-14 | 安徽农业大学 | On-line monitoring method for bamboo shoot quantity and high growth rate based on deep learning |
CN112819001A (en) * | 2021-03-05 | 2021-05-18 | 浙江中烟工业有限责任公司 | Complex scene cigarette packet identification method and device based on deep learning |
CN112861635A (en) * | 2021-01-11 | 2021-05-28 | 西北工业大学 | Fire and smoke real-time detection method based on deep learning |
CN112907885A (en) * | 2021-01-12 | 2021-06-04 | 中国计量大学 | Distributed centralized household image fire alarm system and method based on SCNN |
CN112907886A (en) * | 2021-02-07 | 2021-06-04 | 中国石油化工股份有限公司 | Refinery plant fire identification method based on convolutional neural network |
CN113033553A (en) * | 2021-03-22 | 2021-06-25 | 深圳市安软科技股份有限公司 | Fire detection method and device based on multi-mode fusion, related equipment and storage medium |
CN113192038A (en) * | 2021-05-07 | 2021-07-30 | 北京科技大学 | Method for identifying and monitoring abnormal smoke and fire in existing flame environment based on deep learning |
CN113409923A (en) * | 2021-05-25 | 2021-09-17 | 济南大学 | Error correction method and system in bone marrow image individual cell automatic marking |
CN113553985A (en) * | 2021-08-02 | 2021-10-26 | 中再云图技术有限公司 | High-altitude smoke detection and identification method based on artificial intelligence, storage device and server |
CN113627223A (en) * | 2021-01-07 | 2021-11-09 | 广州中国科学院软件应用技术研究所 | Flame detection algorithm based on deep learning target detection and classification technology |
CN113657250A (en) * | 2021-08-16 | 2021-11-16 | 南京图菱视频科技有限公司 | Flame detection method and system based on monitoring video |
CN113657233A (en) * | 2021-08-10 | 2021-11-16 | 东华大学 | Unmanned aerial vehicle forest fire smoke detection method based on computer vision |
CN113762314A (en) * | 2021-02-02 | 2021-12-07 | 北京京东振世信息技术有限公司 | Smoke and fire detection method and device |
CN113792684A (en) * | 2021-09-17 | 2021-12-14 | 中国科学技术大学 | Fire-fighting robot multi-mode visual flame detection method under weak alignment condition |
CN113947744A (en) * | 2021-10-26 | 2022-01-18 | 华能盐城大丰新能源发电有限责任公司 | Fire image detection method, system, equipment and storage medium based on video |
CN114120208A (en) * | 2022-01-27 | 2022-03-01 | 青岛海尔工业智能研究院有限公司 | Flame detection method, device, equipment and storage medium |
CN114120171A (en) * | 2021-10-28 | 2022-03-01 | 华能盐城大丰新能源发电有限责任公司 | Fire smoke detection method, device and equipment based on video frame and storage medium |
CN114664047A (en) * | 2022-05-26 | 2022-06-24 | 长沙海信智能系统研究院有限公司 | Expressway fire identification method and device and electronic equipment |
CN115170894A (en) * | 2022-09-05 | 2022-10-11 | 深圳比特微电子科技有限公司 | Smoke and fire detection method and device |
CN115862258A (en) * | 2022-11-22 | 2023-03-28 | 中国科学院合肥物质科学研究院 | Fire monitoring and handling system, method, equipment and storage medium |
CN115861898A (en) * | 2022-12-27 | 2023-03-28 | 浙江创悦诚科技有限公司 | Flame smoke identification method applied to gas field station |
CN117011785A (en) * | 2023-07-06 | 2023-11-07 | 华新水泥股份有限公司 | Firework detection method, device and system based on space-time correlation and Gaussian heat map |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180089593A1 (en) * | 2016-09-26 | 2018-03-29 | Acusense Technologies, Inc. | Method and system for an end-to-end artificial intelligence workflow |
CN109903507A (en) * | 2019-03-04 | 2019-06-18 | 上海海事大学 | A kind of fire disaster intelligent monitor system and method based on deep learning |
CN109977790A (en) * | 2019-03-04 | 2019-07-05 | 浙江工业大学 | A kind of video smoke detection and recognition methods based on transfer learning |
CN110232380A (en) * | 2019-06-13 | 2019-09-13 | 应急管理部天津消防研究所 | Fire night scenes restored method based on Mask R-CNN neural network |
CN111539265A (en) * | 2020-04-02 | 2020-08-14 | 申龙电梯股份有限公司 | Method for detecting abnormal behaviors in elevator car |
-
2020
- 2020-09-07 CN CN202010931021.0A patent/CN112052797B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180089593A1 (en) * | 2016-09-26 | 2018-03-29 | Acusense Technologies, Inc. | Method and system for an end-to-end artificial intelligence workflow |
CN109903507A (en) * | 2019-03-04 | 2019-06-18 | 上海海事大学 | A kind of fire disaster intelligent monitor system and method based on deep learning |
CN109977790A (en) * | 2019-03-04 | 2019-07-05 | 浙江工业大学 | A kind of video smoke detection and recognition methods based on transfer learning |
CN110232380A (en) * | 2019-06-13 | 2019-09-13 | 应急管理部天津消防研究所 | Fire night scenes restored method based on Mask R-CNN neural network |
CN111539265A (en) * | 2020-04-02 | 2020-08-14 | 申龙电梯股份有限公司 | Method for detecting abnormal behaviors in elevator car |
Non-Patent Citations (3)
Title |
---|
REDDY, N. DINESH: "Occlusion-Net: 2D/3D Occluded Keypoint Localization Using Graph Networks", IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, 16 July 2020 (2020-07-16) * |
周育新;白宏阳;李伟;郭宏伟;徐啸康;: "基于关键帧的轻量化行为识别方法研究", 仪器仪表学报, no. 07, pages 3 * |
查伟伟;白天;: "公路视频实时车辆检测分类与车流量统计算法", 信息技术与网络安全, no. 03 * |
Cited By (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112699801A (en) * | 2020-12-30 | 2021-04-23 | 上海船舶电子设备研究所(中国船舶重工集团公司第七二六研究所) | Fire identification method and system based on video image |
CN112633231A (en) * | 2020-12-30 | 2021-04-09 | 珠海大横琴科技发展有限公司 | Fire disaster identification method and device |
CN112686190A (en) * | 2021-01-05 | 2021-04-20 | 北京林业大学 | Forest fire smoke automatic identification method based on self-adaptive target detection |
CN113627223A (en) * | 2021-01-07 | 2021-11-09 | 广州中国科学院软件应用技术研究所 | Flame detection algorithm based on deep learning target detection and classification technology |
CN112861635B (en) * | 2021-01-11 | 2024-05-14 | 西北工业大学 | Fire disaster and smoke real-time detection method based on deep learning |
CN112861635A (en) * | 2021-01-11 | 2021-05-28 | 西北工业大学 | Fire and smoke real-time detection method based on deep learning |
CN112907885A (en) * | 2021-01-12 | 2021-06-04 | 中国计量大学 | Distributed centralized household image fire alarm system and method based on SCNN |
CN112800929A (en) * | 2021-01-25 | 2021-05-14 | 安徽农业大学 | On-line monitoring method for bamboo shoot quantity and high growth rate based on deep learning |
CN113762314A (en) * | 2021-02-02 | 2021-12-07 | 北京京东振世信息技术有限公司 | Smoke and fire detection method and device |
CN113762314B (en) * | 2021-02-02 | 2023-11-03 | 北京京东振世信息技术有限公司 | Firework detection method and device |
CN112907886A (en) * | 2021-02-07 | 2021-06-04 | 中国石油化工股份有限公司 | Refinery plant fire identification method based on convolutional neural network |
CN112819001B (en) * | 2021-03-05 | 2024-02-23 | 浙江中烟工业有限责任公司 | Complex scene cigarette packet recognition method and device based on deep learning |
CN112819001A (en) * | 2021-03-05 | 2021-05-18 | 浙江中烟工业有限责任公司 | Complex scene cigarette packet identification method and device based on deep learning |
CN113033553A (en) * | 2021-03-22 | 2021-06-25 | 深圳市安软科技股份有限公司 | Fire detection method and device based on multi-mode fusion, related equipment and storage medium |
CN113033553B (en) * | 2021-03-22 | 2023-05-12 | 深圳市安软科技股份有限公司 | Multi-mode fusion fire detection method, device, related equipment and storage medium |
CN113192038B (en) * | 2021-05-07 | 2022-08-19 | 北京科技大学 | Method for recognizing and monitoring abnormal smoke and fire in existing flame environment based on deep learning |
CN113192038A (en) * | 2021-05-07 | 2021-07-30 | 北京科技大学 | Method for identifying and monitoring abnormal smoke and fire in existing flame environment based on deep learning |
CN113409923A (en) * | 2021-05-25 | 2021-09-17 | 济南大学 | Error correction method and system in bone marrow image individual cell automatic marking |
CN113409923B (en) * | 2021-05-25 | 2022-03-04 | 济南大学 | Error correction method and system in bone marrow image individual cell automatic marking |
CN113553985A (en) * | 2021-08-02 | 2021-10-26 | 中再云图技术有限公司 | High-altitude smoke detection and identification method based on artificial intelligence, storage device and server |
CN113657233A (en) * | 2021-08-10 | 2021-11-16 | 东华大学 | Unmanned aerial vehicle forest fire smoke detection method based on computer vision |
CN113657250A (en) * | 2021-08-16 | 2021-11-16 | 南京图菱视频科技有限公司 | Flame detection method and system based on monitoring video |
CN113792684A (en) * | 2021-09-17 | 2021-12-14 | 中国科学技术大学 | Fire-fighting robot multi-mode visual flame detection method under weak alignment condition |
CN113792684B (en) * | 2021-09-17 | 2024-03-29 | 中国科学技术大学 | Multi-mode visual flame detection method for fire-fighting robot under weak alignment condition |
CN113947744A (en) * | 2021-10-26 | 2022-01-18 | 华能盐城大丰新能源发电有限责任公司 | Fire image detection method, system, equipment and storage medium based on video |
CN114120171A (en) * | 2021-10-28 | 2022-03-01 | 华能盐城大丰新能源发电有限责任公司 | Fire smoke detection method, device and equipment based on video frame and storage medium |
CN114120208A (en) * | 2022-01-27 | 2022-03-01 | 青岛海尔工业智能研究院有限公司 | Flame detection method, device, equipment and storage medium |
CN114664047A (en) * | 2022-05-26 | 2022-06-24 | 长沙海信智能系统研究院有限公司 | Expressway fire identification method and device and electronic equipment |
CN115170894A (en) * | 2022-09-05 | 2022-10-11 | 深圳比特微电子科技有限公司 | Smoke and fire detection method and device |
CN115862258A (en) * | 2022-11-22 | 2023-03-28 | 中国科学院合肥物质科学研究院 | Fire monitoring and handling system, method, equipment and storage medium |
CN115862258B (en) * | 2022-11-22 | 2023-09-22 | 中国科学院合肥物质科学研究院 | Fire monitoring and disposing system, method, equipment and storage medium |
CN115861898A (en) * | 2022-12-27 | 2023-03-28 | 浙江创悦诚科技有限公司 | Flame smoke identification method applied to gas field station |
CN117011785A (en) * | 2023-07-06 | 2023-11-07 | 华新水泥股份有限公司 | Firework detection method, device and system based on space-time correlation and Gaussian heat map |
CN117011785B (en) * | 2023-07-06 | 2024-04-05 | 华新水泥股份有限公司 | Firework detection method, device and system based on space-time correlation and Gaussian heat map |
Also Published As
Publication number | Publication date |
---|---|
CN112052797B (en) | 2024-07-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112052797B (en) | MaskRCNN-based video fire disaster identification method and MaskRCNN-based video fire disaster identification system | |
CN109147254B (en) | Video field fire smoke real-time detection method based on convolutional neural network | |
CN111223088B (en) | Casting surface defect identification method based on deep convolutional neural network | |
CN110298297B (en) | Flame identification method and device | |
CN111598066A (en) | Helmet wearing identification method based on cascade prediction | |
CN111222478A (en) | Construction site safety protection detection method and system | |
CN112598713A (en) | Offshore submarine fish detection and tracking statistical method based on deep learning | |
CN105404847A (en) | Real-time detection method for object left behind | |
CN114639075B (en) | Method and system for identifying falling object of high altitude parabola and computer readable medium | |
CN110490043A (en) | A kind of forest rocket detection method based on region division and feature extraction | |
CN112364740B (en) | Unmanned aerial vehicle room monitoring method and system based on computer vision | |
CN113449606B (en) | Target object identification method and device, computer equipment and storage medium | |
WO2023273010A1 (en) | High-rise littering detection method, apparatus, and device, and computer storage medium | |
CN113033523B (en) | Method and system for constructing falling judgment model and falling judgment method and system | |
CN107944403A (en) | Pedestrian's attribute detection method and device in a kind of image | |
CN111062974A (en) | Method and system for extracting foreground target by removing ghost | |
CN114399734A (en) | Forest fire early warning method based on visual information | |
CN112115878B (en) | Forest fire smoke root node detection method based on smoke area density | |
CN112417955A (en) | Patrol video stream processing method and device | |
CN113177439B (en) | Pedestrian crossing road guardrail detection method | |
CN115311623A (en) | Equipment oil leakage detection method and system based on infrared thermal imaging | |
KR101690050B1 (en) | Intelligent video security system | |
CN111275733A (en) | Method for realizing rapid tracking processing of multiple ships based on deep learning target detection technology | |
CN117475353A (en) | Video-based abnormal smoke identification method and system | |
CN117557937A (en) | Monitoring camera image anomaly detection method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |