CN108537215B - Flame detection method based on image target detection - Google Patents

Flame detection method based on image target detection Download PDF

Info

Publication number
CN108537215B
CN108537215B CN201810243014.4A CN201810243014A CN108537215B CN 108537215 B CN108537215 B CN 108537215B CN 201810243014 A CN201810243014 A CN 201810243014A CN 108537215 B CN108537215 B CN 108537215B
Authority
CN
China
Prior art keywords
picture
frame
model
flame
standard size
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810243014.4A
Other languages
Chinese (zh)
Other versions
CN108537215A (en
Inventor
赵劲松
吴昊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201810243014.4A priority Critical patent/CN108537215B/en
Publication of CN108537215A publication Critical patent/CN108537215A/en
Application granted granted Critical
Publication of CN108537215B publication Critical patent/CN108537215B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/22Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Quality & Reliability (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a flame detection method based on image target detection, and belongs to the field of image processing, fire detection and video monitoring. Firstly, establishing a flame detection data set containing flame pictures and marking information of each picture, and dividing the flame detection data set into a training set and a test set; constructing a deep convolutional neural network model, performing iterative update on the model by using a training set, calculating a loss function on the updated model by using a test set, and finishing the model training if the loss function of the current model is not reduced any more; and shooting a real-time video, detecting each frame of picture by using the trained model, and if flame exists, outputting the coordinate position of the flame in the picture by using the model and marking the flame by using a rectangular frame. According to the invention, the candidate area of suspected flame is generated without manually designing features, and the flame detection can be directly carried out on the whole picture by using the deep convolutional neural network to obtain the position information of the flame, so that the occurrence of fire is early warned, and the harm caused by the fire is reduced to the maximum extent.

Description

Flame detection method based on image target detection
Technical Field
The invention relates to the field of image processing, fire detection and video monitoring, in particular to a flame detection method based on image target detection.
Background
The fire disaster is a frequent disaster seriously harming the life and property safety of people, early warning of the occurrence of the fire disaster can avoid the spread of the fire disaster, and the harm brought by the fire disaster is reduced to the maximum extent. For flame detection tasks, especially for outdoor and large indoor spaces and other scenes, common physical sensors cannot be effectively completed, so that the flame detection method based on the image processing technology is an important research direction in the field of safety engineering at present.
The development of flame detection methods based on image processing techniques is mainly in two stages. The first stage is to judge whether flame exists in the image according to visual characteristics of flame such as color, shape, texture and the like and by combining a classification algorithm in pattern recognition. The methods for manually designing the characteristics have high detection speed, but have poor stability and generalization capability, false alarm is easy to occur when detecting objects with the color similar to that of flame, and false alarm is easy to occur when detecting flame with the color exceeding a set threshold value. To address the problems with such approaches, researchers have recently begun to utilize deep convolutional neural networks for flame detection. The deep convolutional neural network is the most widely studied and applied method in the field of computer vision in recent five years, and features are generally automatically extracted from an image by using a convolutional layer, a pooling layer, a Batch Normalization (BN) layer and the like, and classified by using a full connection layer to judge whether flames exist in the image. The method has the advantages that the tasks of image feature extraction and mode classification can be realized by utilizing the neural network without manually designing features. Compared with the traditional flame detection method, the technology based on the deep convolutional neural network has higher accuracy, recall rate and higher generalization capability. In recent years, with the development of hardware computing power of CPUs, GPUs and the like, the computing speed of the method can meet the real-time requirement of flame detection. The existing flame detection method based on the deep convolutional neural network has the problems that a candidate area of suspected flame still needs to be generated from an original image, and then the neural network is used for classifying the candidate area to judge whether flame exists or not.
At present, a flame detection method based on a deep convolutional neural network is provided, and the method comprises the steps of firstly extracting a suspected flame point in a color image, taking the suspected flame point as a foreground to obtain a binary image, obtaining a series of connected regions by using a connected region method, obtaining candidate regions through screening, and then carrying out classification judgment on whether flames exist in the candidate regions by using the convolutional neural network. The problem with this approach is that the generation of the suspected flame candidate area is still accomplished by a manual design feature. This process still requires the use of the color and dynamic characteristics of the flame, which results in poor adaptability of its detection method to environmental changes. To date, no method has emerged that can achieve image flame detection completely independent of manually designed features.
Disclosure of Invention
The invention aims to overcome the defects of the existing method and provides a flame detection method based on image target detection. According to the method, the candidate area of the suspected flame is generated without using manual design characteristics, the flame of the whole picture can be directly detected by using the deep convolutional neural network, and the position information of the flame is obtained, so that the occurrence of fire is early warned, and the harm caused by the fire is reduced to the maximum extent.
The invention provides a flame detection method based on image target detection, which is characterized by comprising the following steps of:
1) collecting N color pictures with flame pictures to establish a flame picture data set Fimage,N≥5000;
2) For the flame picture data set F established in the step 1)imageManually marking the flame area position coordinates of each picture by using a rectangular frame to obtain a real marking frame of the flame area position and 4 coordinate values of the real marking frame in each picture, wherein the 4 coordinate values are as follows:
Figure BDA0001605794610000021
1, 2, N, wherein
Figure BDA0001605794610000022
The position coordinates of the top left corner vertex of the frame are really marked for the first picture,
Figure BDA0001605794610000023
the position coordinates of the top point of the lower right corner of the frame are really marked for the first picture; storing the information of the real marking frame of each picture as an xml file, wherein the xml files corresponding to all the flame pictures form a flame marking data set Fannotation;FimageAnd FannotationForming a flame detection dataset F0
3) Detecting the flame obtained in step 2) by using a data set F0Random partitioning into training sets FtrainAnd test set FtestWherein the training set FtrainOccupied flame detection dataset F0The ratio of a is more than or equal to 0.6 and less than or equal to 0.9;
4) constructing a deep convolution neural network model and training the model to obtain a trained flame detection model; the method comprises the following specific steps:
4-1) carrying out standardization and normalization processing on each picture of the training set and the test set;
standardizing each picture of the training set and the test set established in the step 1), wherein the standard width and the standard height are respectively W and H, W is more than or equal to 400 and less than or equal to 800, H is more than or equal to 400 and less than or equal to 800, and the channel number Ch is 3; taking the pixel value of the t channel and the coordinate (x, y) of any image after the normalization processing as I (t, x, y), wherein I (t, x, y) is more than or equal to 0 and less than or equal to 255, calculating the normalized pixel value I '(t, x, y) by using the formula (1), wherein I' (t, x, y) is more than or equal to 0 and less than or equal to 1:
Figure BDA0001605794610000024
4-2) constructing a deep convolutional neural network model, wherein the model comprises an input layer, a feature extraction part and an output layer which are sequentially connected; the characteristic extraction part is formed by combining convolution layers and pooling layers and comprises 3-6 pooling layers, wherein each pooling layer comprises 1-5 convolution layers in front of the pooling layer, and a batch standard layer is added behind each convolution layer;
4-3) training the model established in the step 4-2) to obtain a trained flame detection model; the method comprises the following specific steps:
4-3-1) sequentially inputting each picture of the training set subjected to the standardization and normalization processing in the step 4-1) into the model established in the step 4-2); inputting the first picture, outputting a feature map corresponding to the picture by the penultimate convolutional layer in the deep convolutional neural network, wherein the feature map is s multiplied by s units, each unit corresponds to different position areas of the input picture, then connecting the output of the penultimate convolutional layer to an output layer of the whole model through the last convolutional layer, outputting the output of the penultimate convolutional layer as k in the s multiplied by s different position areas of the input picture, wherein k is more than or equal to 5 and less than or equal to 9, and the class probability of each prediction candidate frame
Figure BDA0001605794610000031
And coordinate offset relative to corresponding standard size box
Figure BDA0001605794610000032
1≤i≤s2J is more than or equal to 1 and less than or equal to k, wherein i represents the ith position area, and j represents the jth prediction candidate frame of the position area; standard of meritThe width and height of the size frame relative to the whole input picture are represented by FimageThe width and height of the real mark frame of all pictures relative to the original pictures are obtained by clustering by a k-means method, so thatRespectively representing the width and the height of the jth standard size frame relative to the whole input picture; the center of each standard size frame is the center of the picture position area where the standard size frame is located; each standard size frame corresponds to a prediction candidate frame;
wherein,
Figure BDA0001605794610000034
the probability that flame exists in the jth prediction candidate frame in the ith position area of the ith picture is represented, and the value range is 0-1;the coordinate offset of the jth prediction candidate frame center relative to the corresponding standard size frame center in the ith position area of the ith picture is represented, and the value range is 0-1 (delta wij,Δhij) Representing the width and height offset of a jth prediction candidate frame in an ith position area of the ith picture relative to the width and height of a corresponding standard size frame, wherein the value range is 0-1;
4-3-2) dividing s multiplied by k standard size frames of the input picture in the step 4-3-1) into positive and negative samples;
if the real labeled frame in each picture is G, and any standard size frame of the picture is T, the intersection ratio IoU of T and G is calculated as follows:
Figure BDA0001605794610000036
wherein area represents a region;
setting a threshold η of positive and negative samples, wherein the threshold is not less than 0.5 and not more than η and not more than 0.7, and judging s multiplied by k standard size frames corresponding to each picture, wherein if IoU is not less than η, flames exist in the standard size frames, the standard size frames are positive samples, and if IoU is less than η, the standard size frames do not exist, the standard size frames are negative samples;
4-3-3) calculating the error between the model output information and the real annotation information according to the loss function;
outputting the category probability of s multiplied by k prediction candidate frames corresponding to the first picture and the coordinate offset relative to the corresponding standard size frame according to the model obtained in the step 4-3-1), and calculating the picture prediction information
Figure BDA0001605794610000037
Corresponding real label information
Figure BDA0001605794610000038
The error between; the loss function for the ith picture is defined as follows:
wherein,
Figure BDA0001605794610000042
representing the real category corresponding to the jth prediction candidate frame in the ith position area of the ith picture, if the standard size frame corresponding to the prediction candidate frame is a positive sample, then
Figure BDA0001605794610000043
The value is 1; if the standard size frame corresponding to the prediction candidate frame is negative sample, thenThe value is 0;
Figure BDA0001605794610000045
indicating the coordinate offset of the center of the real labeling frame of the ith picture relative to the center of the jth standard size frame of the ith position area,
Figure BDA0001605794610000046
the width and the height of the real marking frame of the first picture are respectively corresponding to the jth standard of the ith position areaThe offset of the width and height of the size frame; relative offset of real marking frame of first picture
Figure BDA0001605794610000047
The calculation expressions are respectively as follows:
Figure BDA0001605794610000048
Figure BDA0001605794610000049
Figure BDA00016057946100000410
Figure BDA00016057946100000411
wherein,respectively represent F corresponding to the first pictureimageThe width and height of the original picture;
Figure BDA00016057946100000413
coordinate information of the jth standard size frame of the ith position area of the ith picture relative to the whole input picture,
Figure BDA00016057946100000414
is the coordinate of the center of the ith position area relative to the whole input picture,
Figure BDA00016057946100000415
obtained by k-means clustering; lambda [ alpha ]locA weight coefficient representing the position error loss relative to the class error loss, 0.01 ≦ λloc≤0.2;
Figure BDA00016057946100000416
Indicates the ith position areaWhether the jth prediction candidate frame is the Boolean value of the flame or not is judged, and if the standard size frame corresponding to the prediction candidate frame is a positive sample, the standard size frame is judged to be the Boolean value of the flame
Figure BDA00016057946100000417
The value is 1; if the standard size frame corresponding to the prediction candidate frame is negative sample, then
Figure BDA00016057946100000418
The value is 0;
Figure BDA00016057946100000419
a Boolean value indicating whether the jth prediction candidate frame of the ith position area is not the flame, if the standard size frame corresponding to the prediction candidate frame is a positive sample, then
Figure BDA00016057946100000420
The value is 0; if the standard size frame corresponding to the prediction candidate frame is negative sample, then
Figure BDA00016057946100000421
The value is 1; lambda [ alpha ]noobjλ is 0.01 ≦ for adjusting the weighting factor of the class loss ratio of the positive and negative samplesnoobj≤0.2;
4-3-4) repeating steps 4-3-1) to 4-3-3), and combining the training set FtrainInputting each picture into the Model established in the step 4-1), updating the Model through an error back propagation algorithm to obtain a current Model and recording the current Model as a Modelold
4-3-5) Using test set FtestFor ModeloldTesting is carried out, and a test set FtestEach picture is input into a Model in sequenceoldCalculating F according to equation (3)testAll pictures in the ModeloldLoss of total Loss ofold
4-3-6) repeat step 4-3-4), training set FtrainEach picture is input into a Model in sequenceoldObtaining a new current Model and marking as a Modelnew
4-3-7) repeat step 4-3-5), using test set FtestFor ModelnewTesting was carried out to give FtestAll pictures in the ModelnewLoss of total Loss ofnew
4-3-8) training stopping condition judgment:
if Lossnew≤LossoldContinuing to train the Model and setting the current ModelnewUpdate to a new ModeloldWill be the current LossnewUpdate to New LossoldAnd returning to the step 4-3-6);
if Lossnew≥LossoldWhen the Model is not trained, the Model stops training and outputs the current ModeloldTaking the flame detection model as a finally trained flame detection model, and entering the step 5);
5) carrying out flame detection on the real-time video by using the flame detection model obtained in the step 4);
5-1) shooting a real-time video, inputting each frame of picture of the real-time video into the flame detection model trained in the step 4) for detection, and outputting the class probability of s multiplied by k prediction candidate frames corresponding to the frame of picture and the coordinate offset relative to the corresponding standard size frame by the model
Figure BDA0001605794610000051
Where the superscript o represents the flame detection model;
5-2) setting a flame judgment threshold value, wherein the value range is 0-1, and judging the category probability of a prediction candidate frame output by each frame of picture: if it is
Figure BDA0001605794610000052
Then no flame is present in the prediction candidate box; if it is
Figure BDA0001605794610000053
The predicted candidate frame has a flame and the amount of coordinate offset of the predicted candidate frame with respect to the corresponding standard-sized frame is determined based on the flame
Figure BDA0001605794610000054
And corresponding standard size frame coordinatesCalculating to obtain the vertex coordinates of the upper left corner and the lower right corner of the actually predicted flame area as
Figure BDA0001605794610000056
Figure BDA0001605794610000057
Figure BDA0001605794610000059
Figure BDA00016057946100000510
Wherein,
Figure BDA00016057946100000511
respectively representing the original width and height of the frame picture;
5-3) if all the prediction candidate frames of the frame have no flame, returning to the step 5-1) again, and carrying out flame detection on the next frame; if any prediction candidate frame has flame, the model outputs the vertex coordinates of the actually predicted flame area in the upper left corner and the lower right corner of the picture, which are obtained by calculation in the step 5-2), and the coordinates are marked by using a rectangular frame, and then the step 5-1) is returned again to continue to detect the flame of the next frame of picture.
The invention has the characteristics and beneficial effects that:
the invention applies the target detection technology in the field of computer vision, is different from other flame detection methods, does not need to extract a candidate area of suspected flame, and converts the classification problem of flame detection into a regression problem. The invention does not depend on the static and dynamic characteristics of the flame needing manual design, can complete the whole detection process from the input of an original image to the output of flame position information through a deep convolution neural network in the whole process, and avoids the situations of false alarm and false alarm of the manually designed characteristics to a certain extent. In addition, the invention can adapt to image input with different sizes, can meet the requirement of real-time detection, has stronger generalization performance in different scenes, and can be applied to fire early warning tasks in various scenes such as indoor scenes, outdoor scenes and the like so as to avoid fire or reduce the harm brought by fire.
Drawings
FIG. 1 is an overall flow chart of the method of the present invention.
FIG. 2 is a diagram of a standard size box, a true label box, and a prediction candidate box according to an embodiment of the invention.
Fig. 3 is a schematic view of a process for labeling a flame picture according to an embodiment of the invention.
Fig. 4 is a schematic diagram of flame detection results in different scenarios according to an embodiment of the present invention.
Detailed Description
The invention provides a flame detection method based on image target detection, which is further described in detail below with reference to the accompanying drawings and specific embodiments.
The invention provides a flame detection method based on image target detection, the overall flow is shown as figure 1, and the method comprises the following steps:
1) collecting N (N is more than or equal to 5000) color pictures with flame pictures, and establishing a flame picture data set Fimage. The picture source has no special requirement, the pictures of the embodiment are all collected from the network, the pictures have no size requirement, and the pictures are required to have obvious flame pictures.
2) For the flame picture data set F established in the step 1)imageThe position coordinates of the flame area of each picture are manually marked by a rectangular frame (the rectangular frame is the actual position of the flame area in the picture and is framed by the rectangular frame to obtain 4 coordinate values of the rectangular frame), and the real marking frame of the flame area position and the 4 coordinate values of the real marking frame in each picture are obtained as follows:
Figure BDA0001605794610000061
1, 2, N, wherein
Figure BDA0001605794610000062
The position coordinates of the top left corner vertex of the frame are really marked for the first picture,
Figure BDA0001605794610000063
the position coordinates of the top point of the lower right corner of the frame are really marked for the first picture; and storing the flame marking data set as an xml file in a format of a VOC007 data set, wherein each picture corresponds to one xml file, and the xml files corresponding to all the flame pictures form a flame marking data set Fannotation。FimageAnd FannotationForm a complete flame detection data set F0
3) Detecting the flame obtained in step 2) by using a data set F0Random partitioning into training sets FtrainAnd test set Ftest(due to F)imagePictures and F inannotationThe xml files in (1) are in one-to-one correspondence, and the training set F is dividedtrainAnd test set FtestThe images obtained by the middle division and the xml files are in one-to-one correspondence), wherein the training set FtrainOccupied flame detection dataset F0A is more than or equal to 0.6 and less than or equal to 0.9, and a training set FtrainParameters to train the model, test set FtestTo test the generalization performance of the model.
4) Constructing a deep convolution neural network model and training the model to obtain a trained flame detection model; the method comprises the following specific steps:
4-1) carrying out standardization and normalization processing on each picture of the training set and the test set;
because the size of the collected picture in the step 1) is any size, and the input of the convolutional neural network needs to fix the size of the picture, the original picture needs to be scaled to a standard size, the standard width and height are respectively W and H, W is more than or equal to 400 and less than or equal to 800, H is more than or equal to 400 and less than or equal to 800, and Ch is 3 without changing the number of channels. In the training and testing process, the value ranges of all pixel values of the input picture sample are all between 0 and 255, and normalization processing is needed to make the pixel values between 0 and 1. Taking the pixel value of the t channel and the coordinate (x, y) of any normalized picture as I (t, x, y), wherein I (t, x, y) is more than or equal to 0 and less than or equal to 255, calculating the normalized pixel value I '(t, x, y) by using a formula (1), wherein I' (t, x, y) is more than or equal to 0 and less than or equal to 1:
Figure BDA0001605794610000071
4-2) constructing a deep convolutional neural network model, which comprises an input layer, a feature extraction part and an output layer which are sequentially connected; wherein, the normalized picture is input from the input layer and connected with the feature extraction part; the characteristic extraction part is formed by combining a convolution layer and a pooling layer, wherein the pooling layer comprises 3-6 pooling layers, 1-5 convolution layers are arranged in front of each pooling layer, and a batch standard layer is added behind each convolution layer and is used for normalizing each batch of input data in an output characteristic diagram of the convolution layer; and finally, the characteristic extraction part is connected with the output layer to output a prediction result.
4-3) training the model established in the step 4-2) to obtain a trained flame detection model; the method comprises the following specific steps:
4-3-1) randomly selecting a picture from the training set after the standardization and normalization processing in the step 4-1) and inputting the picture into the model established in the step 4-2); let the input be the l-th picture, the feature map corresponding to the picture is output by the penultimate convolutional layer in the deep convolutional neural network, the feature map size corresponding to each picture is s × s units (the output feature map size s × s is determined by the width and height (W, H) of the input picture and the designed network structure), each unit corresponds to a different position region of the input picture, as shown in fig. 2, the large map on the left side of fig. 2 is the input picture, the small map on the right side is the feature map corresponding to the picture, when s is 3, the feature map includes 9 units, which correspond to 9 different position regions in the picture, and the uppermost-left-corner unit of the feature map corresponds to the uppermost-left-corner position region of the input picture. Then the output of the last convolution layer is connected to the output layer of the whole model through the last convolution layer, and the output is the category outline of k prediction candidate frames with k being more than or equal to 5 and less than or equal to 9 in different positions (s multiplied by s) of the input pictureRate of change
Figure BDA0001605794610000072
And coordinate offset relative to corresponding standard size box1≤i≤s2J is more than or equal to 1 and less than or equal to k, wherein i represents the ith position area, and j represents the jth prediction candidate frame of the position area. As shown in fig. 2, the solid line frame in fig. 2 is a real labeled frame, only one frame in fig. 2 is shown, the dotted frame represents a prediction candidate frame, each position region has k equal to 5 prediction candidate frames, the dashed frame represents a standard size frame, each position region has k equal to 5 standard size frames, and the relative size of the standard size frame (i.e., the width and height of the standard size frame relative to the whole input picture, such that the standard size frame corresponds to the whole input picture, and the reference size frame indicates the relative size of the prediction candidate frame
Figure BDA0001605794610000082
Respectively representing the width and height of the jth standard-size frame relative to the whole input picture) is represented by FimageThe width and height of the real marking frames of all the pictures relative to the original pictures (the original pictures are the pictures before standardization processing) are obtained by clustering through a k-means method, and the center of each standard size frame is the center of the picture position area where the standard size frame is located; and each standard size frame corresponds to one prediction candidate frame, and the coordinates of the prediction candidate frame can be calculated through the coordinate offset and the coordinates of the standard size frame.
Wherein
Figure BDA0001605794610000083
The probability of flame existing in the jth prediction candidate frame of the ith position area of the ith picture is represented, the value range is 0-1,
Figure BDA0001605794610000084
closer to 1 indicates a greater likelihood of flame being present in the prediction candidate box;indicates the ith position area of the first pictureCoordinate offset of the jth prediction candidate frame center relative to the corresponding standard size frame center, (Δ w)ij,Δhij) And the width and the height of the jth prediction candidate frame in the ith position area of the ith picture are represented by the offset of the width and the height of the jth prediction candidate frame relative to the width and the height of the corresponding standard size frame, and the value ranges are 0-1.
In actual operation, the input of the model is a batch of batch, the image data with the size of batch being 8 or more and the size of batch being 64 or less, and the size of batch is batch × W × H × Ch.
4-3-2) dividing s multiplied by k standard size frames of the input picture in the step 4-3-1) into positive and negative samples;
and (3) if the real marked frame in each picture is G, and the frame with any standard size of the picture is T, the Intersection-over-Unit ratio of T and G is (IoU):
wherein area represents a region;
setting a threshold η of positive and negative samples, wherein the threshold is more than or equal to 0.5 and less than or equal to η and less than or equal to 0.7, and judging s multiplied by k standard size frames corresponding to each picture, if IoU is more than or equal to η, the standard size frames have flames and are considered as positive samples, and if IoU is less than η, the standard size frames have no flames and are considered as negative samples.
4-3-3) calculating the error between the model output information and the real annotation information according to the loss function;
outputting the category probability of s multiplied by k prediction candidate frames corresponding to the first picture and the coordinate offset relative to the corresponding standard size frame according to the model obtained in the step 4-3-1), and calculating the picture prediction information
Figure BDA0001605794610000087
Corresponding real label information
Figure BDA0001605794610000088
Error between, loss of pictureThe numbers are defined as follows:
Figure BDA0001605794610000091
wherein,
Figure BDA0001605794610000092
indicating the real type corresponding to the jth prediction candidate frame in the ith position area of the ith picture (the same as the real type corresponding to the corresponding standard size frame), and if the standard size frame corresponding to the prediction candidate frame is a positive sample, determining that the current prediction candidate frame is a positive sample
Figure BDA0001605794610000093
The value is 1; if the standard size frame corresponding to the prediction candidate frame is negative sample, then
Figure BDA0001605794610000094
The value is 0;
Figure BDA0001605794610000095
indicating the coordinate offset of the center of the real labeling frame of the ith picture relative to the center of the jth standard size frame of the ith position area,
Figure BDA0001605794610000096
the offset of the width and height of the actual marking frame of the ith picture relative to the width and height of the jth standard size frame of the ith position area (if the standard size frame is a positive sample, namely
Figure BDA0001605794610000097
The coordinate information of the real label box makes sense). Relative offset of real label box
Figure BDA0001605794610000098
From FannotationThe marking information in the xml file is obtained by calculation, for example, the marking information of the flame area of the first picture isThe coordinate information of the jth standard size frame of the ith position area relative to the whole input picture is
Figure BDA00016057946100000910
(
Figure BDA00016057946100000911
Obtained by clustering with k-means,
Figure BDA00016057946100000912
i.e. the coordinates of the center of the ith position area relative to the whole input picture), the relative offset of the real mark frame of the ith picture
Figure BDA00016057946100000913
The calculation expressions are respectively as follows:
Figure BDA00016057946100000915
Figure BDA00016057946100000916
wherein,
Figure BDA00016057946100000918
respectively represent F corresponding to the first pictureimageThe width and height of the original picture; lambda [ alpha ]locThe weight coefficient indicating the position error loss relative to the class error loss can be generally set to 0.01 ≦ λloc≤0.2;
Figure BDA00016057946100000919
To representWhether the jth prediction candidate frame of the ith position region is a Boolean value of the flame (a positive sample is 1, and a negative sample is 0);
Figure BDA00016057946100000920
boolean values (positive sample 0, negative sample 1) indicating whether the ith cell, the jth prediction candidate box, is not a flame; since the number of candidate regions for negative samples is much larger than that for positive samples in the whole picture,
Figure BDA0001605794610000101
the loss of terms will be much greater than
Figure BDA0001605794610000102
Thus introducing a weighting factor lambdanoobjTo adjust the class loss ratio of the positive and negative samples, it can be set to 0.01 ≦ λnoobj≤0.2。
4-3-4) repeating steps 4-3-1) to 4-3-3), and combining the training set FtrainInputting each picture into the Model established in the step 4-1), carrying out iterative training on the whole neural network Model, updating the Model through an error back propagation algorithm, and obtaining a current Model and recording the current Model as a Modelold
4-3-5) Using test set FtestFor ModeloldTesting is carried out, and a test set FtestEach picture is input into a Model in sequenceoldCalculating F according to equation (3)testAll pictures in the ModeloldLoss of total Loss ofold
4-3-6) repeat step 4-3-4), training set FtrainEach picture is input into a Model in sequenceoldObtaining a new current Model and marking as a Modelnew
4-3-7) repeat step 4-3-5), using test set FtestFor ModelnewTesting was carried out to give FtestAll pictures in the ModelnewLoss of total Loss ofnew
4-3-8) training stopping condition judgment:
if Lossnew≤LossoldTo account for test errorsStill descending, ModelnewIn test set FtestThe performance of the system is superior to that of a ModeloldContinuing to train the Model and setting the current ModelnewUpdate to a new ModeloldWill be the current LossnewUpdate to New LossoldAnd returning to the step 4-3-6);
if Lossnew≥LossoldWhen the test error begins to rise, the ModeloldIn test set FtestThe performance of the system is superior to that of a ModelnewIf so, the Model stops training and outputs the current ModeloldAnd 5) as a finally trained flame detection model, entering the step 5).
5) Carrying out flame detection on the real-time video by using the flame detection model obtained in the step 4);
5-1) shooting a real-time video, inputting each frame of picture of the real-time video into the flame detection model trained in the step 4) for detection, and outputting the class probability of s multiplied by k prediction candidate frames corresponding to the frame of picture and the coordinate offset relative to the corresponding standard size frame by the model
Figure BDA0001605794610000103
Where the superscript o represents the flame detection model;
5-2) setting a flame judgment threshold value, wherein the value range is 0-1, and judging the category probability of a prediction candidate frame output by each frame of picture: if it isThen no flame is present in the prediction candidate box; if it is
Figure BDA0001605794610000105
The predicted candidate frame has a flame and the amount of coordinate offset of the predicted candidate frame with respect to the corresponding standard-sized frame is determined based on the flame
Figure BDA0001605794610000106
And corresponding standard size frame coordinates
Figure BDA0001605794610000107
Calculating to obtain the vertex coordinates of the upper left corner and the lower right corner of the actually predicted flame area as
Figure BDA0001605794610000108
Figure BDA0001605794610000111
Figure BDA0001605794610000112
Figure BDA0001605794610000113
Figure BDA0001605794610000114
Wherein,
Figure BDA0001605794610000115
respectively representing the original width and height of the frame picture;
5-3) if all the prediction candidate frames of the frame have no flame, returning to the step 5-1) again, and carrying out flame detection on the next frame; if any prediction candidate frame has flame, the model outputs the vertex coordinates of the actually predicted flame area in the upper left corner and the lower right corner in the picture, which are obtained by calculation in the step 5-2), and the coordinates are marked by using a rectangular frame, and then the step 5-1) is returned again), the flame detection of the next frame of picture is continued, and then the marked picture and the fire alarm signal can be sent to the control center.
The above method can be implemented by programming by those skilled in the art.
One embodiment of the present invention is as follows:
the flame detection method based on image target detection provided by the embodiment comprises the following steps:
1) collecting N (N is 5075) color pictures with flame pictures, and constructingVertical flame picture dataset Fimage. The collected pictures are all gathered from the network, and all the pictures contain obvious flame pictures.
2) For the flame picture dataset F of step 1)imageManually marking the coordinates of the flame area position of each picture by using a rectangular frame, as shown in (a) and (b) of fig. 3, framing the flame area by using the rectangular frame, and obtaining a real marking frame of the flame area position and 4 coordinate values of the real marking frame in each picture, wherein the coordinate values are as follows:
Figure BDA0001605794610000116
1, 2, N, whereinThe position coordinates of the top left corner vertex of the frame are really marked for the first picture,the position coordinates of the top point of the lower right corner of the frame are really marked for the first picture; and storing the labeling information as xml files in the format of a VOC007 data set, wherein each picture corresponds to one xml file, and the xml files corresponding to all flame pictures form a flame labeling data set Fannotation。FimageAnd FannotationForm a complete flame detection data set F0
3) Detecting the flame data set F obtained in the step 2)0Random partitioning into training sets FtrainAnd test set Ftest(due to F)imagePictures and F inannotationThe xml files in (1) are in one-to-one correspondence, and the training set F is dividedtrainAnd test set FtestThe images obtained by the middle division and the xml files are in one-to-one correspondence), wherein the training set FtrainOccupied flame detection dataset F0In a proportion of 80% to FtrainIncluding 4060 pictures and corresponding label data, FtestIncluding 1015 pictures and corresponding labeled data, training set FtrainParameters to train the model, test set FtestTo test the generalization performance of the model.
4) Constructing a deep convolution neural network model and training the model to obtain a trained flame detection model; the method comprises the following specific steps:
4-1) carrying out standardization and normalization processing on each picture of the training set and the test set;
in this embodiment, the standard width and height are set to W and H (W416, H416), and the number Ch of channels of the color picture is 3. And (3) carrying out normalization processing on the pixel data of the zoomed picture according to a formula (1) to enable the pixel value to be between 0 and 1. The input of the model is picture data of one batch of batch 16, and the input size is batch × W × H × Ch.
4-2) constructing a deep convolutional neural network model, which comprises an input layer, a feature extraction part and an output layer: wherein, the normalized picture is input from the input layer and connected with the feature extraction part; the characteristic extraction part consists of 22 convolutional layers and 5 pooling layers, wherein each pooling layer comprises 1-5 convolutional layers in front, and a batch standard layer is added behind each convolutional layer for normalizing each batch of input data in an output characteristic diagram of the convolutional layer; and finally, the characteristic extraction part is connected with the output layer to output a prediction result.
4-3) training the model established in the step 4-2) to obtain a trained flame detection model; the method comprises the following specific steps:
4-3-1) randomly selecting a picture from the training set after the standardization and normalization processing in the step 4-1) and inputting the picture into the model established in the step 4-2); and if the input is the l picture, outputting a feature map corresponding to the picture by using the penultimate convolutional layer in the deep convolutional neural network, wherein the feature map corresponding to each picture is s × s units (s is 13), and each unit corresponds to a different position area of the input picture. The output of the last convolution layer is connected to the output layer of the whole model through the last convolution layer, and the output is the class probability of k-5 prediction candidate frames in different positions (s multiplied by s) of the input picture
Figure BDA0001605794610000121
And coordinate offset relative to corresponding standard size box
Figure BDA0001605794610000122
1≤i≤s2J is more than or equal to 1 and less than or equal to k, wherein i represents the ith position area, and j represents the jth prediction candidate frame of the position area. By making a pair FimageK-means clustering is carried out on the real marking frames of all the pictures relative to the width and the height of the original pictures to obtain the relative width of 5 standard-size frames
Figure BDA0001605794610000123
And height
Figure BDA0001605794610000124
Comprises the following steps: (0.238,0.292),(0.754,0.819),(0.325,0.572),(0.094,0.117),(0.597,0.407).
4-3-2) dividing the s × s × k standard size frames of the input picture in the step 4-3-1) into positive and negative samples, setting the intersection ratio threshold η of the positive and negative samples to be 0.6, and judging the s × s × k standard size frames corresponding to each picture, wherein if IoU is not less than η, flames exist in the standard size frames and are considered as positive samples, and if IoU is less than η, flames do not exist in the standard size frames and are considered as negative samples.
4-3-3) calculating the error between the model output information and the real annotation information according to the loss function;
outputting the category probability of s multiplied by k prediction candidate frames corresponding to the first picture and the coordinate offset relative to the corresponding standard size frame according to the model obtained in the step 4-3-1), and calculating the picture prediction information
Figure BDA0001605794610000125
Corresponding real label informationThe loss function of the first picture is defined in formula (3).
Wherein the weight coefficient of the position error loss relative to the class error loss is set as lambdaloc0.2, weight coefficient setting for adjusting class loss ratio of positive and negative samplesIs λnoobj=0.2。
4-3-4) repeating steps 4-3-1) to 4-3-3), and combining the training set FtrainInputting each picture into the Model established in the step 4-1), carrying out iterative training on the whole neural network Model, updating the Model through an error back propagation algorithm, and obtaining a current Model and recording the current Model as a Modelold
4-3-5) Using test set FtestFor ModeloldTesting is carried out, and a test set FtestEach picture is input into a Model in sequenceoldCalculating F according to equation (3)testAll pictures in the ModeloldLoss of total Loss ofold
4-3-6) repeat step 4-3-4), training set FtrainEach picture is input into a Model in sequenceoldObtaining a new current Model and marking as a Modelnew
4-3-7) repeat step 4-3-5), using test set FtestFor ModelnewTesting was carried out to give FtestAll pictures in the ModelnewLoss of total Loss ofnew
4-3-8) training stopping condition judgment:
if Lossnew≤LossoldModel, which shows that the test error is still decreasingnewIn test set FtestThe performance of the system is superior to that of a ModeloldContinuing to train the Model and setting the current ModelnewUpdate to a new ModeloldWill be the current LossnewUpdate to New LossoldAnd returning to the step 4-3-6);
if Lossnew≥LossoldWhen the test error begins to rise, the ModeloldIn test set FtestThe performance of the system is superior to that of a ModelnewIf so, the Model stops training and outputs the current ModeloldAnd 5) as a finally trained flame detection model, entering the step 5).
5) Carrying out flame detection on the real-time video by using the flame detection model obtained in the step 4);
5-1) shooting a real-time video, and inputting each frame of picture of the real-time video into the training in the step 4)Detecting by a good flame detection model, and outputting the class probability of s × s × k prediction candidate frames corresponding to the frame and the coordinate offset relative to the standard size frame by the model
Figure BDA0001605794610000131
Where the superscript o represents the flame detection model.
5-2) setting a flame judgment threshold value, wherein the value range is 0-1, and judging the category probability of a prediction candidate frame output by each frame of picture: if it is
Figure BDA0001605794610000132
Then no flame is present in the prediction candidate box; if it is
Figure BDA0001605794610000133
The predicted candidate frame has a flame and the coordinate offset of the predicted candidate frame with respect to the corresponding standard-sized frame is used
Figure BDA0001605794610000134
And corresponding standard size frame coordinates
Figure BDA0001605794610000135
Calculating the vertex coordinates of the upper left corner and the lower right corner of the actually predicted flame area according to the formulas (8) to (11) to obtain
5-3) if all the prediction candidate frames of the frame have no flame, returning to the step 5-1) again, and carrying out flame detection on the next frame; if any prediction candidate frame has flame, the model outputs the vertex coordinates of the actually predicted flame area in the upper left corner and the lower right corner of the picture, which are calculated in the step 5-2), and the coordinates are marked by using a rectangular frame, as shown in (a) to (e) in fig. 4, and then the step 5-1 is returned again, the flame detection of the next frame of picture is continued, and then the marked picture and the fire alarm signal can be sent to the control center.

Claims (1)

1. A flame detection method based on image target detection is characterized by comprising the following steps:
1) collecting N color pictures with flame pictures to establish a flame picture data set Fimage,N≥5000;
2) For the flame picture data set F established in the step 1)imageManually marking the flame area position coordinates of each picture by using a rectangular frame to obtain a real marking frame of the flame area position and 4 coordinate values of the real marking frame in each picture, wherein the 4 coordinate values are as follows:
Figure FDA0002165188360000011
wherein
Figure FDA0002165188360000012
The position coordinates of the top left corner vertex of the frame are really marked for the first picture,
Figure FDA0002165188360000013
the position coordinates of the top point of the lower right corner of the frame are really marked for the first picture; storing the information of the real marking frame of each picture as an xml file, wherein the xml files corresponding to all the flame pictures form a flame marking data set Fannotation;FimageAnd FannotationForming a flame detection dataset F0
3) Detecting the flame obtained in step 2) by using a data set F0Random partitioning into training sets FtrainAnd test set FtestWherein the training set FtrainOccupied flame detection dataset F0The ratio of a is more than or equal to 0.6 and less than or equal to 0.9;
4) constructing a deep convolution neural network model and training the model to obtain a trained flame detection model; the method comprises the following specific steps:
4-1) carrying out standardization and normalization processing on each picture of the training set and the test set;
standardizing each picture of the training set and the test set established in the step 3), setting the standard width and height as W and H respectively, setting W to be more than or equal to 400 and less than or equal to 800,400 and less than or equal to 800, and setting the channel number Ch to be 3; taking the pixel value of the t channel and the coordinate (x, y) of any image after the normalization processing as I (t, x, y), wherein I (t, x, y) is more than or equal to 0 and less than or equal to 255, calculating the normalized pixel value I '(t, x, y) by using the formula (1), wherein I' (t, x, y) is more than or equal to 0 and less than or equal to 1:
4-2) constructing a deep convolutional neural network model, wherein the model comprises an input layer, a feature extraction part and an output layer which are sequentially connected; the characteristic extraction part is formed by combining convolution layers and pooling layers and comprises 3-6 pooling layers, wherein each pooling layer comprises 1-5 convolution layers in front of the pooling layer, and a batch standard layer is added behind each convolution layer;
4-3) training the model established in the step 4-2) to obtain a trained flame detection model; the method comprises the following specific steps:
4-3-1) sequentially inputting each picture of the training set subjected to the standardization and normalization processing in the step 4-1) into the model established in the step 4-2); inputting the input image as the first image, outputting a characteristic image corresponding to the image by the penultimate convolutional layer in the deep convolutional neural network, wherein the size of the characteristic image is s × s units, each unit corresponds to different position areas of the input image, then connecting the output of the penultimate convolutional layer to an output layer of the whole model through the last convolutional layer, and outputting the output of the penultimate convolutional layer as the class probability of k prediction candidate frames in the s × s different position areas of the input imageAnd coordinate offset relative to corresponding standard size box
Figure FDA0002165188360000021
Wherein i represents the ith position area, and j represents the jth prediction candidate frame of the position area; the width and height of the standard size frame relative to the whole input picture are represented by FimageThe width and height of the real mark frame of the whole picture relative to the original picture are passed throughObtained by k-means clustering, the order
Figure FDA0002165188360000022
Respectively representing the width and the height of the jth standard size frame relative to the whole input picture; the center of each standard size frame is the center of the picture position area where the standard size frame is located; each standard size frame corresponds to a prediction candidate frame;
wherein,
Figure FDA0002165188360000023
the probability that flame exists in the jth prediction candidate frame in the ith position area of the ith picture is represented, and the value range is 0-1;
Figure FDA0002165188360000024
the coordinate offset of the jth prediction candidate frame center relative to the corresponding standard size frame center in the ith position area of the ith picture is represented, and the value range is 0-1 (delta wij,Δhij) Representing the width and height offset of a jth prediction candidate frame in an ith position area of the ith picture relative to the width and height of a corresponding standard size frame, wherein the value range is 0-1;
4-3-2) dividing s multiplied by k standard size frames of the input picture in the step 4-3-1) into positive and negative samples;
if the real labeled frame in each picture is G, and any standard size frame of the picture is T, the intersection ratio IoU of T and G is calculated as follows:
Figure FDA0002165188360000025
wherein area represents a region;
setting a threshold η of positive and negative samples, wherein the threshold is not less than 0.5 and not more than η and not more than 0.7, and judging s multiplied by k standard size frames corresponding to each picture, wherein if IoU is not less than η, flames exist in the standard size frames, the standard size frames are positive samples, and if IoU is less than η, the standard size frames do not exist, the standard size frames are negative samples;
4-3-3) calculating the error between the model output information and the real annotation information according to the loss function;
outputting the category probability of s multiplied by k prediction candidate frames corresponding to the first picture and the coordinate offset relative to the corresponding standard size frame according to the model obtained in the step 4-3-1), and calculating the picture prediction information
Figure FDA0002165188360000026
Corresponding real label information
Figure FDA0002165188360000027
The error between; the loss function for the ith picture is defined as follows:
Figure FDA0002165188360000028
wherein,representing the real category corresponding to the jth prediction candidate frame in the ith position area of the ith picture, if the standard size frame corresponding to the prediction candidate frame is a positive sample, then
Figure FDA0002165188360000031
The value is 1; if the standard size frame corresponding to the prediction candidate frame is negative sample, thenThe value is 0;
Figure FDA0002165188360000033
indicating the coordinate offset of the center of the real labeling frame of the ith picture relative to the center of the jth standard size frame of the ith position area,the width and the height of the real marking frame of the first picture are respectively relative to the width of the jth standard size frame of the ith position areaAnd the offset of the height; relative offset of real marking frame of first pictureThe calculation expressions are respectively as follows:
Figure FDA0002165188360000037
Figure FDA0002165188360000038
wherein,
Figure FDA00021651883600000310
respectively represent F corresponding to the first pictureimageThe width and height of the original picture;
Figure FDA00021651883600000311
Figure FDA00021651883600000312
coordinate information of the jth standard size frame of the ith position area of the ith picture relative to the whole input picture,
Figure FDA00021651883600000313
is the coordinate of the center of the ith position area relative to the whole input picture,
Figure FDA00021651883600000314
obtained by k-means clustering; lambda [ alpha ]locIndicating loss of position error0.01 ≦ λ for the weight coefficient of class error lossloc≤0.2;
Figure FDA00021651883600000315
A Boolean value indicating whether the jth prediction candidate frame of the ith position area is a flame or not, if the standard size frame corresponding to the prediction candidate frame is a positive sample, the prediction candidate frame is divided into a plurality of regions
Figure FDA00021651883600000316
The value is 1; if the standard size frame corresponding to the prediction candidate frame is negative sample, then
Figure FDA00021651883600000317
The value is 0;a Boolean value indicating whether the jth prediction candidate frame of the ith position area is not the flame, if the standard size frame corresponding to the prediction candidate frame is a positive sample, then
Figure FDA00021651883600000319
The value is 0; if the standard size frame corresponding to the prediction candidate frame is negative sample, thenThe value is 1; lambda [ alpha ]noobjλ is 0.01 ≦ for adjusting the weighting factor of the class loss ratio of the positive and negative samplesnoobj≤0.2;
4-3-4) repeating steps 4-3-1) to 4-3-3), and combining the training set FtrainInputting each picture into the Model established in the step 4-2), updating the Model through an error back propagation algorithm to obtain a current Model and recording the current Model as a Modelold
4-3-5) Using test set FtestFor ModeloldTesting is carried out, and a test set FtestEach picture is input into a Model in sequenceoldCalculating F according to equation (3)testAll pictures in the ModeloldLoss of total Loss ofold
4-3-6) repeat step 4-3-4), training set FtrainEach picture is input into a Model in sequenceoldObtaining a new current Model and marking as a Modelnew
4-3-7) repeat step 4-3-5), using test set FtestFor ModelnewTesting was carried out to give FtestAll pictures in the ModelnewLoss of total Loss ofnew
4-3-8) training stopping condition judgment:
if Lossnew≤LossoldContinuing to train the Model and setting the current ModelnewUpdate to a new ModeloldWill be the current LossnewUpdate to New LossoldAnd returning to the step 4-3-6);
if Lossnew≥LossoldWhen the Model is not trained, the Model stops training and outputs the current ModeloldTaking the flame detection model as a finally trained flame detection model, and entering the step 5);
5) carrying out flame detection on the real-time video by using the flame detection model obtained in the step 4);
5-1) shooting a real-time video, inputting each frame of picture of the real-time video into the flame detection model trained in the step 4) for detection, and outputting the class probability of s multiplied by k prediction candidate frames corresponding to the frame of picture and the coordinate offset relative to the corresponding standard size frame by the modelWhere the superscript o represents the flame detection model;
5-2) setting a flame judgment threshold value, wherein the value range is 0-1, and judging the category probability of a prediction candidate frame output by each frame of picture: if it is
Figure FDA0002165188360000042
Then no flame is present in the prediction candidate box; if it is
Figure FDA0002165188360000043
The predicted candidate frame has a flame and the amount of coordinate offset of the predicted candidate frame with respect to the corresponding standard-sized frame is determined based on the flame
Figure FDA0002165188360000044
And corresponding standard size frame coordinates
Figure FDA0002165188360000045
Calculating to obtain the vertex coordinates of the upper left corner and the lower right corner of the actually predicted flame area as
Figure FDA0002165188360000046
Figure FDA0002165188360000047
Figure FDA0002165188360000049
Wherein,
Figure FDA00021651883600000411
respectively representing the original width and height of the frame picture;
5-3) if all the prediction candidate frames of the frame have no flame, returning to the step 5-1) again, and carrying out flame detection on the next frame; if any prediction candidate frame has flame, the model outputs the vertex coordinates of the actually predicted flame area in the upper left corner and the lower right corner of the picture, which are obtained by calculation in the step 5-2), and the coordinates are marked by using a rectangular frame, and then the step 5-1) is returned again to continue to detect the flame of the next frame of picture.
CN201810243014.4A 2018-03-23 2018-03-23 Flame detection method based on image target detection Active CN108537215B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810243014.4A CN108537215B (en) 2018-03-23 2018-03-23 Flame detection method based on image target detection

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810243014.4A CN108537215B (en) 2018-03-23 2018-03-23 Flame detection method based on image target detection

Publications (2)

Publication Number Publication Date
CN108537215A CN108537215A (en) 2018-09-14
CN108537215B true CN108537215B (en) 2020-02-21

Family

ID=63483653

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810243014.4A Active CN108537215B (en) 2018-03-23 2018-03-23 Flame detection method based on image target detection

Country Status (1)

Country Link
CN (1) CN108537215B (en)

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109360369B (en) * 2018-09-19 2021-09-28 王杰 Method and device for analyzing fire hazard based on clustering result
CN109522819B (en) * 2018-10-29 2020-08-18 西安交通大学 Fire image identification method based on deep learning
CN109568123B (en) * 2018-11-02 2021-02-02 广东数相智能科技有限公司 Acupuncture point positioning method based on YOLO target detection
CN109614990A (en) * 2018-11-20 2019-04-12 成都通甲优博科技有限责任公司 A kind of object detecting device
CN111325047A (en) * 2018-12-13 2020-06-23 博西华电器(江苏)有限公司 Cooking safety auxiliary method and system, kitchen appliance and combination thereof
CN109684982B (en) * 2018-12-19 2020-11-20 深圳前海中创联科投资发展有限公司 Flame detection method based on video analysis and combined with miscible target elimination
CN109829429B (en) * 2019-01-31 2022-08-09 福州大学 Security sensitive article detection method based on YOLOv3 under monitoring scene
CN110119682A (en) * 2019-04-04 2019-08-13 北京理工雷科电子信息技术有限公司 A kind of infrared remote sensing Image Fire point recognition methods
CN110059753A (en) * 2019-04-19 2019-07-26 北京朗镜科技有限责任公司 Model training method, interlayer are every recognition methods, device, equipment and medium
CN110287982A (en) * 2019-05-08 2019-09-27 中国科学技术大学 A kind of CT images classification method, device and medium based on convolutional neural networks
US10885625B2 (en) 2019-05-10 2021-01-05 Advanced New Technologies Co., Ltd. Recognizing damage through image analysis
CN110569703B (en) * 2019-05-10 2020-09-01 阿里巴巴集团控股有限公司 Computer-implemented method and device for identifying damage from picture
CN110348390B (en) * 2019-07-12 2023-05-16 创新奇智(重庆)科技有限公司 Training method, computer readable medium and system for flame detection model
CN110728186B (en) * 2019-09-11 2023-04-07 中国科学院声学研究所南海研究站 Fire detection method based on multi-network fusion
CN112560834B (en) * 2019-09-26 2024-05-10 武汉金山办公软件有限公司 Coordinate prediction model generation method and device and pattern recognition method and device
CN110826439A (en) * 2019-10-25 2020-02-21 杭州叙简科技股份有限公司 Electric welding construction detection method based on deep learning image processing
CN111127433B (en) * 2019-12-24 2020-09-25 深圳集智数字科技有限公司 Method and device for detecting flame
CN111178275A (en) * 2019-12-30 2020-05-19 浙江中创天成科技有限公司 Fire detection method based on convolutional neural network
CN111310662B (en) * 2020-02-17 2021-08-31 淮阴工学院 Flame detection and identification method and system based on integrated deep network
CN111353555A (en) * 2020-05-25 2020-06-30 腾讯科技(深圳)有限公司 Label detection method and device and computer readable storage medium
CN112633061B (en) * 2020-11-18 2023-03-24 淮阴工学院 Lightweight FIRE-DET flame detection method and system
CN113033553B (en) * 2021-03-22 2023-05-12 深圳市安软科技股份有限公司 Multi-mode fusion fire detection method, device, related equipment and storage medium
CN113052226A (en) * 2021-03-22 2021-06-29 淮阴工学院 Time-sequence fire identification method and system based on single-step detector
CN113012383B (en) * 2021-03-26 2022-12-30 深圳市安软科技股份有限公司 Fire detection alarm method, related system, related equipment and storage medium
CN113221768A (en) * 2021-05-18 2021-08-06 北京百度网讯科技有限公司 Recognition model training method, recognition method, device, equipment and storage medium
CN113657250A (en) * 2021-08-16 2021-11-16 南京图菱视频科技有限公司 Flame detection method and system based on monitoring video
CN114841920A (en) * 2022-03-29 2022-08-02 清华大学 Flame identification method and device based on image processing and electronic equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102208018A (en) * 2011-06-01 2011-10-05 西安工程大学 Method for recognizing fire disaster of power transmission line based on video variance analysis
CN105678332A (en) * 2016-01-08 2016-06-15 昆明理工大学 Converter steel-making endpoint determination method and system based on flame image CNN recognizing and modeling process
CN106250845A (en) * 2016-07-28 2016-12-21 北京智芯原动科技有限公司 Flame detecting method based on convolutional neural networks and device

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103116746B (en) * 2013-03-08 2016-08-03 中国科学技术大学 A kind of video flame detection method based on multiple features fusion technology
CN104408469A (en) * 2014-11-28 2015-03-11 武汉大学 Firework identification method and firework identification system based on deep learning of image
CN106934404A (en) * 2017-03-10 2017-07-07 深圳市瀚晖威视科技有限公司 A kind of image flame identifying system based on CNN convolutional neural networks
CN107025443A (en) * 2017-04-06 2017-08-08 江南大学 Stockyard smoke monitoring and on-time model update method based on depth convolutional neural networks

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102208018A (en) * 2011-06-01 2011-10-05 西安工程大学 Method for recognizing fire disaster of power transmission line based on video variance analysis
CN105678332A (en) * 2016-01-08 2016-06-15 昆明理工大学 Converter steel-making endpoint determination method and system based on flame image CNN recognizing and modeling process
CN106250845A (en) * 2016-07-28 2016-12-21 北京智芯原动科技有限公司 Flame detecting method based on convolutional neural networks and device

Also Published As

Publication number Publication date
CN108537215A (en) 2018-09-14

Similar Documents

Publication Publication Date Title
CN108537215B (en) Flame detection method based on image target detection
CN110660052B (en) Hot-rolled strip steel surface defect detection method based on deep learning
CN111126136B (en) Smoke concentration quantification method based on image recognition
CN110717481B (en) Method for realizing face detection by using cascaded convolutional neural network
CN111753828B (en) Natural scene horizontal character detection method based on deep convolutional neural network
CN111898406B (en) Face detection method based on focus loss and multitask cascade
CN112288758B (en) Infrared and visible light image registration method for power equipment
CN112541532B (en) Target detection method based on dense connection structure
CN112861635A (en) Fire and smoke real-time detection method based on deep learning
CN111242026B (en) Remote sensing image target detection method based on spatial hierarchy perception module and metric learning
CN116448019B (en) Intelligent detection device and method for quality flatness of building energy-saving engineering
CN107392251A (en) A kind of method that target detection network performance is lifted using category images
WO2023160666A1 (en) Target detection method and apparatus, and target detection model training method and apparatus
CN115147380A (en) Small transparent plastic product defect detection method based on YOLOv5
CN114841920A (en) Flame identification method and device based on image processing and electronic equipment
CN115294109A (en) Real wood board production defect identification system based on artificial intelligence, and electronic equipment
CN111368625B (en) Pedestrian target detection method based on cascade optimization
CN111178275A (en) Fire detection method based on convolutional neural network
CN106960188A (en) Weather image sorting technique and device
CN113450321B (en) Single-stage target detection method based on edge detection
CN113344005B (en) Image edge detection method based on optimized small-scale features
CN114998222A (en) Automobile differential shell surface detection method, electronic equipment and medium
CN114694090A (en) Campus abnormal behavior detection method based on improved PBAS algorithm and YOLOv5
CN114548376A (en) Intelligent transportation system-oriented vehicle rapid detection network and method
CN113780462A (en) Vehicle detection network establishment method based on unmanned aerial vehicle aerial image and application thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant