CN110853019B - Method for detecting and identifying controlled cutter through security check - Google Patents

Method for detecting and identifying controlled cutter through security check Download PDF

Info

Publication number
CN110853019B
CN110853019B CN201911104406.3A CN201911104406A CN110853019B CN 110853019 B CN110853019 B CN 110853019B CN 201911104406 A CN201911104406 A CN 201911104406A CN 110853019 B CN110853019 B CN 110853019B
Authority
CN
China
Prior art keywords
layer
prediction
frame
characteristic diagram
follows
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911104406.3A
Other languages
Chinese (zh)
Other versions
CN110853019A (en
Inventor
张莉
郭瑞鸿
孙欢
杨莹
曹洋
孟俊熙
谭海燕
韩仪洒
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Polytechnic University
Original Assignee
Xian Polytechnic University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Polytechnic University filed Critical Xian Polytechnic University
Priority to CN201911104406.3A priority Critical patent/CN110853019B/en
Publication of CN110853019A publication Critical patent/CN110853019A/en
Application granted granted Critical
Publication of CN110853019B publication Critical patent/CN110853019B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10116X-ray image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30136Metal

Landscapes

  • Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method for detecting and identifying a controlled cutter in security inspection, which is implemented according to the following steps: step 1, carrying out normalization processing on an image A to obtain an image B; step 2, processing the image B by an SSD-ResNet101 model, a feature fusion method, pooling and a Relu function to obtain a feature map C; step 3, calculating a default frame of the characteristic diagram C by adopting a scale calculation formula; matching the default frame with the target frame by adopting a Jaccard overlay matching strategy and model training to obtain a prediction frame; and 4, screening the final prediction box set by adopting a non-maximum suppression algorithm to obtain a detection result. The invention has high detection and identification precision for dangerous goods of the X-ray tube cutter and strong real-time security check.

Description

Method for detecting and identifying controlled cutter through security check
Technical Field
The invention belongs to the technical field of image detection and identification and X-ray image dangerous goods identification, and particularly relates to a method for detecting and identifying a controlled cutter through security inspection.
Background
An accurate automatic detection technology for controlled articles is a main research direction in the field of security inspection, the existing target detection algorithm based on deep learning has the problems of partial small target false detection, missing detection and the like in detection of a tube-making cutter in a security inspection picture, and the overall detection effect is poor. In 2016, akcay S applies a deep convolutional network to the detection of X-ray dangerous goods for the first time in the literature, and in recent years, target detection algorithms based on CNN are mainly divided into two types: single-stage detection and two-stage detection.
The two-stage detection mainly comprises the following steps: R-CNN, fast RCNN, R-FCN, FPN, etc. Compared with the traditional method, the detection algorithm improves the detection precision, but has large calculation amount and slower detection speed. The single-phase detection algorithm mainly comprises YOLO and SSD. The detection speed is improved by YOLO, but the detection precision of the algorithm is low. Compared with other algorithms, the SSD algorithm has good performance, balanced detection precision and speed, strong robustness and mature application in the field of small target detection, but has the defects. The DSSD algorithm proposed by Fu Cheng-Yang et al in 2017 improves SSD and improves the detection precision of the algorithm, but the detection speed is slow.
Disclosure of Invention
The invention aims to provide a method for detecting and identifying a controlled cutter through security check, which has high detection and identification precision on dangerous goods of the X-ray controlled cutter and strong real-time security check.
The technical scheme adopted by the invention is that the method for detecting and identifying the controlled cutter in the security check is implemented according to the following steps:
step 1, carrying out normalization processing on an image A to obtain an image B;
step 2, processing the image B by an SSD-ResNet101 model, a feature fusion method, pooling and a Relu function to obtain a feature map C;
step 3, calculating a default frame of the characteristic diagram C by adopting a scale calculation formula; matching the default frame with the target frame by adopting a Jaccard overlay matching strategy and model training to obtain a prediction frame;
and 4, screening the final prediction box set by adopting a non-maximum suppression algorithm to obtain a detection result.
The invention is also characterized in that:
the specific process of step 1 is as follows:
the image a is normalized using the following formula:
Figure BDA0002270843070000021
in the formula, y i Representing the result of the normalization processing; x is the number of i A matrix of images representing the input; mu is x i The mean value of (a); sigma 2 Is x i The variance of (a); ε is a positive number, typically 0.0001; gamma is a pull-up parameter; beta is an offset parameter.
The specific process of the step 2 is as follows:
inputting image B into the SSD-ResNet101 network; performing convolution operation on the image B by adopting 512 convolution kernels with the size of 3 multiplied by 3 on the C1 layer, and outputting a characteristic diagram of the C1 layer as the input of the C2 layer; performing convolution operation on the C2 layer by using a convolution kernel of 1 multiplied by 1024, and outputting a characteristic diagram of the C2 layer as the input of the C3 layer; performing convolution operation on the C3 layer by using 512 3 multiplied by 3 convolution kernels, and outputting a C3 layer feature map; introducing a C2 layer characteristic diagram into the C1 layer characteristic diagram, and enhancing by using a Relu function to obtain a C4 layer characteristic diagram; introducing a C3 layer characteristic diagram into the C2 layer characteristic diagram, and enhancing by using a Relu function to obtain a C5 layer characteristic diagram; fusing the C4 layer characteristic diagram and the C5 layer characteristic diagram, enhancing by using a Relu function, and outputting a C6 layer characteristic diagram as the input of a C7 layer;
adopting 512 convolution kernels with the size of 3 multiplied by 3 to carry out convolution operation on the C7 layer, and outputting a C7 layer feature diagram as the input of the C8 layer; carrying out convolution operation on the C8 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C8 layer feature map as the input of a C9 layer; performing convolution operation on the C9 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C9 layer feature map as an input of a C10 layer; performing convolution operation on the C10 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C10 layer feature map as an input of a C11 layer; carrying out convolution operation on the C11 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C10 layer feature map as the input of the C11 layer; and performing pooling operation on the C11 layer, and outputting a C12 layer feature map, namely the feature map C.
The Relu function is as follows:
f(x)=max(0,x) (2)
in the formula, x is an input vector from the neural network of the upper layer, and f (x) is an output vector.
The specific process of step 3 is as follows:
selecting characteristic graphs of C6-C12 layers in the characteristic graph C, and calculating default frames of the characteristic graphs of the C6-C12 layers by adopting a scale calculation formula; and matching the default frame with the target frame by using a Jaccard overlay matching strategy to obtain a prediction frame.
The scale calculation formula is as follows:
Figure BDA0002270843070000041
in the formula, S min The scale of a default frame in the C1 layer is set as S min =0.04;S max The scale of a default frame in the top C12 layer is S max =0.49;m=5;
Wherein, different aspect ratios are set for each default frame, and the calculation formula is as follows:
aspect ratio:
Figure BDA0002270843070000042
width:
Figure BDA0002270843070000043
high:
Figure BDA0002270843070000044
in the formula, when the aspect ratio is 1, a default box is added, and the scale is as follows:
Figure BDA0002270843070000045
)。
when the Jaccard overlay strategy is adopted to match the target frame with the default frame, when the Jaccard overlay is larger than the threshold value of 0.5, the default frame is considered to be matched with the target frame; the Jaccard Overlap expression is as follows:
Figure BDA0002270843070000046
in the formula, A represents a set area formed by all pixels in a default frame; b denotes an aggregate area made up of all pixels within the target frame.
Model training is used for reducing the total prediction error; the total error of prediction is obtained by weighted summation of classified and detected errors, and the expression is as follows:
Figure BDA0002270843070000047
wherein L is the total error; l is conf (x, c) is a classification error; l is loc (x, l, g) is detectionAn error; the weight value a =1; x is a matching flag of the prediction box and the target box, x =1 represents matching, and x =0 represents mismatching; c is the score of the class prediction; l is the pixel value of the prediction frame; g is the pixel value of the target frame; if N =0, L =0;
wherein L is conf (x, c) is calculated by a Softmax loss function, and the calculation formula is as follows:
Figure BDA0002270843070000051
wherein, L is adopted loc (x, L, g) is obtained by performing position offset regression calculation by a Smooth L1 function, and the calculation formula is as follows:
Figure BDA0002270843070000052
Figure BDA0002270843070000053
Figure BDA0002270843070000054
Figure BDA0002270843070000055
Figure BDA0002270843070000056
in the formula (I), the compound is shown in the specification,
Figure BDA0002270843070000057
l represents a prediction box; g represents a target frame; d denotes a default box.
The process of non-maxima suppression is as follows:
sorting all the prediction frames from top to bottom according to the scores, judging the Jaccard Over lap of the prediction frame with the highest score and other prediction frames, if the score is larger than the threshold value, discarding other prediction frames, then judging the Jaccard Over lap of the prediction frame with the highest score and other prediction frames, and discarding the prediction frames larger than the threshold value; then judging a prediction frame with the third highest score; and so on until all prediction boxes are traversed.
The invention has the beneficial effects that:
1) The mAP value of the invention on the VOC2007+2012 universal data set is higher than that of the original SSD algorithm;
2) The invention solves the problems of partial small target false detection and missing detection existing in the detection of the tube cutter in the security inspection picture by the current target detection algorithm based on deep learning, and improves the overall detection effect of the X-ray image dangerous goods.
Drawings
FIG. 1 is a comparison graph of detection speed and detection precision of an algorithm and an SSD algorithm, an FSSD algorithm and an ESSD algorithm in the method for detecting and identifying the managed cutter by security inspection of the invention;
FIG. 2 is a diagram showing the experimental results of the method for detecting and identifying the controlled cutting tools in the security inspection of the present invention.
Detailed Description
The present invention will be described in detail below with reference to the accompanying drawings and specific embodiments.
The invention relates to a method for detecting and identifying a controlled cutter in security check, which is implemented according to the following steps:
step 1, carrying out normalization processing on an image A to obtain an image B;
the specific process of step 1 is as follows:
the image a is normalized using the following formula:
Figure BDA0002270843070000061
in the formula, y i Representing the result of the normalization processing; x is the number of i A matrix of images representing the input; mu is x i The mean value of (a); sigma 2 Is x i The variance of (a); ε is a positive number, typically 0.0001; gamma is a tensile stressA parameter; beta is an offset parameter.
Step 2, processing the image B by an SSD-ResNet101 model, a feature fusion method, pooling and a Relu function to obtain a feature map C;
the SSD-ResNet101 network structure adopts a residual error network ResNet101 to replace the original VGGNet as a basic network of the SSD; the model is mainly divided into two parts, wherein a basic network is a residual error ResNet101 network, res3b3 and Res5c are residual error structural layers, conv6_2, new layer 2-New layer6 are newly added auxiliary convolutional layers, and New layer7 is a pooling layer; conv:3 × 3 × 512 represents convolving the feature map using 512 convolution kernels of size 3 × 3;
the specific process of the step 2 is as follows:
inputting image B into the SSD-ResNet101 network; performing convolution operation on the image B by adopting 512 convolution kernels with the size of 3 multiplied by 3 on the C1 layer, and outputting a characteristic diagram of the C1 layer as the input of the C2 layer; performing convolution operation on the C2 layer by using a convolution kernel of 1 multiplied by 1024, and outputting a characteristic diagram of the C2 layer as the input of the C3 layer; performing convolution operation on the C3 layer by using 512 3 multiplied by 3 convolution kernels, and outputting a C3 layer feature map; introducing a C2 layer characteristic diagram into the C1 layer characteristic diagram, and enhancing by using a Relu function to obtain a C4 layer characteristic diagram; introducing a C3 layer characteristic diagram into the C2 layer characteristic diagram, and enhancing by using a Relu function to obtain a C5 layer characteristic diagram; fusing the C4 layer characteristic diagram and the C5 layer characteristic diagram, enhancing by using a Relu function, and outputting a C6 layer characteristic diagram as the input of a C7 layer;
adopting 512 convolution kernels with the size of 3 multiplied by 3 to carry out convolution operation on the C7 layer, and outputting a characteristic diagram of the C7 layer as the input of the C8 layer; performing convolution operation on the C8 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C8 layer feature map as an input of a C9 layer; carrying out convolution operation on the C9 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C9 layer feature map as the input of a C10 layer; performing convolution operation on the C10 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C10 layer feature map as an input of a C11 layer; performing convolution operation on the C11 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C10 layer feature map as an input of the C11 layer; performing pooling operation on the C11 layer, and outputting a C12 layer characteristic diagram, namely a characteristic diagram C;
wherein the Relu function is as follows:
f(x)=max(0,x) (2)
in the formula, x is an input vector from the neural network of the upper layer, and f (x) is an output vector.
Step 3, calculating a default frame of the characteristic diagram C by adopting a scale calculation formula; matching the default frame with the target frame by adopting a Jaccard overlay matching strategy and model training to obtain a prediction frame;
the specific process of step 3 is as follows:
selecting characteristic graphs of C6-C12 layers in the characteristic graph C, and calculating default frames of the characteristic graphs of the C6-C12 layers by adopting a scale calculation formula; and matching the default frame with the target frame by using a Jaccard overlay matching strategy to obtain a prediction frame.
The scale calculation formula is as follows:
Figure BDA0002270843070000081
in the formula, S min The scale of a default frame in the C1 layer is set as S min =0.04;S max The scale of a default frame in the top C12 layer is S max =0.49;m=5;
Wherein, different aspect ratios are set for each default frame, and the calculation formula is as follows:
aspect ratio:
Figure BDA0002270843070000082
width:
Figure BDA0002270843070000083
high:
Figure BDA0002270843070000091
in the formula, when the aspect ratio is 1, a default box is added, and the scale is as follows:
Figure BDA0002270843070000092
)。
when the Jaccard overlay strategy is adopted to match the target frame with the default frame, when the Jaccard overlay is larger than the threshold value of 0.5, the default frame is considered to be matched with the target frame; the Jaccard Overlap expression is as follows:
Figure BDA0002270843070000093
in the formula, A represents a set area formed by all pixels in a default frame; b denotes a collective region composed of all pixels in the target frame.
Model training is used for reducing the total prediction error; the total error of prediction is obtained by weighted summation of classified and detected errors, and the expression is as follows:
Figure BDA0002270843070000094
wherein L is the total error; l is a radical of an alcohol conf (x, c) is a classification error; l is loc (x, l, g) is detection error; the weight value a =1; x is a matching flag of the prediction box and the target box, x =1 represents matching, and x =0 represents mismatching; c is the score of the class prediction; l is the pixel value of the prediction frame; g is the pixel value of the target frame; if N =0, L =0;
wherein L is conf (x, c) is calculated by a Softmax loss function, and the calculation formula is as follows:
Figure BDA0002270843070000095
wherein, L is adopted loc (x, L, g) is obtained by performing position offset regression calculation by a Smooth L1 function, and the calculation formula is as follows:
Figure BDA0002270843070000096
Figure BDA0002270843070000101
Figure BDA0002270843070000102
Figure BDA0002270843070000103
Figure BDA0002270843070000104
in the formula (I), the compound is shown in the specification,
Figure BDA0002270843070000105
l represents a prediction box; g represents a target frame; d denotes a default box.
Step 4, screening the final prediction box set by adopting a non-maximum suppression algorithm to obtain a detection result;
the process of non-maxima suppression is as follows:
sorting all the prediction frames from top to bottom according to the scores, judging the Jaccard Over lap of the prediction frame with the highest score and other prediction frames, if the score is larger than the threshold value, discarding other prediction frames, then judging the Jaccard Over lap of the prediction frame with the highest score and other prediction frames, and discarding the prediction frames larger than the threshold value; then judging a prediction box with the third highest score; and so on until all prediction boxes are traversed.
1. Simulation experiment and result analysis
1. Experimental Environment and parameter configuration
The experimental process is carried out at the Linux PC side, and the specific hardware and software environments used for training the model and testing the algorithm performance are shown in Table 1:
table 1 experimental environment configuration
Table 1 experimental environment configuration
Figure BDA0002270843070000111
Setting the initial learning rate of the training model to be 0.001, changing the initial learning rate to be 0.0001 after the number of iterations reaches 4000, and finishing the model training after the iterations continue for 4000; model training was performed on a Tensorflow architecture, with tests performed every 20000 intervals, with momentum and weight decay rates set to 0.9000 and 0.0005, respectively.
2. Comparative method analysis
In order to verify the improved algorithm performance of the invention, an SSD algorithm, an FSSD algorithm and an ESSD algorithm are selected, five groups of tests are performed on the SDCK2019 managed cutter data set on the performance of each algorithm, and the test results are shown in Table 2:
TABLE 2
mAP of each algorithm on SDCK2019 data set
Table 2 mAP of each algorithm on SDCK2019 data set
Figure BDA0002270843070000121
It can be seen that compared with the classic SSD algorithm, the precision of the improved SSD detection algorithm is improved by 1.6% on the SDCK2019 managed cutter data, and the improved SSD detection algorithm is superior to the FSSD algorithm and the ESSD algorithm.
3. Detecting speed contrast
In order to test the detection speed of the improved SSD algorithm, the FSSD algorithm and the ESSD algorithm are selected, a GPU server carrying two NVIDIA RTX2080ti is tested on the SDCK2019 control tool data set, and the Frames pre second of each algorithm is calculated; the table of comparison of the detection speeds of the respective algorithms is shown in table 3:
TABLE 3
Detection speed of each algorithm on SDCK2019 data set
Table 5 detection speed of each algorithm on SDCK2019 data set
Figure BDA0002270843070000122
It can be seen that the detection precision of the improved SSD algorithm on the SDCK2019 control tool data set is 12.5Fps, which is equivalent to the detection speed of the ESSD algorithm, but lower than the detection speeds of the SSD and the FSSD algorithm; the reason is that the SSD basic network is replaced by ResNet101 from VGGNet, the network structure is deepened, and the algorithm detection speed is reduced due to complex calculation; however, the detection speed of the current security check machine for the controlled articles is required to be 10 frames per second, and the real-time requirement of the algorithm applied to the security check machine can be met as long as the detection speed of the algorithm is not lower than 10Fps, so that the improved SSD algorithm can meet the actual security check requirement.
4. In order to compare the overall detection performance of the improved SSD algorithm, FSSD algorithm and ESSD algorithm on the SDCK2019 managed cutter data set, a comparison graph of the detection speed and the detection precision of each algorithm is made by combining the tables 2 and 3, and the comparison graph is shown in FIG. 1; in the figure, the horizontal and vertical coordinates respectively represent the number of picture frames detected by the algorithm per second and the average detection precision (mAP value); comparing and analyzing the comprehensive detection performance of various algorithms on the small target of the cutting tool by using a speed precision comparison graph of the algorithms, wherein the closer to the upper right corner in the graph, the better the comprehensive detection performance of the algorithms is; the improved SSD algorithm has the highest detection precision on the small targets in the machining tool, solves the problems of partial false detection and missing detection of the small targets in the machining tool in the security inspection picture by the conventional target detection algorithm based on deep learning to a certain extent, and improves the overall detection effect.
5. To verify whether the improved SSD algorithm and feature fusion of the present invention is effective on the traditional SSD improved method, two steps of the algorithm of the present invention are tested in a distribution on SDCK2019 managed tool dataset, testing the maps of each step as shown in table 4:
TABLE 4
Improved algorithm step-by-step mAP test
Table 7 improved algorithm test mAP step by step
Figure BDA0002270843070000141
The improved SSD algorithm enhances the feature extraction capability of the detection model on the small target cutter; the deep layer and the shallow layer of the improved SSD algorithm are subjected to feature fusion, the receptive field of a shallow layer feature map is increased, and richer semantic information in the deep layer is fused to the shallow layer, so that the detection capability of the algorithm on the small cutter control target in the security inspection picture is comprehensively improved, the algorithm is high in precision and good in robustness; the results of the detection are shown in FIG. 2.

Claims (9)

1. A method for detecting and identifying a controlled cutter in security inspection is characterized by comprising the following steps:
step 1, carrying out normalization processing on an image A to obtain an image B;
step 2, processing the image B by an SSD-ResNet101 model, a feature fusion method, pooling and a Relu function to obtain a feature map C;
step 3, calculating a default frame of the characteristic diagram C by adopting a scale calculation formula; matching the default frame with the target frame by adopting a Jaccard overlay matching strategy and model training to obtain a prediction frame;
and 4, screening the final prediction box set by adopting a non-maximum suppression algorithm to obtain a detection result.
2. The method for safety inspection and identification of control tools according to claim 1, wherein the specific process of step 1 is as follows:
the image a is normalized using the following formula:
Figure FDA0002270843060000011
in the formula, y i Representing the result of the normalization processing; x is the number of i A matrix of images representing the input; mu is x i The mean value of (a); sigma 2 Is x i The variance of (a); ε is a positive number, typically 0.0001; gamma is a pull-up parameter; beta is an offset parameter.
3. The method for safety inspection and identification of controlled cutters as claimed in claim 2, wherein the step 2 comprises the following specific processes:
inputting image B into the SSD-ResNet101 network; carrying out convolution operation on the image B by adopting 512 3 multiplied by 3 convolution kernels on the C1 layer, and outputting a C1 layer feature map as the input of the C2 layer; performing convolution operation on the C2 layer by using a convolution kernel of 1 multiplied by 1024, and outputting a characteristic diagram of the C2 layer as the input of the C3 layer; performing convolution operation on the C3 layer by using 512 3 multiplied by 3 convolution kernels, and outputting a C3 layer feature map; introducing a C2 layer characteristic diagram into the C1 layer characteristic diagram, and enhancing by using a Relu function to obtain a C4 layer characteristic diagram; introducing a C3 layer characteristic diagram into the C2 layer characteristic diagram, and enhancing by using a Relu function to obtain a C5 layer characteristic diagram; fusing the C4 layer characteristic diagram and the C5 layer characteristic diagram, enhancing by using a Relu function, and outputting a C6 layer characteristic diagram as the input of a C7 layer;
adopting 512 convolution kernels with the size of 3 multiplied by 3 to carry out convolution operation on the C7 layer, and outputting a characteristic diagram of the C7 layer as the input of the C8 layer; performing convolution operation on the C8 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C8 layer feature map as an input of a C9 layer; performing convolution operation on the C9 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C9 layer feature map as an input of a C10 layer; performing convolution operation on the C10 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C10 layer feature map as an input of a C11 layer; performing convolution operation on the C11 layer by using 256 3 multiplied by 3 convolution kernels, and outputting a C10 layer feature map as an input of the C11 layer; and performing pooling operation on the C11 layer, and outputting a C12 layer feature map, namely the feature map C.
4. The method for safety inspection identifying a regulatory tool as in claim 3, wherein the Relu function is as follows:
f(x)=max(0,x) (2)
in the formula, x is an input vector from the neural network of the upper layer, and f (x) is an output vector.
5. The method for safety inspection and identification of control tools according to claim 3 or 4, wherein the specific process of step 3 is as follows:
selecting characteristic graphs of C6-C12 layers in the characteristic graph C, and calculating default frames of the characteristic graphs of the C6-C12 layers by adopting a scale calculation formula; and matching the default frame with the target frame by using a Jaccard overlay matching strategy to obtain a prediction frame.
6. The method for safety inspection and identification of a managed tool as claimed in claim 5, wherein the scale calculation formula is as follows:
Figure FDA0002270843060000031
in the formula, S min The scale of a default frame in the C1 layer is set as S min =0.04;S max Is the scale of a default frame in the topmost C12 layer and takes the value of S max =0.49;m=5;
Wherein, different aspect ratios are set for each default frame, and the calculation formula is as follows:
aspect ratio:
Figure FDA0002270843060000032
width:
Figure FDA0002270843060000033
high:
Figure FDA0002270843060000034
in the formula, when the aspect ratio is 1, a default box is added, and the scale is as follows:
Figure FDA0002270843060000035
7. the method for safety inspection and recognition of the managed cutter according to claim 5, wherein when the Jaccard overlay strategy is adopted to match the target frame with the default frame, when the Jaccard overlay is greater than the threshold value of 0.5, the default frame is considered to be matched with the target frame; the Jaccard Overlap expression is as follows:
Figure FDA0002270843060000036
in the formula, A represents a set area formed by all pixels in a default frame; b denotes an aggregate area made up of all pixels within the target frame.
8. The method for safety inspection recognition of a regulatory tool as claimed in claim 5, wherein the model training is used to reduce the predicted total error; the total error of prediction is obtained by weighted summation of classified and detected errors, and the expression is as follows:
Figure FDA0002270843060000037
wherein L is the total error; l is conf (x, c) is a classification error; l is loc (x, l, g) is detection error; the weight value a =1; x is a matching flag of the prediction box and the target box, x =1 represents matching, and x =0 represents mismatching; c is the score of the class prediction; l is the pixel value of the prediction frame; g is the pixel value of the target frame; if N =0, then L =0;
wherein L is conf (x, c) is calculated by a Softmax loss function, and the calculation formula is as follows:
Figure FDA0002270843060000041
wherein, L is adopted loc (x, L, g) is obtained by performing position offset regression calculation by a Smooth L1 function, and the calculation formula is as follows:
Figure FDA0002270843060000042
Figure FDA0002270843060000043
Figure FDA0002270843060000044
Figure FDA0002270843060000045
Figure FDA0002270843060000046
in the formula (I), the compound is shown in the specification,
Figure FDA0002270843060000047
l represents a prediction box; g represents a target frame; d denotes a default box.
9. The method for safety inspection and identification of a regulatory tool as claimed in claim 7 or 8 wherein the process of non-maximum suppression is as follows:
sorting all the prediction frames from top to bottom according to the scores, judging the Jaccard Over lap of the prediction frame with the highest score and other prediction frames, if the score is larger than the threshold value, discarding other prediction frames, then judging the Jaccard Over lap of the prediction frame with the highest score and other prediction frames, and discarding the prediction frames larger than the threshold value; then judging a prediction frame with the third highest score; and so on until all prediction boxes are traversed.
CN201911104406.3A 2019-11-13 2019-11-13 Method for detecting and identifying controlled cutter through security check Active CN110853019B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911104406.3A CN110853019B (en) 2019-11-13 2019-11-13 Method for detecting and identifying controlled cutter through security check

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911104406.3A CN110853019B (en) 2019-11-13 2019-11-13 Method for detecting and identifying controlled cutter through security check

Publications (2)

Publication Number Publication Date
CN110853019A CN110853019A (en) 2020-02-28
CN110853019B true CN110853019B (en) 2023-02-24

Family

ID=69601470

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911104406.3A Active CN110853019B (en) 2019-11-13 2019-11-13 Method for detecting and identifying controlled cutter through security check

Country Status (1)

Country Link
CN (1) CN110853019B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114549647B (en) * 2022-04-22 2022-08-12 成都飞机工业(集团)有限责任公司 Method for detecting placement orientation of HSK knife handle

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108446617B (en) * 2018-03-09 2022-04-22 华南理工大学 Side face interference resistant rapid human face detection method
CN108960198A (en) * 2018-07-28 2018-12-07 天津大学 A kind of road traffic sign detection and recognition methods based on residual error SSD model
CN109506628A (en) * 2018-11-29 2019-03-22 东北大学 Object distance measuring method under a kind of truck environment based on deep learning

Also Published As

Publication number Publication date
CN110853019A (en) 2020-02-28

Similar Documents

Publication Publication Date Title
CN109711474B (en) Aluminum product surface defect detection algorithm based on deep learning
CN108898610A (en) A kind of object contour extraction method based on mask-RCNN
CN110796186A (en) Dry and wet garbage identification and classification method based on improved YOLOv3 network
CN111898406B (en) Face detection method based on focus loss and multitask cascade
CN112200045B (en) Remote sensing image target detection model establishment method based on context enhancement and application
CN109272500B (en) Fabric classification method based on adaptive convolutional neural network
CN109671071B (en) Underground pipeline defect positioning and grade judging method based on deep learning
CN110992349A (en) Underground pipeline abnormity automatic positioning and identification method based on deep learning
CN112200225B (en) Steel rail damage B display image identification method based on deep convolution neural network
CN110008899B (en) Method for extracting and classifying candidate targets of visible light remote sensing image
CN114926407A (en) Steel surface defect detection system based on deep learning
CN111127454A (en) Method and system for generating industrial defect sample based on deep learning
CN110992314A (en) Pavement defect detection method and device and storage medium
CN110853019B (en) Method for detecting and identifying controlled cutter through security check
CN107871315B (en) Video image motion detection method and device
CN113657423A (en) Target detection method suitable for small-volume parts and stacked parts and application thereof
CN116029979A (en) Cloth flaw visual detection method based on improved Yolov4
CN107247967A (en) A kind of vehicle window annual test mark detection method based on R CNN
CN111860265B (en) Multi-detection-frame loss balanced road scene understanding algorithm based on sample loss
Lim et al. An optimized lightweight model for real-time wood defects detection based on yolov4-tiny
Lee A Study on Fruit Quality Identification Using YOLO V2 Algorithm
CN116542962A (en) Improved Yolov5m model-based photovoltaic cell defect detection method
CN111179239A (en) Tire X-ray flaw detection method for performing re-ranking by using background features
CN116630989A (en) Visual fault detection method and system for intelligent ammeter, electronic equipment and storage medium
CN111368625B (en) Pedestrian target detection method based on cascade optimization

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant