CN109214399B - Improved YOLOV3 target identification method embedded in SENET structure - Google Patents
Improved YOLOV3 target identification method embedded in SENET structure Download PDFInfo
- Publication number
- CN109214399B CN109214399B CN201811187750.9A CN201811187750A CN109214399B CN 109214399 B CN109214399 B CN 109214399B CN 201811187750 A CN201811187750 A CN 201811187750A CN 109214399 B CN109214399 B CN 109214399B
- Authority
- CN
- China
- Prior art keywords
- yolov3
- senet
- recognition method
- training
- algorithm
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
- G06F18/2321—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
- G06F18/23213—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention relates to the field of deep learning, in particular to an improved Yolov3 target recognition algorithm embedded in a SENET structure, which comprises the following steps of S100: collecting characteristic information of an object to be identified, and making a data set; the feature information includes image information; step S300: taking one part of the data set as a training set, and taking the other part of the data set as a test set; step S500: embedding an SE structure in a YOLOV3 algorithm to obtain an SE-YOLOV3 algorithm; step S600: training SE-YOLOV3 on a training set; step S700: SE-YOLOVE3 performance was tested on the test set. According to the improved YOLOV3 target identification algorithm embedded in the SENET structure, when more interference of defective parts exists in a sample picture, the method can still accurately identify the target parts, and higher precision ratio and recall ratio are obtained.
Description
Technical Field
The invention relates to the field of deep learning, in particular to an improved Yolov3 target recognition algorithm embedded in a SENET structure.
Background
Deep learning originates from the study of artificial neural networks, with the aim of simulating the human brain to capture and distinguish things. The method is characterized in that low-level features are combined to form an abstract high level, so that distributed features of data are discovered. The Convolutional Neural Network (CNN) is a deep learning method, and has excellent performance for image processing.
SENET is a convolutional neural network structure proposed by Hujie and its team in 2017, and obtains ILSVRC 2017 classification task champion. The structure is composed of three parts of Squeeze, interaction and weight, and the interdependence relation among characteristic channels is explicitly constructed. Since it does not change the size of the feature map, there is no need to introduce new functions, SE has the great advantage of being embedded in almost all network architectures. Experimental results show that by embedding the SE structure in classical networks such as ResNet and inclusion, the generalization capability and accuracy of the model can be remarkably improved.
In the past, image recognition is mainly realized by means of a traditional image processing technology, and features need to be manually extracted and input into a classifier for recognition. With the continuous development of deep learning, various target detection algorithms are proposed to gradually replace the traditional detection method. At present, the image recognition by using the convolutional neural network mainly includes two types: one is a region-based target identification method, such as R-CNN, Faster R-CNN, Mask R-CNN, etc., which has high target positioning accuracy but has the disadvantage of low detection speed; the other is a regression-based target recognition method, such as a method of YOLO.
The YoLO algorithm was proposed by Joseph Redmon et al in 2015, collectively referred to as You Only Look one. The Yolov3 proposed in 2018 is an integrated product of a Yolo series algorithm, has the characteristics of high speed and high accuracy, and can meet the requirement of industrial instantaneity.
At present, convolutional neural networks are often utilized to solve the problem of image recognition. As shown in fig. 6, there are 4 types of target parts to be sorted, and the postures of the parts are arbitrary but do not overlap. For such a problem, the conventional image processing algorithm relying on manual feature extraction is difficult to solve, and can be processed by using YOLOV3 with better performance at present. Due to the fact that positions of parts are disordered, a large number of defective parts (edge positions in fig. 6) exist in a shot picture, when the YOLOV3 network training is used, the defective parts cannot be identified correctly, a background area is judged as the parts by mistake, and the precision ratio on a training set is low (only 72.3%), so that the YOLOV3 algorithm needs to be improved to solve the problems.
Disclosure of Invention
In order to solve the problems in the background art, the invention provides an improved YOLOV3 target identification algorithm embedded in a SENET structure, and when more interference of defective parts exists in a sample picture, the method can still accurately identify the target parts to obtain higher precision ratio and recall ratio.
The technical scheme for solving the problems is as follows: an improved Yolov3 target recognition algorithm embedded in SENET structure, which is characterized by comprising the following steps:
step S100: collecting characteristic information of an object to be identified, and making a data set; the feature information includes image information;
step S300: taking one part of the data set as a training set, and taking the other part of the data set as a test set;
step S500: embedding an SE structure in a YOLOV3 algorithm to obtain an SE-YOLOV3 algorithm;
step S600: training SE-YOLOV3 on a training set;
step S700: SE-YOLOVE3 performance was tested on the test set.
Further, the method also comprises the step S200: carrying out data processing and labeling on the collected characteristic parameters of the object to be identified; wherein the step S200 is between the step S100 and the step S300.
Further, in the step S200, the data processing includes enlarging and/or reducing and/or enhancing and/or reducing and/or flipping and/or adding noise to the acquired feature parameters of the object to be recognized.
Further, in the step S200, the labels are specifically: and marking the acquired characteristic parameters of the object to be identified by using an image marking tool.
Further, the method further comprises the step S400: recalculating an anchor for the data set obtained in step S200, specifically: reading the marked data set, randomly taking out W, H values as initialized seed points, performing iteration by using a K-means algorithm, and calculating to obtain an anchor; wherein the step S400 is between the step S200 and the step S300.
Further, in the step S100, the collecting of the feature information of the object to be recognized is to collect the feature information of the object to be recognized by using a CMOS camera.
Further, in step S500, an SE structure is embedded in the YOLOV3 algorithm to obtain an SE-YOLOV3 algorithm, which specifically includes: embedding an SE structure behind each short cut; modifying the configuration file cfg of the YOLOV3, adding an SE substructure after the 4 th, 8 th, 11 th, 15 th, 18 th, 21 th, 24 th, 27 th, 30 th, 33 th, 36 th, 40 th, 43 th, 46 th, 49 th, 52 th, 55 th, 58 th, 61 th, 65 th, 68 th, 71 th and 74 th shortcuts, and designating the channel value of the global average pooling as the number of feature map channels output by the shortcuts for calculating the global average pooling;
specifically, the channel values are 64, 128, 128, 256, 256, 256, 256, 256, 256, 256, 256, 256, 512, 512, 512, 512, 512, 512, 512, 512, 512, 1024, 1024, 1024 in sequence.
Further, in step S600, SE-YOLOV3 is trained on the training set, and a weight file is obtained after the training is finished.
Further, in step S700, the performance of SE-yoloeve 3 is tested on the test set, specifically: and (5) loading the weight data obtained in the step (S600) and carrying out performance test on the SE-YOLOV3 on the test set.
The invention has the advantages that:
according to the SENET structure embedded improved YOLOV3 target recognition algorithm, the output feature map weight has a global receptive field, information of each channel can be integrated, the expression capability of the feature map can be enhanced, the detection effect of a model in an edge area is improved, when the edge of a sample picture has more incomplete part interference, the target can still be accurately recognized, the false recognition of the edge area is reduced, the number of false positive cases is reduced, and therefore a higher precision ratio is obtained.
Drawings
FIG. 1 is a schematic flow diagram of an embodiment of the present invention;
FIG. 2 is a diagram of a YOLOV3 network architecture;
FIG. 3 is a diagram of the SE-YOLOV3 network architecture of the present invention;
FIG. 4 is a YOLOV3 training data curve;
FIG. 5 is a plot of the SE-YOLOV3 training data of the present invention;
FIG. 6 is a graph of the results of the YOLOV3 assay;
FIG. 7 is a graph showing the results of the test of SE-YOLOV3 according to the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention more apparent, the technical solutions of the embodiments of the present invention will be described clearly and completely with reference to the accompanying drawings of the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all embodiments of the present invention. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, are within the scope of the present invention. Thus, the following detailed description of the embodiments of the present invention, presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, are within the scope of the present invention.
Referring to fig. 1, the object of the present invention is to provide an improved YOLOV3 target recognition algorithm embedded in a SENet structure, comprising the following steps:
step S100: collecting characteristic information of an object to be identified, and making a data set; the feature information includes image information.
In this embodiment, specifically, an annular red light source is fixed below the CMOS camera, and 4 types of small fastener pictures are collected to obtain an RGB three-channel picture as a convolutional neural network data set.
This example uses a Basler company acA2500-20gc camera to take pictures of class 4 mini fasteners (size 4-7 mm) with a ring red light source fixed under the camera. 529 pictures are obtained, the resolution is 2592 multiplied by 2048 pixels (RGB three channels), and the picture is converted into a single channel by a compiling program so as to improve the calculation speed.
Step S200: and carrying out data processing and labeling on the acquired characteristic parameters of the object to be identified. The data processing comprises the steps of amplifying and/or reducing and/or enhancing the brightness and/or weakening the brightness and/or turning over and/or increasing noise of the acquired characteristic parameters of the object to be identified; the labels are specifically: and marking the acquired characteristic parameters of the object to be identified by using an image marking tool.
The method specifically comprises the following steps: and performing data enhancement on the acquired original picture, and labeling 4 types of parts in the picture by using an image labeling tool Labelimg.
The present embodiment fabricates the acquired image into a data set usable by a convolutional neural network. Writing a program, and performing data enhancement on an original picture by 6 methods of zooming (zooming the image width and the image height to 1.5 times), zooming (zooming the image width to 1/3 and the image height to 1/2 to ensure that the image size is a multiple of 32), brightness enhancement, brightness reduction, overturning (90 degrees and 180 degrees) and noise addition (specklenoid), thereby obtaining 4232 pictures. And (3) marking the 4 types of parts in the 529 pictures by using an image marking tool Labelimg. The specific method comprises the following steps: and (4) selecting all complete parts in the picture frame (ignoring the parts with defective edges of the picture) to obtain the VOC format xml file. Meanwhile, a program needs to be written, and the label file of 3703 pictures subjected to data enhancement is calculated according to the label file of 529 original pictures. The format conversion program is written to convert the xml file into a txt file in "tag" + "X" + "Y" + "W" + "H" format.
Step S300: and taking one part of the data set as a training set, and taking the rest part of the data set as a testing set.
The embodiment specifically includes: 3400 out of the 4232 pictures obtained are randomly selected as a training set, wherein 40612 parts are contained, and the rest 832 parts are used as a test set. Including 9963 parts.
And 4, step 4: and building a YOLOV3 model, and recalculating partial parameters.
In this embodiment, a development platform is first built, a CPU selects intel (r) core (tm) i7-8700, a GPU selects NVIDIA GeForce GTX 1080Ti, an operating system selects Ubuntu 16.04LTS, and a deep learning framework is Pytorch. Then build a YOLOV3 model based on GPU acceleration, as shown in fig. 2. In order to adapt to the data set, part of the parameters need to be modified, and the specific contents include: setting the number of channels in a YOLOV3 algorithm cfg configuration file as 1, and reading in a single-channel gray-scale picture; the number of categories in the cfg configuration file is 4, and the number of channels of the three-scale feature maps (13 × 13, 26 × 26, 52 × 52) for detection is set to be (4+1+4) × 3, namely 27; the label that sets up 4 kinds of parts simultaneously does: type1, type2, type3, type 4; the anchors in the YOLOV3 source code were designed for the coco dataset, and for the datasets herein, the K-means + + algorithm was used to save computation time. The specific method comprises the following steps: w, H data in 4232 txt files which are marked are read, 9 groups of W, H values are randomly taken out to serve as initial seed points, the number of iterations is set to 10000, and then the K-means algorithm is used for iteration. The final calculation yields an anchorbox suitable for the data set herein: (39, 56),(47, 58),(51, 66),(42, 82),(63, 57),(69, 53),(49, 81),(65, 67),(61, 80).
And 5: and embedding an SE structure in the YOLOV3 algorithm to obtain an improved SE-YOLOV3 algorithm.
This example is an improvement over the YOLOV3 model established in step 3. In fig. 3, CBR1 represents a convolution layer with a convolution kernel size of 1 × 1, and the batch normalization operation was performed using ReLU as the activation function. CBR3 is similar to CBR1 except that the convolution kernel size is 3 x 3. The shortcut layer is composed of CBR1 and CBR3, and shortcut n in the figure represents that n (CBR1+ CBR3) substructures are used. The SE-YOLOV3 algorithm is shown in fig. 3, the SE structure performs global average pooling on the input feature map to obtain a feature map (C is the number of feature map channels) with a size of C × 1 × 1, the feature map is activated by a sigmoid function after passing through two full-connected layers (dimensionality reduction and dimensionality increase), a weight with a size of C × 1 × 1 is obtained, and the weight is multiplied by the original input feature map at a corresponding position to obtain an output. The SE-shortcutn layer is composed of a conventional shortcutn layer and an SE structure. After each shortcut, a SE structure is embedded, and SE-shortcut represents that n (CBR1+ CBR3+ SE) substructures are used. Modifying the configuration file cfg of the YOLOV3, adding an SE substructure after the 4 th, 8 th, 11 th, 15 th, 18 th, 21 th, 24 th, 27 th, 30 th, 33 th, 36 th, 40 th, 43 th, 46 th, 49 th, 52 th, 55 th, 58 th, 61 th, 65 th, 68 th, 71 th and 74 th layers (all shortcut layers), and designating the global average pooled channel value as the number of feature map channels output by the shortcut layer for calculating the global average pooling. Specifically, the channel values are 64, 128, 128, 256, 256, 256, 256, 256, 256, 256, 256, 256, 512, 512, 512, 512, 512, 512, 512, 512, 512, 1024, 1024, 1024 in sequence. Taking a 4 th layer short cut layer of YOLOV3 as an example, an SE structure input feature map is 208 × 208 × 64, a feature map of 1 × 1 × 64 is obtained after global average pooling, the feature map is changed into 1 × 1 × 4 after passing through a first full connection layer, the feature map is changed into 1 × 1 × 64 after passing through a second full connection layer, finally a weight value of 1 × 1 × 64 is obtained through a sigmoid activation function, and the input feature map and the weight value are multiplied to obtain an output of 208 × 208 × 64. The network depth of the YOLOV3 is 106 layers, and the network depth of the improved SE-YOLOV3 reaches 129 layers.
Step 6: and respectively training a Yolov3 model and a SE-Yolov3 model on a training set to obtain weight data.
This embodiment is used to train the model. And (3) training the models of YOLOV3 and SE-YOLOV3 built in the steps 3 and 4 on a training set for 150 times from the beginning, wherein the specific parameters are set as follows: the batch number was 12, the gradient optimization used the Adam algorithm, the initial learning rate was 0.0001, the weight attenuation was set to 0.0005, and the learning rate was attenuated to 1/10 after each 50 iterations. The method for judging the true example TP comprises the following steps: in the grid unit where the part exists, the confidence coefficient is greater than 0.99, the intersection ratio (IoU) of the prediction frame (bbox) and the real frame is greater than 0.5, the prediction category is correct, and 3 conditions are met, so that the method is a true example. The false positive case FP contains two cases, the first case: in the grid unit where the part exists, if the confidence coefficient is greater than 0.99 but the part is not a true case, the part is judged to be a false positive case; in the second case: if the confidence is greater than 0.99 in the grid cell where the part does not exist, it is determined as a false positive. The false negative FN determination method comprises the following steps: in the grid cells where the part exists, the confidence is less than 0.99. The precision ratio P is TP/(TP + FP), and the recall ratio R is TP/(TP + FN). The data after each iteration is saved and plotted as a curve, as shown in fig. 4 and 5. And storing the weight data after the training is finished.
And 7: the performance of the model YOLOV3 and SE-YOLOV3 was tested on a test set.
This embodiment is for a test model. Loading two weight data obtained in the step 5 in a model YOLOV3 and a model SE-YOLOV3, and respectively verifying on a test set, wherein the specific parameters are set as follows: the number of batches was 32, the IoU threshold was set to 0.5, the confidence threshold was set to 0.98, and the non-maximum suppression threshold was set to 0.4. Finally, the precision ratio and the recall ratio of the test are obtained, and the test picture is saved, as shown in fig. 6 and 7.
The invention has the following beneficial effects: as can be seen from fig. 4 and 5, on the training set, the precision ratio of YOLOV3 was 72.3%, and the recall ratio was 98.5%, while that of SE-YOLOV3 was 92.5%, and the recall ratio was 96.1%. On the test set, the precision of YOLOV3 was 71.6%, the recall was 97.7%, and the precision of SE-YOLOV3 was 91.9%, the recall was 95.7%. As can be seen from fig. 6 and 7, YOLOV3 has a situation where the edge defect part is identified incorrectly, and at the same time, the background is identified incorrectly as the part, SE-YOLOV3 only identifies the complete part, and there is no situation where the background is identified as the part. By adding the SE structure on the basis of the YOLOV3, the invention integrates the advantages of the YOLOV3 and the SE model, enhances the anti-interference performance of the YOLOV3 algorithm on incomplete parts in images, reduces the number of false positive cases for identification, and improves the precision ratio by 20.3 percent under the condition of ensuring higher recall ratio.
The above description is only an embodiment of the present invention, and not intended to limit the scope of the present invention, and all equivalent structures or equivalent flow transformations that are made by using the contents of the present specification and the drawings, or directly or indirectly applied to other related system fields, are included in the scope of the present invention.
Claims (8)
1. An improved Yolov3 target recognition method embedded in a SENET structure is characterized by comprising the following steps:
step S100: collecting characteristic information of an object to be identified, and making a data set; the feature information includes image information;
step S300: taking one part of the data set as a training set, and taking the other part of the data set as a test set;
step S500: embedding an SE structure in a YOLOV3 algorithm to obtain an SE-YOLOV3 algorithm, which specifically comprises the following steps: embedding an SE structure behind each short cut; modifying the configuration file cfg of the YOLOV3, adding an SE substructure after the 4 th, 8 th, 11 th, 15 th, 18 th, 21 th, 24 th, 27 th, 30 th, 33 th, 36 th, 40 th, 43 th, 46 th, 49 th, 52 th, 55 th, 58 th, 61 th, 65 th, 68 th, 71 th and 74 th shortcuts, and designating the channel value of the global average pooling as the number of feature map channels output by the shortcuts for calculating the global average pooling;
specifically, the channel values are 64, 128, 128, 256, 256, 256, 256, 256, 512, 512, 512, 512, 512, 512, 512, 512, 512, 1024, 1024, 1024 in sequence;
step S600: training SE-YOLOV3 on a training set;
step S700: SE-YOLOVE3 performance was tested on the test set.
2. The SENET structure embedded improved YOLOV3 target recognition method according to claim 1, wherein: further comprising step S200: carrying out data processing and labeling on the collected characteristic parameters of the object to be identified; wherein the step S200 is between the step S100 and the step S300.
3. The SENET structure embedded improved YOLOV3 target recognition method according to claim 2, wherein: in step S200, the data processing includes amplifying and/or reducing and/or enhancing brightness and/or reducing brightness and/or flipping and/or adding noise to the acquired feature parameters of the object to be identified.
4. The SENET structure embedded improved YOLOV3 target recognition method according to claim 3, wherein: in step S200, the labeling specifically includes: and marking the acquired characteristic parameters of the object to be identified by using an image marking tool.
5. The SENET structure embedded improved YOLOV3 target recognition method according to claim 4, wherein: further comprising step S400: recalculating an anchor for the data set obtained in step S200, specifically: reading the marked data set, randomly taking out W, H values as initialized seed points, performing iteration by using a K-means algorithm, and calculating to obtain an anchor; wherein the step S400 is between the step S200 and the step S300.
6. The SENET structure embedded improved YOLOV3 object recognition method according to any one of claims 1 to 5, wherein: in step S100, the collecting the feature information of the object to be recognized is to collect the feature information of the object to be recognized by using a CMOS camera.
7. The SENET structure embedded improved YOLOV3 object recognition method according to any one of claims 1 to 5, wherein: in step S600, SE-YOLOV3 is trained on a training set, and a weight file is obtained after training is completed.
8. The SENET structure embedded improved YOLOV3 target recognition method according to claim 7, wherein: in step S700, the performance of SE-yoloeve 3 is tested on the test set, specifically: and (5) loading the weight data obtained in the step (S600) and carrying out performance test on the SE-YOLOV3 on the test set.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811187750.9A CN109214399B (en) | 2018-10-12 | 2018-10-12 | Improved YOLOV3 target identification method embedded in SENET structure |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811187750.9A CN109214399B (en) | 2018-10-12 | 2018-10-12 | Improved YOLOV3 target identification method embedded in SENET structure |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109214399A CN109214399A (en) | 2019-01-15 |
CN109214399B true CN109214399B (en) | 2021-01-01 |
Family
ID=64979487
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811187750.9A Active CN109214399B (en) | 2018-10-12 | 2018-10-12 | Improved YOLOV3 target identification method embedded in SENET structure |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109214399B (en) |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109886925A (en) * | 2019-01-19 | 2019-06-14 | 天津大学 | A kind of aluminium material surface defect inspection method that Active Learning is combined with deep learning |
CN109978868A (en) * | 2019-03-29 | 2019-07-05 | 北京百度网讯科技有限公司 | Toy appearance quality determining method and its relevant device |
CN110011727B (en) * | 2019-04-09 | 2020-03-20 | 浩鲸云计算科技股份有限公司 | ODF equipment port-oriented detection system |
CN111832336B (en) * | 2019-04-16 | 2022-09-02 | 四川大学 | Improved C3D video behavior detection method |
CN110084166B (en) * | 2019-04-19 | 2020-04-10 | 山东大学 | Intelligent substation smoke and fire identification and monitoring method based on deep learning |
CN110210571B (en) * | 2019-06-10 | 2023-01-06 | 腾讯医疗健康(深圳)有限公司 | Image recognition method and device, computer equipment and computer readable storage medium |
CN110929577A (en) * | 2019-10-23 | 2020-03-27 | 桂林电子科技大学 | Improved target identification method based on YOLOv3 lightweight framework |
CN110781888B (en) * | 2019-10-25 | 2022-07-12 | 北京字节跳动网络技术有限公司 | Method and device for returning to screen in video picture, readable medium and electronic equipment |
CN110909666B (en) * | 2019-11-20 | 2022-10-25 | 西安交通大学 | Night vehicle detection method based on improved YOLOv3 convolutional neural network |
CN111428550A (en) * | 2019-11-29 | 2020-07-17 | 长沙理工大学 | Vehicle detection method based on improved YO L Ov3 |
CN111382678B (en) * | 2020-02-25 | 2023-04-18 | 浙江大学 | Tourist bus passenger flow statistical algorithm based on improved CNN network |
CN111429418A (en) * | 2020-03-19 | 2020-07-17 | 天津理工大学 | Industrial part detection method based on YO L O v3 neural network |
CN111476131B (en) * | 2020-03-30 | 2021-06-11 | 北京微播易科技股份有限公司 | Video processing method and device |
CN111814862A (en) * | 2020-06-30 | 2020-10-23 | 平安国际智慧城市科技股份有限公司 | Fruit and vegetable identification method and device |
CN112070713A (en) * | 2020-07-03 | 2020-12-11 | 中山大学 | Multi-scale target detection method introducing attention mechanism |
CN111882554B (en) * | 2020-08-06 | 2022-05-03 | 桂林电子科技大学 | SK-YOLOv 3-based intelligent power line fault detection method |
TWI768432B (en) * | 2020-08-18 | 2022-06-21 | 新加坡商鴻運科股份有限公司 | Method,system,electronic device and storage medium for classifying and processing parts before assembly |
CN112163602A (en) * | 2020-09-14 | 2021-01-01 | 湖北工业大学 | Target detection method based on deep neural network |
CN112561872B (en) * | 2020-12-09 | 2022-05-24 | 福州大学 | Corrosion defect segmentation method for tower crane |
CN112749734B (en) * | 2020-12-29 | 2024-01-05 | 北京环境特性研究所 | Domain-adaptive target detection method based on movable attention mechanism |
CN112884090A (en) * | 2021-04-14 | 2021-06-01 | 安徽理工大学 | Fire detection and identification method based on improved YOLOv3 |
CN113553936A (en) * | 2021-07-19 | 2021-10-26 | 河北工程大学 | Mask wearing detection method based on improved YOLOv3 |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107358596B (en) * | 2017-04-11 | 2020-09-18 | 阿里巴巴集团控股有限公司 | Vehicle loss assessment method and device based on image, electronic equipment and system |
CN108154120A (en) * | 2017-12-25 | 2018-06-12 | 上海七牛信息技术有限公司 | video classification model training method, device, storage medium and electronic equipment |
CN108304820B (en) * | 2018-02-12 | 2020-10-13 | 腾讯科技(深圳)有限公司 | Face detection method and device and terminal equipment |
CN108573303A (en) * | 2018-04-25 | 2018-09-25 | 北京航空航天大学 | It is a kind of that recovery policy is improved based on the complex network local failure for improving intensified learning certainly |
CN108510012B (en) * | 2018-05-04 | 2022-04-01 | 四川大学 | Target rapid detection method based on multi-scale feature map |
-
2018
- 2018-10-12 CN CN201811187750.9A patent/CN109214399B/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN109214399A (en) | 2019-01-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109214399B (en) | Improved YOLOV3 target identification method embedded in SENET structure | |
CN111310862B (en) | Image enhancement-based deep neural network license plate positioning method in complex environment | |
CN111444821B (en) | Automatic identification method for urban road signs | |
CN108564097B (en) | Multi-scale target detection method based on deep convolutional neural network | |
CN110443143B (en) | Multi-branch convolutional neural network fused remote sensing image scene classification method | |
CN110348319B (en) | Face anti-counterfeiting method based on face depth information and edge image fusion | |
CN110532920B (en) | Face recognition method for small-quantity data set based on FaceNet method | |
CN104361313B (en) | A kind of gesture identification method merged based on Multiple Kernel Learning heterogeneous characteristic | |
CN103984948B (en) | A kind of soft double-deck age estimation method based on facial image fusion feature | |
CN105574550A (en) | Vehicle identification method and device | |
CN113420607A (en) | Multi-scale target detection and identification method for unmanned aerial vehicle | |
CN113919442B (en) | Tobacco maturity state identification method based on convolutional neural network | |
CN114998220B (en) | Tongue image detection and positioning method based on improved Tiny-YOLO v4 natural environment | |
CN111178451A (en) | License plate detection method based on YOLOv3 network | |
CN113487610B (en) | Herpes image recognition method and device, computer equipment and storage medium | |
CN111340096A (en) | Weakly supervised butterfly target detection method based on confrontation complementary learning | |
CN110543906A (en) | Skin type automatic identification method based on data enhancement and Mask R-CNN model | |
WO2020119624A1 (en) | Class-sensitive edge detection method based on deep learning | |
CN116524189A (en) | High-resolution remote sensing image semantic segmentation method based on coding and decoding indexing edge characterization | |
CN113177528B (en) | License plate recognition method and system based on multi-task learning strategy training network model | |
CN103268494B (en) | Parasite egg recognition methods based on rarefaction representation | |
CN113378642B (en) | Method for detecting illegal occupation buildings in rural areas | |
CN117036948A (en) | Sensitized plant identification method based on attention mechanism | |
CN117011614A (en) | Wild ginseng reed body detection and quality grade classification method and system based on deep learning | |
CN116189130A (en) | Lane line segmentation method and device based on image annotation model |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |