CN116343150A - Road sign target detection method based on improved YOLOv7 - Google Patents

Road sign target detection method based on improved YOLOv7 Download PDF

Info

Publication number
CN116343150A
CN116343150A CN202310293143.5A CN202310293143A CN116343150A CN 116343150 A CN116343150 A CN 116343150A CN 202310293143 A CN202310293143 A CN 202310293143A CN 116343150 A CN116343150 A CN 116343150A
Authority
CN
China
Prior art keywords
yolov7
improved
detection
layer
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310293143.5A
Other languages
Chinese (zh)
Inventor
邓月明
谢竞
陈正浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hunan Normal University
Original Assignee
Hunan Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hunan Normal University filed Critical Hunan Normal University
Priority to CN202310293143.5A priority Critical patent/CN116343150A/en
Publication of CN116343150A publication Critical patent/CN116343150A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/588Recognition of the road, e.g. of lane markings; Recognition of the vehicle driving pattern in relation to the road
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/776Validation; Performance evaluation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Multimedia (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a target detection method based on road identification recognition of improved YOLOv7, which comprises the steps of obtaining a public road identification data set CCTSDB-2021; dividing the data set into a test set and a training set; carrying out data enhancement on the image; constructing an improved YOLOv7-SDB network model, and importing a training set into a detection model for training; and carrying out target detection on the test set by the trained detection model. According to the invention, the Conv convolution of 6 conv blocks in the ELAN block in the backup is subjected to the concat splicing operation with the CBS block, so that the number of channels is increased under the condition that each piece of characteristic information is not increased; introducing a GAM attention module after each concat operation in the Head to improve the detection precision; in the space pyramid pooling module, SPPCSPC is replaced by SPPFCSPC by referring to the thought of SPPF, so that the speed is improved and the calculated amount is reduced under the condition of keeping the receptive field unchanged; the original Mosaic data enhancement is changed into Mosaic-9 data enhancement, so that the data set samples are enriched. Compared with the original YOLOv7 model, the method has the advantages that the detection precision and the detection speed are obviously improved.

Description

Road sign target detection method based on improved YOLOv7
Technical Field
The invention relates to the field of computer vision target detection, in particular to a target detection method based on improved YOLOv7 road identification.
Background
Road traffic sign detection refers to detecting and identifying various traffic signs on roads, such as speed limit signs, forbidden signs, indicating signs, warning signs, and the like. Object detection based on road traffic identification is an important task in computer vision, aiming at identifying and locating various traffic identifications on roads. The technology has important significance in the fields of automatic driving, intelligent traffic systems and the like.
At present, algorithms for road traffic sign target detection are mainly divided into two categories: based on traditional vision methods and based on deep learning methods. The road traffic sign target detection algorithm based on the traditional vision method is mainly based on the characteristics of template matching, edge detection, shape matching and the like, such as a Canny algorithm based on edge detection, a color segmentation algorithm based on image color distribution and the like. And the deep learning-based mode extracts information from the image through convolutional neural network learning, and utilizes various optimization strategies to identify and position the target. The algorithms currently in widespread use include R-CNN, faster R-CNN, YOLO, SSD, and the like.
YOLO series is widely used in industry for its good detection performance and detection speed, where YOLO v7 brings new characteristics compared to previous generation versions, including multi-scale feature fusion (MSFF) layer, smooth tag (SL) algorithm, class Balance (CB) loss function, etc. Although the YOLOv7 algorithm has certain optimization in the design of the multi-scale fusion and depth deformable convolution network, the detection effect of the algorithm can be limited for some small-size and slender targets, and the problems of missed detection and the like are easy to generate.
Disclosure of Invention
The invention provides a road sign target detection method based on an improved YOLOv7 model based on an original YOLOv7 model, and aims to improve the detection precision of medium and small-sized road signs under the condition of properly improving the parameter quantity.
The invention is realized by the following technical scheme, and the road sign target detection method based on the improved YOLOv7 provided by the invention comprises the following steps:
1. acquiring a public road identification data set CCTSDB-2021, wherein the data set comprises 20491 road identification pictures and 3 categories which are respectively indication identifications, warning identifications and forbidden identifications;
2. 7500 pictures in the data set are taken, and the test set and the training set are divided according to a ratio of 8:2;
3. performing Mosaic-9 data enhancement processing on the selected picture;
4. k-means cluster analysis is carried out on the real label samples in the data set;
5. constructing a road identification detection model YOLOv7-SDB based on the improved YOLOv 7;
6. leading the training set into a detection model for training to obtain a road identification detection model;
7. carrying out road identification detection on the test set by the detection model;
the invention provides a preferable scheme, wherein the Mosaic-9 data enhancement process comprises the following steps: randomly selecting 9 pictures, performing augmentation operation on the 9 pictures respectively, and pasting the 9 pictures to the corresponding positions of the masks with the same size as the final output image respectively; augmentation operations include random clipping, scaling, alignment, and gamut variation; and obtaining a new picture and enriching the data set.
According to the preferred scheme provided by the invention, the K-means clustering algorithm clusters the width and height of all target frames contained in the training set to obtain 9 most representative width and height combinations, and the 9 anchors are divided into 3 groups according to the width and height.
According to the preferred scheme provided by the invention, the construction of the improved YOLOv7-SDB structure improvement process based on the improved YOLOv7 road sign detection model comprises the following steps: firstly, performing concha splicing operation on 6 conv convolutions in an ELAN block in a back band part in a detection network and a CBS block, then respectively introducing GAM (Global Attention Mechanism) attention modules after each concha operation in a Head part, and finally replacing an SPPCSPC module in the Head part with an SPPFCSPC module by referring to the idea of SPPF. For the final output 32 times downsampled feature map C5 of the Backbone, then the number of channels goes from 1024 to 512 through the SPPFCSPC module. Fusing the P3 layer, the P4 layer and the P5 layer with the C4 layer and the C3 layer in a top-down mode; and then fusing with the P4 layer and the P5 layer in a bottom-up mode. Finally, three detection layers with different sizes of 80×80, 40×40 and 20×20 are obtained, wherein 80×80 is used for detecting small targets, 40×40 is used for detecting small targets and 20×20 is used for detecting medium targets.
In a preferred embodiment of the present invention, the GAM attention module includes: the GAM attention mechanism can stably improve the performance of model detection by designing a mechanism for reducing feature information generation and amplifying global dimension interaction features. Defining an input feature map as
Figure SMS_1
Intermediate state variable ∈ ->
Figure SMS_2
And final output +.>
Figure SMS_3
The expression of (2) is shown in formula 1:
Figure SMS_4
Figure SMS_5
(1)
wherein,,
Figure SMS_6
and->
Figure SMS_7
Channel and spatial attention patterns, respectively, +.>
Figure SMS_8
Representing a calculated multiplication of the element pattern.
Input features
Figure SMS_9
Finally pass through the channelThe sub-attention unit and the space sub-attention unit output a value of +>
Figure SMS_10
Is used to preserve cross-three dimensional information using 3D placement by activating channel attention subunits and multi-layer MLP (multi-layer perceptron) is used to amplify cross-dimensional channel spatial correlation. In the spatial attention subunit, two convolutions are used to merge spatial information in order to fully focus on the spatial information.
The SPPFCSPC module of a preferred scheme provided by the invention comprises: the SPPFCSPC module is based on the SPPCSPC module according to the idea of SPPF, with the difference that the SPPFCSPC module processes the 3 largest pooling layers sequentially, and can achieve a speed boost with unchanged receptive field.
According to the preferred scheme provided by the invention, the super-parameter setting of the road sign detection model based on the improved YOLOv7 is as follows:
the Learning rate (Learning rate) was set to 0.01, the optimizer was chosen to be random gradient descent (SGD), the Momentum (Momentum) was set to 0.937, the number of training iterations (Batch) was 300 rounds, the Batch Size (batch_size) was 16, and the sio_loss was used as the boundary loss function.
The invention discloses a road identification target detection method based on improved YOLOv7, which is based on an original YOLOv7 target detection network model, combines a road identification data set CCTSDB-2021 to carry out optimization and improvement, can carry out real-time detection on medium and small-sized road identifications, and improves the detection precision and detection speed of the model.
Drawings
Fig. 1 is a flowchart of a road sign target detection method based on improved YOLOv 7.
Fig. 2 is a schematic diagram of a detection network structure of a road sign target detection method based on improved YOLOv 7.
Fig. 3 is a schematic diagram of the structure of the GAM attention module.
Fig. 4 is a training effect diagram of the YOLOv7-SDB road identification detection network.
Fig. 5 is a diagram showing the detection effect of the YOLOv7-SDB road identification detection network.
FIG. 6 is a schematic diagram of a confusion matrix for a Yolov7-SDB road identification detection network.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, a detailed description of the present invention will be further described with reference to the accompanying drawings.
Fig. 1 is a schematic flow chart of the method of the present invention, and the method for detecting a road sign target based on improved YOLOv7 provided by the present invention specifically includes the following steps:
step 1: and acquiring a public road identification image, and dividing a target detection data set required by training and testing.
The operation in the step 1 is as follows: 7500 pictures in the CCTSDB-2021 data set are taken, and the training set and the testing set are divided according to the proportion of 8:2. And respectively creating summary files train. Txt and val. Txt for storing the absolute paths and the label positions and the categories of the pictures in rows by writing a python script file, and finally placing label labels and jpg pictures which are separated in a training set and a testing set under the same catalog.
Step 2: and performing Mosaic-9 data enhancement processing on the selected picture.
The operation in the step 2 is as follows: randomly selecting 9 pictures, performing augmentation operation on the 9 pictures respectively, and pasting the 9 pictures to the corresponding positions of the masks with the same size as the final output image respectively; augmentation operations include random clipping, scaling, alignment, and gamut variation; and obtaining a new picture and enriching the data set.
Randomly cutting: the obstacle is caused to appear in different positions of the original picture in different proportions by image random cropping.
Scaling: the original picture is scaled in size.
Arrangement: the combination of the original pictures and the arrangement of the frames.
Color gamut variation: the brightness, saturation and tone of the original picture are changed.
Step 3: and carrying out K-means cluster analysis on the real label sample in the data set.
The operation in the step 3 is as follows: the K-means clustering algorithm clusters the width and height of all target frames contained in the training set to obtain 9 most representative width and height combinations, and the 9 Anchor are divided into 3 groups according to the width and height. The 9 different prior frame sizes were calculated to be [6,7], [11,10], [14,15], [19,19], [25, 25], [20,52], [37,36], [63,65], [98,119], respectively.
Step 4: and constructing a road identification detection model YOLOv7-SDB based on the improved YOLOv7, wherein the network structure is shown in figure 2.
The operation in the step 4 is as follows: firstly, performing concha splicing operation on 6 conv convolutions in an ELAN block in a backhaul part in a detection network and a CBS block, and then respectively introducing a GAM attention module after each concha operation in a Head to increase the feature extraction capability of the network. The GAM attention module is shown in fig. 3.
The GAM attention mechanism can stably improve the performance of model detection by designing a mechanism for reducing feature information generation and amplifying global dimension interaction features. Defining an input feature map as
Figure SMS_11
Intermediate state variable
Figure SMS_12
And final output +.>
Figure SMS_13
The expression of (2) is shown in formula 1:
Figure SMS_14
Figure SMS_15
(1)
wherein,,
Figure SMS_16
and->
Figure SMS_17
Channel and spatial attention patterns, respectively, +.>
Figure SMS_18
Representing a calculated multiplication of the element pattern.
Input features
Figure SMS_19
Finally, a signal is output as +.>
Figure SMS_20
Is used to preserve cross-three dimensional information using 3D placement by activating channel attention subunits and multi-layer MLP (multi-layer perceptron) is used to amplify cross-dimensional channel spatial correlation. In the spatial attention subunit, two convolutions are used to merge spatial information in order to fully focus on the spatial information.
Finally, the SPPCSPC module of the head part is replaced by the SPPFCSPC module by referring to the thought of SPPF. The SPPFCSPC module is based on the SPPCSPC module according to the idea of SPPF, and differs from the SPPCSPC module in that the SPPFCSPC module sequentially processes the 3 largest pooling layers, and can obtain a speed improvement without changing the receptive field.
For the final output 32 times downsampled feature map C5 of the Backbone, then the number of channels goes from 1024 to 512 through the SPPFCSPC module. Fusing the P3 layer, the P4 layer and the P5 layer with the C4 layer and the C3 layer in a top-down mode; and then fusing with the P4 layer and the P5 layer in a bottom-up mode. Finally, three detection layers with different sizes of 80×80, 40×40 and 20×20 are obtained, wherein 80×80 is used for detecting small targets, 40×40 is used for detecting small targets and 20×20 is used for detecting medium targets.
The Learning rate (Learning rate) based on the road identification detection model of improved YOLOv7 was set to 0.01, the optimizer was chosen to be random gradient descent (SGD), the Momentum (Momentum) was set to 0.937, the training iteration number (Batch) was 300 rounds, the Batch Size (batch_size) was 16, and the sio_loss was used as the boundary loss function.
The Siou_loss boundary loss function mainly includes a 4-part loss: angle loss (Angle cost), distance loss (Distance cost), shape loss (Shape cost), and cross ratio loss (IoU cost).
Angle cost (Angle cost):
Figure SMS_21
wherein the method comprises the steps of
Figure SMS_22
For the difference in height between the center points of the real and predicted frames,/for the difference in height between the center points of the real and predicted frames>
Figure SMS_23
For the distance of the center point of the real frame and the predicted frame, in fact +.>
Figure SMS_24
Equal to angle->
Figure SMS_25
Distance loss (Distance cost):
Figure SMS_26
wherein the method comprises the steps of
Figure SMS_27
,/>
Figure SMS_28
,/>
Figure SMS_29
Figure SMS_30
The width and height of the minimum bounding rectangle for the real and predicted frames. />
Figure SMS_31
For the center coordinates of the real frame, < >>
Figure SMS_32
Is the predicted frame center coordinates.
Shape penalty (Shape cost):
Figure SMS_33
=/>
Figure SMS_34
wherein the method comprises the steps of
Figure SMS_35
,/>
Figure SMS_36
,/>
Figure SMS_37
And->
Figure SMS_38
Width and height of prediction frame and real frame, respectively,/-for>
Figure SMS_39
Controlling the degree of concern over shape loss.
Cross-ratio loss (IoU cost):
Figure SMS_40
wherein the method comprises the steps of
Figure SMS_41
For predicting frame area, ++>
Figure SMS_42
Is the real frame area.
Finally, the total loss function is:
Figure SMS_43
step 5: the training results of the road identification detection model YOLOv7-SDB based on the improved YOLOv7 were evaluated.
The operation in the step 5 is as follows: the parameter of the statistical model, the detection speed FPS on the test set, the average accuracy mAP, and the original YOLOv7 target detection method were compared, and the results are shown in table 1 below.
Figure SMS_44
TABLE 1
As shown in the results of Table 1, compared with the original YOLOv7 network, the model of the invention improves mAP by 12.4% on the basis of properly improving parameter, improves the detection speed FPS by approximately 20%, and satisfies the detection accuracy and instantaneity.
The model training environment of the invention is: CPU uses Intel (R) Xeon (R) W-2102@ 2.90GHz, GPU uses GeForce RTX 2080Ti, running memory 64GB, operating system is Ubuntu 18.04.3 LTS, and deep learning framework is PyTorch.

Claims (7)

1. The road identification target detection method based on the improved YOLOv7 is characterized by comprising the following steps of:
1.1. acquiring a public road identification data set CCTSDB-2021, wherein the data set comprises 20491 road identification pictures and 3 categories which are respectively indication identifications, warning identifications and forbidden identifications;
1.2. 7500 pictures in the data set are taken, and the test set and the training set are divided according to a ratio of 8:2;
1.3. performing Mosaic-9 data enhancement processing on the selected picture;
1.4. k-means cluster analysis is carried out on the real label samples in the data set;
1.5. constructing a road identification detection model YOLOv7-SDB based on the improved YOLOv 7;
1.6. leading the training set into a detection model for training to obtain a road identification detection model;
1.7. and carrying out road identification detection on the test set by the detection model.
2. The improved YOLOv 7-based road marking target detection method of claim 1, wherein the step 1.3. Mosaic-9 data enhancement procedure comprises:
randomly selecting 9 pictures, performing augmentation operation on the 9 pictures respectively, and pasting the 9 pictures to the corresponding positions of the masks with the same size as the final output image respectively; augmentation operations include random clipping, scaling, alignment, and gamut variation; and obtaining a new picture and enriching the data set.
3. The improved YOLOv 7-based road identification target detection method according to claim 1, wherein the K-means clustering algorithm in step 1.4 clusters the widths and heights of all target frames contained in the training set to obtain 9 most representative combinations of widths and heights, and the 9 anchors are divided into 3 groups according to the sizes of the widths and the heights.
4. The method for detecting a road marking target based on improved YOLOv7 as claimed in claim 1, wherein the construction of the road marking detection model YOLOv7-SDB structure improvement based on improved YOLOv7 in step 1.5 comprises:
firstly, performing concha splicing operation on 6 conv convolutions in an ELAN block in a back band part in a detection network and a CBS block, then respectively introducing GAM (Global Attention Mechanism) attention modules after each concha operation in a Head part, and finally replacing an SPPCSPC module in the Head part with an SPPFCSPC module by referring to the idea of SPPF. For the final output 32 times downsampled feature map C5 of the Backbone, then the number of channels goes from 1024 to 512 through the SPPFCSPC module. Fusing the P3 layer, the P4 layer and the P5 layer with the C4 layer and the C3 layer in a top-down mode; and then fusing with the P4 layer and the P5 layer in a bottom-up mode. Finally, three detection layers with different sizes of 80×80, 40×40 and 20×20 are obtained, wherein 80×80 is used for detecting small targets, 40×40 is used for detecting small targets and 20×20 is used for detecting medium targets.
5. The improved YOLOv 7-based road marking target detection method of claim 1, wherein the GAM attention module comprises:
the GAM attention mechanism can stably promote by designing a mechanism for reducing feature information generation and amplifying global dimension interaction featuresPerformance of model test. Defining an input feature map as
Figure QLYQS_1
Intermediate state variable ∈ ->
Figure QLYQS_2
And final output +.>
Figure QLYQS_3
The expression of (2) is shown in formula 1:
Figure QLYQS_4
,/>
Figure QLYQS_5
(1)
wherein,,
Figure QLYQS_6
and->
Figure QLYQS_7
Channel and spatial attention patterns, respectively, +.>
Figure QLYQS_8
Representing a calculated multiplication of the element pattern. Input features->
Figure QLYQS_9
Finally, a signal is output as +.>
Figure QLYQS_10
Is used to preserve cross-three dimensional information using 3D placement by activating channel attention subunits and multi-layer MLP (multi-layer perceptron) is used to amplify cross-dimensional channel spatial correlation. In the spatial attention subunit, two convolutions are used to merge spatial information in order to fully focus on the spatial information.
6. The improved YOLOv 7-based road marking target detection method of claim 1, wherein the SPPFCSPC module comprises:
the SPPFCSPC module is based on the SPPCSPC module according to the idea of SPPF, with the difference that the SPPFCSPC module processes the 3 largest pooling layers sequentially, and can achieve a speed boost with unchanged receptive field.
7. The road marking target detection method based on modified YOLOv7 as claimed in claim 1, wherein a Learning rate (Learning rate) of the road marking detection model based on modified YOLOv7 is set to 0.01, an optimizer is selected to be a random gradient descent (SGD), a Momentum (Momentum) is set to 0.937, a training iteration number (Batch) is 300 rounds, a Batch Size (batch_size) is 16, and a sio_loss is used as a boundary loss function.
CN202310293143.5A 2023-03-24 2023-03-24 Road sign target detection method based on improved YOLOv7 Pending CN116343150A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310293143.5A CN116343150A (en) 2023-03-24 2023-03-24 Road sign target detection method based on improved YOLOv7

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310293143.5A CN116343150A (en) 2023-03-24 2023-03-24 Road sign target detection method based on improved YOLOv7

Publications (1)

Publication Number Publication Date
CN116343150A true CN116343150A (en) 2023-06-27

Family

ID=86894411

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310293143.5A Pending CN116343150A (en) 2023-03-24 2023-03-24 Road sign target detection method based on improved YOLOv7

Country Status (1)

Country Link
CN (1) CN116343150A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117152846A (en) * 2023-10-30 2023-12-01 云南师范大学 Student behavior recognition method, device and system and computer readable storage medium
CN117315614A (en) * 2023-11-28 2023-12-29 南昌大学 Traffic target detection method based on improved YOLOv7
CN117315934A (en) * 2023-09-25 2023-12-29 阜阳交通能源投资有限公司 Expressway traffic flow real-time monitoring and congestion prediction system based on unmanned aerial vehicle
CN117689731A (en) * 2024-02-02 2024-03-12 陕西德创数字工业智能科技有限公司 Lightweight new energy heavy-duty truck battery pack identification method based on improved YOLOv5 model
CN118247286A (en) * 2024-05-30 2024-06-25 佛山科学技术学院 Yolov-based steel surface defect detection method and yolov-based steel surface defect detection system

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117315934A (en) * 2023-09-25 2023-12-29 阜阳交通能源投资有限公司 Expressway traffic flow real-time monitoring and congestion prediction system based on unmanned aerial vehicle
CN117152846A (en) * 2023-10-30 2023-12-01 云南师范大学 Student behavior recognition method, device and system and computer readable storage medium
CN117152846B (en) * 2023-10-30 2024-01-26 云南师范大学 Student behavior recognition method, device and system and computer readable storage medium
CN117315614A (en) * 2023-11-28 2023-12-29 南昌大学 Traffic target detection method based on improved YOLOv7
CN117315614B (en) * 2023-11-28 2024-03-29 南昌大学 Traffic target detection method based on improved YOLOv7
CN117689731A (en) * 2024-02-02 2024-03-12 陕西德创数字工业智能科技有限公司 Lightweight new energy heavy-duty truck battery pack identification method based on improved YOLOv5 model
CN117689731B (en) * 2024-02-02 2024-04-26 陕西德创数字工业智能科技有限公司 Lightweight new energy heavy-duty battery pack identification method based on improved YOLOv model
CN118247286A (en) * 2024-05-30 2024-06-25 佛山科学技术学院 Yolov-based steel surface defect detection method and yolov-based steel surface defect detection system

Similar Documents

Publication Publication Date Title
CN116343150A (en) Road sign target detection method based on improved YOLOv7
CN112200161B (en) Face recognition detection method based on mixed attention mechanism
Zhang et al. CDNet: A real-time and robust crosswalk detection network on Jetson nano based on YOLOv5
CN102289686B (en) Method for identifying classes of moving targets based on transfer learning
CN111582339B (en) Vehicle detection and recognition method based on deep learning
CN114495029B (en) Traffic target detection method and system based on improved YOLOv4
CN113762209A (en) Multi-scale parallel feature fusion road sign detection method based on YOLO
CN112084901A (en) GCAM-based high-resolution SAR image airport runway area automatic detection method and system
Ghosh On-road vehicle detection in varying weather conditions using faster R-CNN with several region proposal networks
CN107346420A (en) Text detection localization method under a kind of natural scene based on deep learning
CN105354568A (en) Convolutional neural network based vehicle logo identification method
CN106326858A (en) Road traffic sign automatic identification and management system based on deep learning
CN112016605B (en) Target detection method based on corner alignment and boundary matching of bounding box
CN108921120B (en) Cigarette identification method suitable for wide retail scene
CN110619279A (en) Road traffic sign instance segmentation method based on tracking
CN110032952B (en) Road boundary point detection method based on deep learning
CN113378815B (en) Scene text positioning and identifying system and training and identifying method thereof
CN113239753A (en) Improved traffic sign detection and identification method based on YOLOv4
CN116206112A (en) Remote sensing image semantic segmentation method based on multi-scale feature fusion and SAM
CN108960175A (en) A kind of licence plate recognition method based on deep learning
CN113554030A (en) Multi-type license plate recognition method and system based on single character attention
CN117975418A (en) Traffic sign detection method based on improved RT-DETR
CN114066937B (en) Multi-target tracking method for large-scale remote sensing image
CN117912058A (en) Cattle face recognition method
CN111881914B (en) License plate character segmentation method and system based on self-learning threshold

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination