CN117333845A - Real-time detection method for small target traffic sign based on improved YOLOv5s - Google Patents
Real-time detection method for small target traffic sign based on improved YOLOv5s Download PDFInfo
- Publication number
- CN117333845A CN117333845A CN202311453323.1A CN202311453323A CN117333845A CN 117333845 A CN117333845 A CN 117333845A CN 202311453323 A CN202311453323 A CN 202311453323A CN 117333845 A CN117333845 A CN 117333845A
- Authority
- CN
- China
- Prior art keywords
- model
- real
- traffic sign
- yolov5s
- small target
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000011897 real-time detection Methods 0.000 title claims abstract description 33
- 238000000034 method Methods 0.000 title claims abstract description 25
- 238000001514 detection method Methods 0.000 claims abstract description 40
- 238000012549 training Methods 0.000 claims abstract description 25
- 238000012360 testing method Methods 0.000 claims abstract description 21
- 238000012795 verification Methods 0.000 claims abstract description 15
- 230000006870 function Effects 0.000 claims description 36
- 238000004364 calculation method Methods 0.000 claims description 13
- 238000011156 evaluation Methods 0.000 claims description 5
- 238000013459 approach Methods 0.000 claims description 4
- 238000000691 measurement method Methods 0.000 claims description 3
- 238000010276 construction Methods 0.000 claims description 2
- 238000005259 measurement Methods 0.000 claims 1
- 238000012544 monitoring process Methods 0.000 claims 1
- 238000013527 convolutional neural network Methods 0.000 description 6
- 230000008569 process Effects 0.000 description 6
- 238000011161 development Methods 0.000 description 3
- 230000018109 developmental process Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000004913 activation Effects 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000001537 neural effect Effects 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 230000001902 propagating effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 238000001629 sign test Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/582—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
技术领域Technical field
本发明涉及人工智能领域,具体涉及一种小目标交通标志实时检测方法。The invention relates to the field of artificial intelligence, and specifically relates to a real-time detection method of small target traffic signs.
背景技术Background technique
随着自动驾驶技术的不断发展,道路交通标志识别(Traffic Sign Recognition,TSR)已经成为辅助驾驶系统(Advanced Driver Assistance Systems,ADAS)的关键组件之一。TSR能够对摄像头捕获的图像进行高效且准确的识别,实时地为驾驶员提供指示、警告等信息。在实际道路行驶过程中,交通标志在刚进入视野时所占的图像比例通常小于1%,属于小目标检测,小目标交通标志在图像中存在模糊、细节特征不明显的问题,因此,对小目标交通标志的快速、准确的识别,对于辅助驾驶系统的发展具有重要意义。With the continuous development of autonomous driving technology, Traffic Sign Recognition (TSR) has become one of the key components of the Advanced Driver Assistance Systems (ADAS). TSR can efficiently and accurately identify images captured by cameras and provide drivers with instructions, warnings and other information in real time. In the actual road driving process, the proportion of the image occupied by traffic signs when they first enter the field of view is usually less than 1%, which belongs to small target detection. Small target traffic signs have problems such as blur and unclear details in the image. Therefore, it is difficult to detect small target traffic signs. The rapid and accurate recognition of target traffic signs is of great significance to the development of assisted driving systems.
传统交通标志检测方法主要集中在颜色分割,形状、轮廓等特征进行特征提取,然后通过分类器完成特征分类,从而实现对道路交通标志的检测与识别。传统交通标志检测算法利用手工制作的方式提取特征,然后将这些特征作为算法的输入,这些检测算法具有维度高、计算量大、费时费力,交通标志检测过程中存在实时性、稳定性和精确性不佳。Traditional traffic sign detection methods mainly focus on color segmentation, feature extraction of shapes, contours and other features, and then complete feature classification through classifiers to achieve detection and recognition of road traffic signs. Traditional traffic sign detection algorithms use manual methods to extract features, and then use these features as input to the algorithm. These detection algorithms have high dimensions, large amounts of calculation, and are time-consuming and labor-intensive. There are real-time, stability, and accuracy issues in the traffic sign detection process. Not good.
随着深度学习的快速发展,改进卷积神经模型的目标检测技术已成为当下研究热点。卷积神经模型(Convolutional Neural Network,CNN)作为一种主流的深度学习算法,包括R-CNN、Fast-R-CNN、Fast-R-CNN、Mask-R-CNN、Alex Net、YOLO等多种模型。这些模型在不同应用场景下的目标检测取得了较好的效果。目标检测算法可分为两类:使用区域候选框的双阶段检测算法和不使用区域候选框的单阶段检测算法。双阶段算法精度高但检测速度较慢,改进后的算法在一定程度上提高了速度,但由于候选区域的设置,双阶段模型不能较好的满足高速驾驶过程中小目标交通标志的实时检测要求,因此在交通标志的识别中通常采用单阶段算法。单阶段目标检测算法方面,相较于双阶段检测算法虽然实时性有所提高,但精度却较低。尤其在小目标交通标志识别检测中,存在检测精度不高,容易导致漏检和误检,且由于模型参数量大、计算复杂度高导致检测实时性不好,模型迁移到可移动设备中具有一定困难。With the rapid development of deep learning, improving the target detection technology of convolutional neural models has become a current research hotspot. Convolutional Neural Network (CNN) is a mainstream deep learning algorithm, including R-CNN, Fast-R-CNN, Fast-R-CNN, Mask-R-CNN, Alex Net, YOLO, etc. Model. These models have achieved good results in target detection in different application scenarios. Target detection algorithms can be divided into two categories: two-stage detection algorithms that use region candidate frames and single-stage detection algorithms that do not use region candidate frames. The two-stage algorithm has high accuracy but slow detection speed. The improved algorithm improves the speed to a certain extent. However, due to the setting of the candidate area, the two-stage model cannot better meet the real-time detection requirements of small target traffic signs during high-speed driving. Therefore, a single-stage algorithm is usually used in traffic sign recognition. In terms of single-stage target detection algorithm, although the real-time performance is improved compared to the two-stage detection algorithm, the accuracy is lower. Especially in the recognition and detection of small target traffic signs, the detection accuracy is not high, which can easily lead to missed detections and false detections. Moreover, due to the large number of model parameters and high computational complexity, the real-time detection performance is not good. Migrating the model to mobile devices has the disadvantage of It must be difficult.
发明内容Contents of the invention
针对现有技术中存在的问题,本发明提供了一种基于改进YOLOv5s的小目标交通标志实时检测方法,通过改进YOLOv5s目标检测模型,利用特征增强以及采用NWD度量方法,解决小目标交通标志检测精度不高、实时性不好、模型参数量大、计算复杂度高、不适合部署在移动设备上的问题。In view of the problems existing in the existing technology, the present invention provides a real-time detection method of small target traffic signs based on improved YOLOv5s. By improving the YOLOv5s target detection model, utilizing feature enhancement and adopting the NWD measurement method, the detection accuracy of small target traffic signs is solved. It is not high, has poor real-time performance, has a large number of model parameters, has high computational complexity, and is not suitable for deployment on mobile devices.
本发明提供的技术方案包括以下步骤:The technical solution provided by the invention includes the following steps:
步骤1:获取交通标志图像,形成第一数据集;Step 1: Obtain traffic sign images to form the first data set;
步骤2:为所述第一数据集中的图像添加标注信息,形成第二数据集,并将所述第二数据集划分训练集、验证集和测试集;Step 2: Add annotation information to the images in the first data set to form a second data set, and divide the second data set into a training set, a verification set and a test set;
步骤3:构建基于改进YOLOv5s的交通标志检测模型;Step 3: Build a traffic sign detection model based on improved YOLOv5s;
步骤4:采用新的损失函数;Step 4: Use a new loss function;
步骤5:利用所述训练集和验证集对所述基于改进YOLOv5s的交通标志检测模型进行训练;Step 5: Use the training set and verification set to train the traffic sign detection model based on improved YOLOv5s;
步骤6:使用交通标志测试集对训练获取的最优模型进行测试,并获取最终的小目标交通标志实时检测模型。Step 6: Use the traffic sign test set to test the optimal model obtained through training, and obtain the final small target traffic sign real-time detection model.
进一步的,所述步骤1中,第一数据集中的交通标志图像可通过数码相机进行拍摄,也可从网络中收集获取,或者从监控视频中获取。Further, in step 1, the traffic sign images in the first data set can be captured by a digital camera, collected from the Internet, or obtained from surveillance videos.
优选的,所述步骤2中,可按照7∶2∶1的比例进行划分训练集、验证集和测试集。Preferably, in step 2, the training set, verification set and test set can be divided according to the ratio of 7:2:1.
进一步的,所述步骤3具体包括步骤3.1至步骤3.3:Further, the step 3 specifically includes step 3.1 to step 3.3:
步骤3.1:将YOLOv5s模型中的C3模块和ConvMixer相结合,构成CSPCM模块,用CSPCM模块分别替换YOLOv5s主干网络最后一层的C3模块和颈部网络中最后一层的C3模块;Step 3.1: Combine the C3 module and ConvMixer in the YOLOv5s model to form the CSPCM module, and use the CSPCM module to replace the C3 module at the last layer of the YOLOv5s backbone network and the C3 module at the last layer in the neck network respectively;
步骤3.2:用轻量化卷积模块C3_Faster替换YOLOv5s模型主干网络和颈部网络中其余的C3模块,即采用C3_Faster替换除所述步骤3.1中已经替换的主干网络和颈部网络中最后一层的C3模块之外的其余的C3模块;Step 3.2: Replace the remaining C3 modules in the backbone network and neck network of the YOLOv5s model with the lightweight convolution module C3_Faster, that is, use C3_Faster to replace the last layer of C3 in the backbone network and neck network except the ones that have been replaced in step 3.1. The remaining C3 modules outside the module;
步骤3.3:输出层在已有的3个检测头的基础上增加一个小目标检测头。Step 3.3: The output layer adds a small target detection head to the existing three detection heads.
进一步的,所述步骤4中,在CIoU损失函数中引入NWD度量方法,使用NWD度量优化YOLOv5s的CIoU损失函数,优化后的损失函数公式为:Further, in step 4, the NWD metric method is introduced into the CIoU loss function, and the NWD metric is used to optimize the CIoU loss function of YOLOv5s. The optimized loss function formula is:
L=(1-β)*(1-NWD(Na,Nb))+β*(1-CIoU) (1)L=(1-β)*(1-NWD(N a , N b ))+β*(1-CIoU) (1)
式(1)中,L是优化后的损失函数,NWD是归一化的Wasserstein距离,Na,Nb是由 和/>建模的高斯分布,a表示真实框,b表示预测框,cxa,,cya代表真实框的中心点坐标,wa、ha代表真实框的宽度和高度;cxb,cyb代表预测框的中心点坐标,wb,hb代表预测框的宽度和高度;β是权重比例系数,CIoU是原YOLOv5s中的损失函数,其计算公式为:In formula (1), L is the optimized loss function, NWD is the normalized Wasserstein distance, N a and N b are given by and/> Modeled Gaussian distribution, a represents the real box, b represents the predicted box, cx a , cy a represents the center point coordinates of the real box, w a , ha represents the width and height of the real box; cx b , cy b represents prediction The center point coordinates of the box, w b and h b represent the width and height of the prediction box; β is the weight proportion coefficient, CIoU is the loss function in the original YOLOv5s, and its calculation formula is:
式(2)中,ρ2(bA,bB)表示真实框与预测框中心点之间的欧几里得距离,c表示预测框和真实框的最小限定矩形的对角线距离,α为权重因子,v为纵横比一致性,IoU为真实框与预测框之间的重叠比例。In formula (2), ρ 2 (b A ,b B ) represents the Euclidean distance between the center point of the real box and the prediction box, c represents the diagonal distance of the minimum limiting rectangle between the prediction box and the real box, α is the weight factor, v is the aspect ratio consistency, and IoU is the overlap ratio between the real box and the predicted box.
进一步的,所述步骤5具体包括步骤5.1至步骤5.4:Further, the step 5 specifically includes step 5.1 to step 5.4:
步骤5.1:设定所述基于改进YOLOv5s的小目标交通标志实时检测模型训练参数,模型训练参数具体包括:学习率,动量,权重衰减,优化器,迭代轮数,批大小;Step 5.1: Set the training parameters of the real-time detection model of small target traffic signs based on improved YOLOv5s. The model training parameters specifically include: learning rate, momentum, weight attenuation, optimizer, number of iteration rounds, and batch size;
步骤5.2:将训练集和验证集图像以及对应标签输入到所述改进YOLOv5s的小目标交通标志实时检测模型中,使用反向传播算法计算损失函数对模型参数的梯度,通过最小化损失函数调整模型参数,使其逐渐接近最优解;Step 5.2: Input the training set and verification set images and corresponding labels into the improved YOLOv5s small target traffic sign real-time detection model, use the back propagation algorithm to calculate the gradient of the loss function on the model parameters, and adjust the model by minimizing the loss function parameters so that it gradually approaches the optimal solution;
步骤5.3:使用优化器来更新模型参数,使其朝着梯度下降的方向更新;直到训练集和验证集的损失函数不再下降,同时准确率P、召回率R、mAP等评价指标也不再提高;Step 5.3: Use the optimizer to update the model parameters in the direction of gradient descent; until the loss functions of the training set and verification set no longer decrease, and the evaluation indicators such as accuracy P, recall R, and mAP no longer decrease. improve;
步骤5.4:将训练好的模型参数保存为最优模型。Step 5.4: Save the trained model parameters as the optimal model.
进一步的,所述步骤6具体包括步骤6.1至步骤6.3:Further, the step 6 specifically includes step 6.1 to step 6.3:
步骤6.1:将所述测试集输入到步骤5所述的改进后的最优模型;Step 6.1: Input the test set into the improved optimal model described in step 5;
步骤6.2:计算模型性能指标:性能指标具体包括准确率P(Precision)、召回率R、mAP、参数量、计算复杂度(GFLOPs)、模型大小,具体公式如下:Step 6.2: Calculate model performance indicators: Performance indicators specifically include accuracy P (Precision), recall rate R, mAP, number of parameters, computational complexity (GFLOPs), and model size. The specific formula is as follows:
其中,P为准确率,R为召回率,mAP为所有类别的平均精确率均值,AP为平均精确率,m为交通标志总类别数,TP表示正样本被正确识别为正样本的个数,FP表示负样本被错误识别为正样本的个数,FN表示正样本被错误识别为负样本的个数;Among them, P is the precision rate, R is the recall rate, mAP is the average precision rate mean of all categories, AP is the average precision rate, m is the total number of traffic sign categories, TP represents the number of positive samples that are correctly recognized as positive samples, FP represents the number of negative samples that are incorrectly identified as positive samples, and FN represents the number of positive samples that are incorrectly identified as negative samples;
步骤6.3:当性能指标满足精度要求时,获得最终基于改进YOLOv5s的小目标交通标志实时检测模型。Step 6.3: When the performance indicators meet the accuracy requirements, obtain the final real-time detection model of small target traffic signs based on improved YOLOv5s.
与现有技术相比,本发明的有益效果是:Compared with the prior art, the beneficial effects of the present invention are:
本发明公开的一种基于改进YOLOv5s的小目标交通标志实时检测方法,通过引入CSPCM模块和增加小目标检测头,提高了小目标交通标志的检测精度;采用C3_Faster网络结构减少了模型参数和计算复杂度,提高了交通标志检测的实时性;在CIoU损失函数中引入NWD度量方法,采用了新的损失函数,有效避免了小目标带来的较大损失,加快了模型的收敛。The invention discloses a real-time detection method of small target traffic signs based on improved YOLOv5s. By introducing the CSPCM module and adding a small target detection head, the detection accuracy of small target traffic signs is improved; the C3_Faster network structure is used to reduce model parameters and calculation complexity. degree, which improves the real-time performance of traffic sign detection; the NWD measurement method is introduced in the CIoU loss function, and a new loss function is adopted, which effectively avoids large losses caused by small targets and accelerates the convergence of the model.
附图说明Description of drawings
图1为本发明基于改进YOLOv5s的小目标交通标志实时检测方法的流程图;Figure 1 is a flow chart of the real-time detection method of small target traffic signs based on improved YOLOv5s according to the present invention;
图2为本发明基于改进YOLOv5s的小目标交通标志实时检测方法模型结构示意图;Figure 2 is a schematic structural diagram of the model structure of the real-time detection method of small target traffic signs based on improved YOLOv5s according to the present invention;
图3为CSPCM结构示意图;Figure 3 is a schematic diagram of the CSPCM structure;
图4为C3_Faster结构示意图;Figure 4 is a schematic diagram of the C3_Faster structure;
具体实施方式Detailed ways
为了使本发明的技术方案、构造特点、所实现的目的及优点更加清楚明白,以下结合具体实施方式并配合附图,对本发明进行进一步详细说明。需要说明的是,此处所描述的具体实施例仅用于更清楚的解释本发明,并不用于限定本发明。In order to make the technical solutions, structural features, achieved objectives and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with specific embodiments and accompanying drawings. It should be noted that the specific embodiments described here are only used to explain the present invention more clearly and are not used to limit the present invention.
图1是本发明公开的一种基于改进YOLOv5s的小目标交通标志实时检测方法流程图,其实现过程如下:Figure 1 is a flow chart of a real-time detection method for small target traffic signs based on improved YOLOv5s disclosed by the present invention. The implementation process is as follows:
步骤1:获取交通标志图像,形成第一数据集;所述第一数据集中的交通标志图像可通过数码相机进行拍摄、从网络中收集获取,或者从监控视频中获取;Step 1: Obtain traffic sign images to form a first data set; the traffic sign images in the first data set can be photographed by a digital camera, collected from the Internet, or obtained from surveillance videos;
在本实施例中,为了更好的评价本发明公开的一种基于改进YOLOv5s的小目标交通标志实时检测方法的检测效果,采取了公开数据集TT100K(Tsinghua-Tencent 100K)交通标志数据集。采用TT100K数据集,需要对数据集进行预处理:首先,统计TT100K数据集中出现次数小于200次的交通标志类,删除类别出现次数小于200次的交通标志类所对应的标签和只包含该类别的图片,最终制成含36类交通标志的数据集,即形成所述第一数据集。In this embodiment, in order to better evaluate the detection effect of a real-time detection method for small target traffic signs based on improved YOLOv5s disclosed in the present invention, the public data set TT100K (Tsinghua-Tencent 100K) traffic sign data set was used. Using the TT100K data set, the data set needs to be preprocessed: First, count the traffic sign categories that appear less than 200 times in the TT100K data set, delete the labels corresponding to the traffic sign categories that appear less than 200 times, and the labels that only contain this category. The pictures are finally made into a data set containing 36 types of traffic signs, which forms the first data set.
步骤2:为所述第一数据集中的图像使用lableme标注工具添加标注信息。由于本实施例中采取的公开数据集TT100K中已有标注信息,故略过该步骤。将本实施例中所述的第一数据集中json格式的标签文件转换为YOLOv5s所需要的txt格式标签文件,形成第二数据集,并按照7:2:1将所述第二数据集划分训练集、验证集和测试集;划分后的训练集包括6598张图片,测试集包括1889张图片,验证集包括970张图片,本实施例中所述第二数据集中图像尺寸大小为640*640。Step 2: Use labelme annotation tool to add annotation information to the images in the first data set. Since the public data set TT100K used in this embodiment already has annotation information, this step is skipped. Convert the json format label file in the first data set described in this embodiment into the txt format label file required by YOLOv5s to form a second data set, and divide the second data set according to 7:2:1 for training Set, verification set and test set; the divided training set includes 6598 pictures, the test set includes 1889 pictures, and the verification set includes 970 pictures. The image size in the second data set described in this embodiment is 640*640.
步骤3:构建基于改进YOLOv5s的交通标志检测模型,改进后的YOLOv5s模型结构如图2所示,模型构建过程具体包括步骤3.1至步骤3.3:Step 3: Construct a traffic sign detection model based on improved YOLOv5s. The improved YOLOv5s model structure is shown in Figure 2. The model construction process specifically includes steps 3.1 to 3.3:
步骤3.1:将C3模块中的Bottleneck模块用ConvMixer模块进行替换,形成新的卷积模块CSPCM;用新的卷积CSPCM模块替换了Backbone最后一层的C3模块和Neck最后一层的C3模块;Step 3.1: Replace the Bottleneck module in the C3 module with the ConvMixer module to form a new convolution module CSPCM; replace the C3 module of the last layer of Backbone and the C3 module of the last layer of Neck with the new convolution CSPCM module;
进一步的,ConvMixer模块的结构如图3所示,它采用分离空间和通道维度的混合方式进行操作,并在整个模型中保持相同的大小和分辨率;在ConvMixer中,输入先通过一个Depthwise Conv(Dwconv),即组数等于通道数的分组卷积;然后通过一个PointwiseConv(PWCONV),即1×1卷积;每个卷积操作之后,都跟随一个激活函数z′1和一个z1+ 1BatchNorm层。这种设计能够有效地减少模型参数量,有利于提高模型的实时性,并有效提高模型特征表达能力。Further, the structure of the ConvMixer module is shown in Figure 3. It operates in a mixed manner of separating space and channel dimensions, and maintains the same size and resolution throughout the model; in ConvMixer, the input is first passed through a Depthwise Conv( Dwconv), that is, a grouped convolution with the number of groups equal to the number of channels; and then through a PointwiseConv (PWCONV), which is a 1×1 convolution; after each convolution operation, it is followed by an activation function z′ 1 and a z 1+ 1 BatchNorm layer. This design can effectively reduce the amount of model parameters, improve the real-time performance of the model, and effectively improve the model's feature expression ability.
步骤3.2:使用轻量化卷积模块C3_Faster替换YOLOv5s模型主干网络和颈部网络中其余的C3模块,即采用C3_Faster替换除步骤3.1中已经替换的主干网络和颈部网络中最后一层的C3模块之外的其余的C3模块;Step 3.2: Use the lightweight convolution module C3_Faster to replace the remaining C3 modules in the backbone network and neck network of the YOLOv5s model, that is, use C3_Faster to replace the C3 modules in the last layer of the backbone network and neck network that have been replaced in step 3.1. The remaining C3 modules;
轻量级卷积模块C3_Faster使用了Pconv卷积;相较于深度卷积(DWConv)和组卷积(GConv),PConv可以减少FLOPs的计算量,减少计算冗余和内存访问,图4为C3_Faster的结构,由一个PConv层后跟2个Conv层组成,它们一起显示为倒置残差块,中间层具有扩展的通道数量,每个中间Conv之后,都会有一个标准化和激活层。The lightweight convolution module C3_Faster uses Pconv convolution; compared with depth convolution (DWConv) and group convolution (GConv), PConv can reduce the calculation amount of FLOPs, reduce computational redundancy and memory access, Figure 4 shows C3_Faster The structure consists of a PConv layer followed by 2 Conv layers, which are shown together as an inverted residual block, the intermediate layer has an extended number of channels, and after each intermediate Conv, there is a normalization and activation layer.
步骤3.3:输出层在已有的3个检测头的基础上增加一个小目标检测头,能够更高效地处理多尺度特征信息,使检测器对小目标更加敏感,提高了对小目标交通标志的检测性能。Step 3.3: The output layer adds a small target detection head to the existing three detection heads, which can process multi-scale feature information more efficiently, make the detector more sensitive to small targets, and improve the detection of small target traffic signs. Detection performance.
步骤4:采用新的损失函数,在CIoU损失函数中引入NWD度量方法,将CIoU与NWD度量相结合,提出一个新的损失函数;新的损失函数能有效地避免小目标带来的较大损失,加快模型的收敛,具体步骤包括步骤4.1至步骤4.3:Step 4: Use a new loss function, introduce the NWD metric method into the CIoU loss function, combine CIoU with the NWD metric, and propose a new loss function; the new loss function can effectively avoid large losses caused by small targets. , to speed up the convergence of the model. The specific steps include steps 4.1 to 4.3:
步骤4.1:将边界框建模为二维高斯分布,利用预测目标与实际目标对应的高斯分布,计算预测框与真实框之间的Wasserstein距离,计算公式为:Step 4.1: Model the bounding box as a two-dimensional Gaussian distribution, and use the Gaussian distribution corresponding to the predicted target and the actual target to calculate the Wasserstein distance between the predicted box and the real box. The calculation formula is:
式(1)中,是Wasserstein距离,Na,Nb是由/>和 建模的高斯分布,a表示真实框,b表示预测框,cxa,,cya代表真实框的中心点坐标,wa、ha代表真实框的宽度和高度;cxb,cyb代表预测框的中心点坐标,wb,hb代表预测框的宽度和高度;In formula (1), is the Wasserstein distance, N a and N b are given by/> and Modeled Gaussian distribution, a represents the real box, b represents the predicted box, cx a , cy a represents the center point coordinates of the real box, w a , ha represents the width and height of the real box; cx b , cy b represents prediction The center point coordinates of the box, w b , h b represent the width and height of the prediction box;
步骤4.2:计算它们之间归一化的Wasserstein距离,计算公式为:Step 4.2: Calculate the normalized Wasserstein distance between them. The calculation formula is:
式(2)中,NWD(Na,Nb)是归一化的Wasserstein距离,C是与数据集密切相关的常数。步骤4.3:根据CIoU与NWD的比例关系,提出新的损失函数,其计算公式为:In formula (2), NWD (N a , N b ) is the normalized Wasserstein distance, and C is a constant closely related to the data set. Step 4.3: Based on the proportional relationship between CIoU and NWD, a new loss function is proposed, and its calculation formula is:
L=(1-β)*(1-NWD(Na,Nb))+β*(1-CIoU) (3)L=(1-β)*(1-NWD(N a , N b ))+β*(1-CIoU) (3)
式(3)中,L是优化后的损失函数,β是权重比例系数,CIoU是原YOLOv5s中的损失函数,其计算公式为:In formula (3), L is the optimized loss function, β is the weight proportion coefficient, CIoU is the loss function in the original YOLOv5s, and its calculation formula is:
式(4)中,ρ2(bA,bB)表示真实框与预测框中心点之间的欧几里得距离,c表示预测框和真实框的最小限定矩形的对角线距离,α为权重因子,v为纵横比一致性,IntersectionoverUnion(IoU)是目标检测用于衡量预测边界框和真实边界框之间的重叠程度。IoU值越高,表示预测框和真实框之间的重叠程度越大,检测效果越好,IoU计算公式为:In formula (4), ρ 2 (b A , b B ) represents the Euclidean distance between the center point of the real frame and the prediction frame, c represents the diagonal distance of the minimum limiting rectangle between the prediction frame and the real frame, α is the weight factor, v is the aspect ratio consistency, and IntersectionoverUnion (IoU) is the object detection method used to measure the overlap between the predicted bounding box and the real bounding box. The higher the IoU value, the greater the overlap between the predicted frame and the real frame, and the better the detection effect. The IoU calculation formula is:
式(5)中:B为预测框;BGT为真实框。In formula (5): B is the prediction box; B GT is the real box.
步骤5:将所述训练集和验证集输入到步骤3所述的基于改进YOLOv5s的小目标交通标志模型中进行训练,具体包括步骤5.1至步骤5.4:Step 5: Input the training set and verification set into the small target traffic sign model based on improved YOLOv5s described in step 3 for training, specifically including steps 5.1 to 5.4:
步骤5.1:设定所述基于改进YOLOv5s的小目标交通标志实时检测模型的训练参数,模型训练参数包括:学习率,动量,权重衰减,优化器,迭代轮数,批大小;Step 5.1: Set the training parameters of the small target traffic sign real-time detection model based on improved YOLOv5s. The model training parameters include: learning rate, momentum, weight attenuation, optimizer, number of iteration rounds, and batch size;
在本实施例中,优化器optimizer为SGD,初始学习率1r0为0.01,动量momentum为0.937,权重衰减weight_decay为0.0005,批大小batchsize为16,迭代轮次Epoch为300。In this embodiment, the optimizer is SGD, the initial learning rate 1r0 is 0.01, the momentum is 0.937, the weight decay weight_decay is 0.0005, the batch size is 16, and the iteration round Epoch is 300.
步骤5.2:将训练集和验证集图像以及对应标签输入到所述改进YOLOv5s的小目标交通标志实时检测模型中,使用反向传播算法计算损失函数对模型参数的梯度。反向传播算法是一种有效的计算梯度的方法,它使用链式法则来计算每个参数对于损失函数的梯度。具体的,反向传播通过将损失函数从输出层向后传播,逐层计算每个参数的梯度。在这个过程中,每个参数的梯度表示了损失函数对于该参数的变化率,即损失函数如何随着该参数的变化而变化。通过最小化损失函数调整模型参数使其逐渐接近最优解。Step 5.2: Input the training set and verification set images and corresponding labels into the improved YOLOv5s small target traffic sign real-time detection model, and use the back propagation algorithm to calculate the gradient of the loss function on the model parameters. The backpropagation algorithm is an efficient method for calculating gradients. It uses the chain rule to calculate the gradient of each parameter with respect to the loss function. Specifically, backpropagation calculates the gradient of each parameter layer by layer by propagating the loss function backward from the output layer. In this process, the gradient of each parameter represents the rate of change of the loss function for that parameter, that is, how the loss function changes as the parameter changes. Adjust the model parameters by minimizing the loss function to gradually approach the optimal solution.
步骤5.3:计算得到模型参数的梯度后,使用优化器来更新这些参数。优化器根据参数的梯度信息,按照梯度的反方向进行参数更新。梯度较大的参数会以更大的步伐进行更新,而梯度较小的参数则以较小的步伐进行更新。通过不断迭代更新模型参数,可以逐渐减小损失函数的值。通过最小化损失函数,调整模型参数的取值,使其逐渐接近最优解,即损失函数达到最小值的参数取值,模型的预测结果与真实值之间的差异最小化,同时mAP、召回率R、准确率P等评价指标也不再提高。Step 5.3: After calculating the gradients of the model parameters, use the optimizer to update these parameters. The optimizer updates parameters in the opposite direction of the gradient based on the gradient information of the parameters. Parameters with larger gradients are updated in larger steps, while parameters with smaller gradients are updated in smaller steps. By continuously updating the model parameters iteratively, the value of the loss function can be gradually reduced. By minimizing the loss function, the values of the model parameters are adjusted so that they gradually approach the optimal solution, that is, the parameter value at which the loss function reaches the minimum value, the difference between the model's prediction results and the true value is minimized, and at the same time, mAP, recall Evaluation indicators such as rate R and accuracy P also no longer improve.
步骤5.4:将训练好的模型参数保存为最优模型。Step 5.4: Save the trained model parameters as the optimal model.
步骤6:采用所述测试集对步骤5所述的最优模型进行测试,对测试集测试结果进行评估,满足精度要求,即获得最终基于改进YOLOv5s的小目标交通标志实时检测模型,具体的,所述步骤6进一步包括步骤6.1至步骤6.3:Step 6: Use the test set to test the optimal model described in step 5, evaluate the test set test results, and meet the accuracy requirements, that is, obtain the final real-time detection model of small target traffic signs based on improved YOLOv5s. Specifically, The step 6 further includes steps 6.1 to 6.3:
步骤6.1:将所述测试集输入到步骤5所述的最优模型中;Step 6.1: Input the test set into the optimal model described in step 5;
步骤6.2:计算模型性能指标:性能指标具体包括准确率P、召回率R、mAP、参数量、计算复杂度GFLOPs、模型大小,具体计算公式如下:Step 6.2: Calculate model performance indicators: Performance indicators specifically include accuracy P, recall rate R, mAP, number of parameters, computational complexity GFLOPs, and model size. The specific calculation formula is as follows:
其中,P为准确率,R为召回率,mAP为所有类别的平均精确率均值,AP为平均精确率,m为交通标志总类别数,TP表示正样本被正确识别为正样本的个数,FP表示负样本被错误识别为正样本的个数,FN表示正样本被错误识别为负样本的个数;Among them, P is the precision rate, R is the recall rate, mAP is the average precision rate mean of all categories, AP is the average precision rate, m is the total number of traffic sign categories, TP represents the number of positive samples that are correctly recognized as positive samples, FP represents the number of negative samples that are incorrectly identified as positive samples, and FN represents the number of positive samples that are incorrectly identified as negative samples;
步骤6.3:当性能指标满足精度要求时,获得最终基于改进YOLOv5s的小目标交通标志实时检测模型。Step 6.3: When the performance indicators meet the accuracy requirements, obtain the final real-time detection model of small target traffic signs based on improved YOLOv5s.
在本实施例中,为了验证本发明公开的改进后的模型效果,将测试集输入到所述基于改进YOLOv5s的小目标交通标志实时检测模型中以及YOLOv5s模型进行测试。对两个模型分别计算评价指标,评价指标数据见表1所示。由表1可知,本发明公开的小目标交通标志实时检测模型比原YOLOv5s模型在准确率P、mAP@0.5等指标都获得了较高的检测精度。在模型参数量、模型大小、计算复杂度GFLOPs等性能指标上,本发明公开的改进模型比原YOLOv5s模型更低,从而使模型具有较好的轻量化指标,更容易部署到移动设备上。In this embodiment, in order to verify the improved model effect disclosed in the present invention, the test set is input into the small target traffic sign real-time detection model based on improved YOLOv5s and the YOLOv5s model is tested. The evaluation indicators are calculated separately for the two models, and the evaluation indicator data are shown in Table 1. As can be seen from Table 1, the real-time detection model of small target traffic signs disclosed in the present invention has achieved higher detection accuracy than the original YOLOv5s model in terms of accuracy P, mAP@0.5 and other indicators. In terms of performance indicators such as model parameter quantity, model size, and computational complexity GFLOPs, the improved model disclosed in the present invention is lower than the original YOLOv5s model, so that the model has better lightweight indicators and is easier to deploy on mobile devices.
表1对比实验结果Table 1 Comparative experimental results
以上所述仅为本发明的一个实施例,并非因此限制本发明的专利范围,对于本领域的技术人员来说,本发明可以有各种更改和变化,凡在本发明的精神和原则之内,所做的任何修改、同等替换、改进等,均应包含在本发明的保护范围之内。The above is only an embodiment of the present invention, and does not limit the patent scope of the present invention. For those skilled in the art, the present invention can have various modifications and changes, all within the spirit and principles of the present invention. , any modifications, equivalent substitutions, improvements, etc. shall be included in the protection scope of the present invention.
Claims (3)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311453323.1A CN117333845B (en) | 2023-11-03 | 2023-11-03 | A real-time detection method for small target traffic signs based on improved YOLOv5s |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311453323.1A CN117333845B (en) | 2023-11-03 | 2023-11-03 | A real-time detection method for small target traffic signs based on improved YOLOv5s |
Publications (2)
Publication Number | Publication Date |
---|---|
CN117333845A true CN117333845A (en) | 2024-01-02 |
CN117333845B CN117333845B (en) | 2025-02-07 |
Family
ID=89275667
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311453323.1A Active CN117333845B (en) | 2023-11-03 | 2023-11-03 | A real-time detection method for small target traffic signs based on improved YOLOv5s |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117333845B (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118230079A (en) * | 2024-05-27 | 2024-06-21 | 中国科学院西安光学精密机械研究所 | A detection method for small remote sensing targets based on improved YOLO |
CN118570740A (en) * | 2024-08-02 | 2024-08-30 | 南京信息工程大学 | Bus stop scene vehicle pedestrian detection method based on improvement YOLOV5 |
CN118710607A (en) * | 2024-06-18 | 2024-09-27 | 东北电力大学 | A lightweight road surface defect detection method based on improved YOLOv7 |
CN118865174A (en) * | 2024-07-18 | 2024-10-29 | 电子科技大学 | A fast detection method for small moving targets in UAVs |
CN119693923A (en) * | 2025-02-24 | 2025-03-25 | 浙江理工大学 | Path planning method combining small-scale traffic sign recognition and DWA algorithm |
CN119810761A (en) * | 2024-12-05 | 2025-04-11 | 东北电力大学 | Unmanned aerial vehicle detection method based on improved YOLO11n |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200210826A1 (en) * | 2018-12-29 | 2020-07-02 | Northeastern University | Intelligent analysis system using magnetic flux leakage data in pipeline inner inspection |
US20210370993A1 (en) * | 2020-05-27 | 2021-12-02 | University Of South Carolina | Computer vision based real-time pixel-level railroad track components detection system |
CN114998831A (en) * | 2022-05-23 | 2022-09-02 | 重庆科技学院 | Fire detection method and system based on edge calculation and improved YOLOv3 |
CN116092115A (en) * | 2022-11-28 | 2023-05-09 | 中国计量大学 | A real-time lightweight detection method for safety clothing of construction workers |
CN116259032A (en) * | 2022-11-30 | 2023-06-13 | 长春工业大学 | Road traffic sign detection and identification algorithm based on improved YOLOv5 |
CN116434188A (en) * | 2023-03-26 | 2023-07-14 | 复旦大学 | Traffic sign detection method based on improved_yolov5s network |
CN116469021A (en) * | 2023-04-30 | 2023-07-21 | 桂林电子科技大学 | Unmanned aerial vehicle aerial photographing target detection method based on improved YOLOv5 |
CN116597413A (en) * | 2023-04-28 | 2023-08-15 | 哈尔滨工业大学 | Real-time traffic sign detection method based on improved YOLOv5 |
CN116630920A (en) * | 2023-05-18 | 2023-08-22 | 天津科技大学 | Lane line type recognition method of improved YOLOv5s network model |
US11790640B1 (en) * | 2022-06-22 | 2023-10-17 | Ludong University | Method for detecting densely occluded fish based on YOLOv5 network |
US20230351573A1 (en) * | 2021-03-17 | 2023-11-02 | Southeast University | Intelligent detection method and unmanned surface vehicle for multiple type faults of near-water bridges |
-
2023
- 2023-11-03 CN CN202311453323.1A patent/CN117333845B/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200210826A1 (en) * | 2018-12-29 | 2020-07-02 | Northeastern University | Intelligent analysis system using magnetic flux leakage data in pipeline inner inspection |
US20210370993A1 (en) * | 2020-05-27 | 2021-12-02 | University Of South Carolina | Computer vision based real-time pixel-level railroad track components detection system |
US20230351573A1 (en) * | 2021-03-17 | 2023-11-02 | Southeast University | Intelligent detection method and unmanned surface vehicle for multiple type faults of near-water bridges |
CN114998831A (en) * | 2022-05-23 | 2022-09-02 | 重庆科技学院 | Fire detection method and system based on edge calculation and improved YOLOv3 |
US11790640B1 (en) * | 2022-06-22 | 2023-10-17 | Ludong University | Method for detecting densely occluded fish based on YOLOv5 network |
CN116092115A (en) * | 2022-11-28 | 2023-05-09 | 中国计量大学 | A real-time lightweight detection method for safety clothing of construction workers |
CN116259032A (en) * | 2022-11-30 | 2023-06-13 | 长春工业大学 | Road traffic sign detection and identification algorithm based on improved YOLOv5 |
CN116434188A (en) * | 2023-03-26 | 2023-07-14 | 复旦大学 | Traffic sign detection method based on improved_yolov5s network |
CN116597413A (en) * | 2023-04-28 | 2023-08-15 | 哈尔滨工业大学 | Real-time traffic sign detection method based on improved YOLOv5 |
CN116469021A (en) * | 2023-04-30 | 2023-07-21 | 桂林电子科技大学 | Unmanned aerial vehicle aerial photographing target detection method based on improved YOLOv5 |
CN116630920A (en) * | 2023-05-18 | 2023-08-22 | 天津科技大学 | Lane line type recognition method of improved YOLOv5s network model |
Non-Patent Citations (9)
Title |
---|
PEILIN LIU等: "UCN-YOLOv5: Traffic Sign Object Detection Algorithm Based on Deep Learning", 《 IEEE ACCESS ( VOLUME: 11)》, 5 October 2023 (2023-10-05) * |
WANCHANG JIANG等: "Top Invulnerability Nodes Mining in Dual-Direction Different-Weight Complex Network Based on Node Double-Level Local Structure Weighted Entropy", 《 IEEE ACCESS ( VOLUME: 7)》, 28 June 2019 (2019-06-28) * |
XIA, KW等: "Global contextual attention augmented YOLO with ConvMixer prediction heads for PCB surface defect detection", 《SCIENTIFIC REPORTS》, 16 June 2023 (2023-06-16) * |
姜万昌;杨旸;隋吉生;: "基于小波变换的ARMA-RBF光功率组合预测", 光通信技术, no. 12, 15 December 2013 (2013-12-15) * |
崔家睿: "基于多目标追踪的交通场景异常检测", 《硕士电子期刊出版信息》, 15 July 2023 (2023-07-15) * |
张钟文;高宇;王静;曹登平;: "基于YOLOv3的深度学习交通标志识别系统", 建筑电气, no. 07, 25 July 2020 (2020-07-25) * |
殷浩东: "基于深度学习的行人跟踪算法研究", 《硕士电子期刊出版信息》, 15 March 2022 (2022-03-15) * |
葛明进;孙作雷;孔薇;: "基于anchor-free的交通场景目标检测技术", 计算机工程与科学, no. 04, 15 April 2020 (2020-04-15) * |
郑冬;李向群;许新征;: "基于轻量化SSD的车辆及行人检测网络", 南京师大学报(自然科学版), no. 01, 20 March 2019 (2019-03-20) * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118230079A (en) * | 2024-05-27 | 2024-06-21 | 中国科学院西安光学精密机械研究所 | A detection method for small remote sensing targets based on improved YOLO |
CN118710607A (en) * | 2024-06-18 | 2024-09-27 | 东北电力大学 | A lightweight road surface defect detection method based on improved YOLOv7 |
CN118865174A (en) * | 2024-07-18 | 2024-10-29 | 电子科技大学 | A fast detection method for small moving targets in UAVs |
CN118570740A (en) * | 2024-08-02 | 2024-08-30 | 南京信息工程大学 | Bus stop scene vehicle pedestrian detection method based on improvement YOLOV5 |
CN119810761A (en) * | 2024-12-05 | 2025-04-11 | 东北电力大学 | Unmanned aerial vehicle detection method based on improved YOLO11n |
CN119693923A (en) * | 2025-02-24 | 2025-03-25 | 浙江理工大学 | Path planning method combining small-scale traffic sign recognition and DWA algorithm |
CN119693923B (en) * | 2025-02-24 | 2025-06-24 | 浙江理工大学 | Path planning method combining small-scale traffic sign recognition and DWA algorithm |
Also Published As
Publication number | Publication date |
---|---|
CN117333845B (en) | 2025-02-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN117333845A (en) | Real-time detection method for small target traffic sign based on improved YOLOv5s | |
CN108416307B (en) | An aerial image pavement crack detection method, device and equipment | |
CN112150821B (en) | Method, system and device for constructing lightweight vehicle detection model | |
CN114863348B (en) | Video target segmentation method based on self-supervision | |
CN110660052A (en) | A deep learning-based detection method for surface defects of hot-rolled strip steel | |
CN112149547B (en) | Water Body Recognition Method Based on Image Pyramid Guidance and Pixel Pair Matching | |
CN110443173A (en) | A kind of instance of video dividing method and system based on inter-frame relation | |
CN113420643B (en) | Lightweight underwater target detection method based on depth separable cavity convolution | |
CN110276264A (en) | A Crowd Density Estimation Method Based on Foreground Segmentation Map | |
CN109977895A (en) | A kind of wild animal video object detection method based on multi-characteristic fusion | |
CN113420727B (en) | Training method and device for table detection model, and table detection method and device | |
CN113139969A (en) | Attention mechanism-based weak supervision image semantic segmentation method and system | |
CN111680705A (en) | MB-SSD Method and MB-SSD Feature Extraction Network for Object Detection | |
CN114972753A (en) | A lightweight semantic segmentation method and system based on contextual information aggregation and assisted learning | |
CN113221956A (en) | Target identification method and device based on improved multi-scale depth model | |
CN111291760A (en) | Semantic segmentation method and device for image and electronic equipment | |
CN114360067A (en) | A dynamic gesture recognition method based on deep learning | |
WO2023159527A1 (en) | Detector training method and apparatus, and storage medium | |
CN114758255A (en) | Unmanned aerial vehicle detection method based on YOLOV5 algorithm | |
CN117975418A (en) | Traffic sign detection method based on improved RT-DETR | |
CN114298187A (en) | Target detection algorithm integrating improved attention mechanism | |
CN110263731A (en) | A kind of single step face detection system | |
CN113033444A (en) | Age estimation method and device and electronic equipment | |
CN113673534A (en) | RGB-D image fruit detection method based on fast RCNN | |
CN112070080A (en) | Method for classifying cartoon characters playing songs based on Faster R-CNN |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |