CN112215071A - Vehicle-mounted multi-target coupling identification and tracking method for automatic driving under heterogeneous traffic flow - Google Patents

Vehicle-mounted multi-target coupling identification and tracking method for automatic driving under heterogeneous traffic flow Download PDF

Info

Publication number
CN112215071A
CN112215071A CN202010948875.XA CN202010948875A CN112215071A CN 112215071 A CN112215071 A CN 112215071A CN 202010948875 A CN202010948875 A CN 202010948875A CN 112215071 A CN112215071 A CN 112215071A
Authority
CN
China
Prior art keywords
model
deep learning
vehicle
training
establishing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010948875.XA
Other languages
Chinese (zh)
Inventor
万千
谢振友
彭国庆
林初染
龙朝党
陆盛康
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hualan Design Group Co ltd
Guilin University of Electronic Technology
Original Assignee
Hualan Design Group Co ltd
Guilin University of Electronic Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hualan Design Group Co ltd, Guilin University of Electronic Technology filed Critical Hualan Design Group Co ltd
Priority to CN202010948875.XA priority Critical patent/CN112215071A/en
Publication of CN112215071A publication Critical patent/CN112215071A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Traffic Control Systems (AREA)

Abstract

The invention relates to an automatic driving vehicle-mounted multi-target coupling identification and tracking method under heterogeneous traffic flow, which comprises the steps of establishing a vehicle-mounted embedded application program resource scheduler, wherein the identification and tracking method comprises the steps of carrying out model pruning and restoration processing on a driving scene identification model and establishing a resource distribution framework for supporting deep learning model application program dynamics; and establishing a deep learning model operation resource allocation scheduler, flexibly allocating resources for the deep learning models operated concurrently through the resource scheduler, and outputting an optimized scheduling scheme. The invention can reduce the memory occupation and the switching energy consumption of the deep learning model on the mobile visual equipment, provide flexible resource allocation and accuracy rate balance, reduce the processing delay, improve the multi-target identification efficiency of the automatic driving automobile, ensure that the processing of the automobile road cooperation is more timely and accurate when the automobile is automatically driven, and further improve the safety of the automatic driving automobile and the application prospect in the traffic field.

Description

Vehicle-mounted multi-target coupling identification and tracking method for automatic driving under heterogeneous traffic flow
Technical Field
The invention relates to the fields of computer vision technology and intelligent traffic, in particular to an automatic driving vehicle-mounted multi-target coupling identification and tracking method under heterogeneous traffic flow.
Background
With the development of deep learning technology, various artificial intelligence applications are attracting strong attention. The automatic driving is an important field for realizing intelligent traffic and establishing a strong traffic country in the future, and the attention of people is more and more attracted, the safety is always a key consideration of the automatic driving technology, and the automatic driving under the cooperation of the vehicle and the road is very important for improving the safety of the automatic driving. The visual perception is one of key technologies in automatic driving, and a camera and a sensor acquire data from an environment outside a vehicle and transmit the data to a processor; the recognition algorithm in the processor can recognize driving scene targets such as people, vehicles, roads, traffic sign and marking lines; the recognition algorithm is based on sensing and understanding of the environment, and the automatic driving vehicle can safely drive on the road. The application of automatic driving can comprehensively improve the driving safety and comfort, and has great market demand.
Although the deep learning algorithm has a good calculation rate at the PC end and the server, most of the current embedded mobile devices are far less computationally intensive than the server, so that many of the well-performing deep learning network models cannot be deployed on the embedded mobile devices. However, real-time visual recognition is required in an autonomous automobile, cloud support is not available, and computing resources of an embedded mobile device are limited. When it comes to compressing deep learning models reduces the resource requirements but reduces the accuracy and the resource budget of the compressed model is fixed, i.e. static. Then, on one hand, the deep learning model pays attention to the resource requirement, the resources required by the vehicle-mounted scene recognition system during operation are dynamically changed, and when the application programs are parallel and reach the maximum available resources, the application programs mutually compete for the resources, so that the frame frequency of streaming media video processing is lowered; on the other hand, if the mobile vision system has additional resources to run, the compression model cannot recover its reduced accuracy using the additional resources. Under a mixed traffic scene, a plurality of targets need to be identified, cameras and sensors of an automatic driving vehicle are distributed, the front, the back, the left and the right of the vehicle and the roof of the vehicle, streaming video data transmitted by each camera needs to be processed simultaneously, a mobile vision system needs to identify and process in real time, and under the limited computing resources of an embedded mobile device at a vehicle-mounted end, the resource requirement of a deep learning network model with excellent western performance cannot be well met, so that the accuracy of the automatic driving vehicle on scene identification is greatly reduced, the processing delay is increased, and the accuracy of the targets is affected and the real-time performance cannot be achieved.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides an automatic driving vehicle-mounted multi-target coupling identification and tracking method under heterogeneous traffic flow, which aims at embedded mobile equipment such as an automatic driving automobile and the like supporting a deep learning model, integrates and optimizes the existing algorithm framework, improves the resource distribution efficiency of the deep learning model at an embedded mobile terminal, realizes real-time visual processing of the automatic driving vehicle through equipment such as a camera and the like under a mixed traffic scene, improves the accuracy of the targets such as people, vehicles, roads and the like, reduces processing delay, and has important significance for improving driving safety and improving the efficiency of a traffic system.
The purpose of the invention is realized by the following technical scheme: the identification and tracking method comprises the steps of establishing a vehicle-mounted embedded application program resource scheduler, and comprises the following steps:
performing model pruning and restoration processing on the driving scene recognition model, and establishing a resource allocation framework supporting the dynamic state of the deep learning model application program;
and establishing a deep learning model operation resource allocation scheduler, flexibly allocating resources for the deep learning models operated concurrently through the resource scheduler, and outputting an optimized scheduling scheme.
Further, the identification and tracking method also comprises the step of establishing and training a driving scene identification model before establishing the vehicle-mounted embedded application program resource scheduler; the step of establishing and training a driving scene recognition model comprises the following steps: and establishing a driving scene sample data set and a recognition tracking model for training multiple targets in a driving scene.
Further, the performing model pruning and restoration processing on the driving scene recognition model and establishing a resource allocation framework supporting deep learning model application dynamics includes:
pruning neurons of the deep learning model, and compressing the deep learning model;
restoring the deep learning model to generate a multi-capacity model;
and analyzing the given vehicle-mounted traffic scene recognition system model, selecting the optimal resource precision balance according to the requirement of each traffic scene recognition model, and optimizing the inference accuracy, the memory occupation and the processing delay of each derived model.
Further, the pruning neurons of the deep learning model, and compressing the deep learning model includes:
setting the matrix of each layer of the convolution layer as | Mij|,MijRepresenting a specific value in each convolution kernel, summing the absolute values of the weight values of all convolution kernels, i.e.
Figure BDA0002676237410000021
mi is the channel number of the convolution kernel;
according to importance to SiSorting, setting a minimum threshold and comparing S below the minimum thresholdiPruning is carried out, and the feature graph of the convolution kernel of the next layer corresponding to the current layer is deleted;
and (3) creating a new matrix and generating new weights for the next layer corresponding to the pruned convolutional layer, integrating the model structure to complete pruning, retraining the pruned model to complete model compression, and taking the minimum pruning model as a seed model.
Further, the restoring the deep learning model and the generating the multi-capacity model include:
iteration is carried out by using the seed model to carry out model solidification on the filter parameters, and the existing model parameter data are saved;
according to the record of the neuron pruning path during model compression, on the basis of the reverse pruning direction, applying filter growth to supplement the pruned filter to increase the capacity of the model, and then increasing the accuracy of the model through training;
and continuously generating a new derivative model by repeatedly iterating the previous model, and finally generating a derivative model containing all the parameter capacities of the previous model to obtain the multi-capacity model.
Further, the establishing of the deep learning model and the running of the resource allocation scheduler flexibly allocate resources for the deep learning model running concurrently through the resource scheduler, and the outputting of the optimized scheduling scheme includes:
by setting a cost function in the vehicle-mounted driving scene recognition system
Figure BDA0002676237410000031
Figure BDA0002676237410000032
Analyzing all concurrently running application programs, and selecting an optimal derivative model for each application program;
the method comprises the steps of designing a scheduling scheme for the deep learning model in the parallel driving scene, minimizing the total cost of the application programs of the deep learning model running concurrently, carrying out optimization constraint on a resource perception scheduler and reasonably distributing the running resources of the parallel application programs so as to balance the running performance.
Further, the training of the recognition tracking model of multiple targets in the driving scene comprises:
training a driving scene multi-target recognition model based on a YOLOv3 algorithm by using a Dartnet deep learning framework, and importing a training set and a verification set;
setting training parameters, setting a learning weight parameter to be 0.0005, setting a momentum to be 0.9, setting a batch size to be 64, setting a learning rate to be 0.001 initially, and reducing the learning rate to be 0.0001 and 0.00001 in sequence, wherein the network iterates 25000, 20000 and 15000 times respectively corresponding to the learning rate of each stage, and the iteration is carried out for 60000 times in total;
the loss function adopts CIOU loss, when the descending gradient of the loss function tends to be smooth, the model training is finished, and the model parameters of the optimal training result are stored;
importing a Deepsort algorithm model into the trained model, modifying classes into the number of model identification classes 16, and setting filters as 3 x (5+ len (classes));
and training the model of YOLOv3 combined with Deepsort, and storing the model after finishing.
The invention has the following advantages: model pruning compression, restoration and analysis are carried out on the visual recognition deep learning model based on a NestDNN framework, and a resource scheduler is built so as to ensure the correctness and the effectiveness of compression of the visual recognition deep learning model and provide technical support for embedded vehicle-mounted equipment for automatic driving. Experimental results show that the method for recognizing and tracking the vehicle-mounted multi-target coupling of the automatic driving under the heterogeneous traffic flow can reduce the memory occupation and switching energy consumption of a deep learning model on mobile visual equipment, provide flexible balance of resource allocation and accuracy, reduce processing delay, improve the efficiency of multi-target recognition of the automatic driving vehicle, enable the vehicle to be more timely and accurate in the aspect of processing of vehicle-road cooperation during automatic driving, and further improve the safety of the automatic driving vehicle and the application prospect in the traffic field. .
Drawings
FIG. 1 is a flow chart of an embodiment of the present invention;
FIG. 2 is a graph of model training process loss variation;
FIG. 3 is a schematic view of the model before and after trimming;
FIG. 4 is a schematic diagram of pruning compression for a deep learning model network;
FIG. 5 is a graph of model accuracy effect after using the architectural approach;
FIG. 6 is a graph illustrating the effect of model processing time after the architectural approach is used.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present application clearer, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all the embodiments. The components of the embodiments of the present application, generally described and illustrated in the figures herein, can be arranged and designed in a wide variety of different configurations. Thus, the following detailed description of the embodiments of the present application, as presented in the figures, is not intended to limit the scope of the claimed application, but is merely representative of selected embodiments of the application. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present application without making any creative effort, shall fall within the protection scope of the present application. The invention is further described below with reference to the accompanying drawings.
As shown in fig. 1, the invention relates to a vehicle-mounted multi-target coupling identification and tracking method for automatic driving under heterogeneous traffic flow, which relates to the field of automatic driving automobiles with computer vision, and specifically comprises the following steps:
s1, training a driving scene recognition model:
s11, collecting a data set of a driving scene, and in order to explain implementation cases conveniently, in the embodiment, a German computer vision algorithm is used for evaluating the data set Kitti, and 9000, 1000 and 1000 images are selected as a training set, a verification set and a test set respectively.
S12, training a multi-target recognition tracking model in a driving scene, specifically comprising:
s121, training a driving scene multi-target recognition model based on a YOLOv3 algorithm by using a Dartnet deep learning framework, and importing a training set and a verification set;
s122, set training parameters, set Learning weight parameter (decay) to 0.0005, momentum to 0.9, Batch size (Batch) to 64, Learning rate (Learning _ rate) to 0.001 initially, and decrease to 0.0001 and 0.00001 in sequence. Corresponding to the learning rate of each stage, the network respectively iterates 25000, 20000 and 15000 times, and 60000 iterations are carried out in total;
and S123, adopting CIOU loss as a loss function. The loss function is formulated as:
Figure BDA0002676237410000041
wherein b, bgtRepresenting a prediction box B and a real box BgtC is a diagonal line of a minimum box containing the prediction box and the real box;
wherein
Figure BDA0002676237410000042
Is a weight function; wherein
Figure BDA0002676237410000043
Is a parameter for measuring the similarity of aspect ratio, particularly in the length and width of [0,1 ]]In the case of (a) in (b),
Figure BDA0002676237410000044
taking the value 1.
As shown in fig. 2, YOLOv3 model training ends when the loss function descent gradient tends to level off.
S124, after the training is finished, storing the model parameter frame of the optimal training result to obtain a trained YOLOv3 model;
s125, copying the trained model weights.h5 into a model _ data folder, and copying the deep _ sort _ Yolov3 model; generating a mars-small128.pb file and placing the mars-small pb file into a model _ data folder of a deep _ sort _ Yolov3 model; modifying classes to the number of model identification categories 16, setting filters to 3 × (5+ len (classes));
s126, training the model of YOLOv3 combined with Deepsort, and storing the model after finishing training.
S2, establishing a vehicle-mounted embedded application program resource scheduler:
s21, constructing a NestDNN resource scheduling framework;
s211, performing model compression on the driving scene recognition model;
as shown in fig. 3 and 4, for the neurons with smaller importance in the trained deep learning model, the convolutional layers are trimmed layer by layer and retrained again to complete model compression; and completing the model compression. And the number of parameters is reduced by removing unimportant parameters in the network, and the model compression is completed.
Specifically, the matrix for each layer of the convolutional layer is set to be | Mij|,MijRepresenting a specific value in each convolution kernel, summing the absolute values of the weight values of all convolution kernels, i.e.
Figure BDA0002676237410000051
mi is the channel number of the convolution kernel; according to importance to SiSorting, setting a minimum threshold and comparing S below the minimum thresholdiPruning is carried out, and the convolution kernel characteristic graph of the next layer corresponding to the current layer is deleted; creating a new matrix and generating new weights for the pruned convolutional layers and the corresponding next layer, and integrating the model structure to complete pruning; retraining the trimmed model, and repeating the training to trim the whole deep learning model network to obtain a seed model and complete model compression;
s212, performing model restoration on the driving scene recognition model;
iteration is carried out by using a seed model, model solidification is carried out on filter parameters, existing model parameter data are stored according to a neuron trimming path record during model compression, a trimmed filter is added back by applying filter growth based on a reverse trimming direction, and the model accuracy is increased by retraining; the new model generated based on the seed model nesting can share model parameters, and through repeated iteration, the previous model generates a new derivative model until the iteration is finished, and a derivative model containing all the previous model parameter capacities is generated, so that a multi-capacity model is obtained;
and S213, performing model analysis on the driving scene recognition model, taking the video acquired by each camera of the vehicle-mounted end as a deep learning model for traffic scene recognition, selecting the optimal resource precision balance according to the requirement of each traffic scene recognition model, and optimizing and deducing the accuracy, the memory occupation and the processing delay of each derivative model.
S22, establishing a driving scene recognition resource perception scheduler;
s221, establishing a deep learning model operation resource allocation scheduler, and setting a cost function C for the vehicle-mounted driving scene recognition application program, wherein the formula is as follows:
Figure BDA0002676237410000052
wherein V is an application program of the vehicle-mounted scene recognition system, V is a set of application programs which run in the vehicle-mounted traffic light recognition system concurrently, and V belongs to V. m isvFor derived models, uv∈[0,1]For the proportion of computing resources allocated to v,
Figure BDA0002676237410000061
processing delay allocated to v for all computing resources, Amin(v) For minimum processing precision, Lmax(v) Is 0 max processing delay target.
The reasoning accuracy and the processing delay of each scene recognition application program are converted into a cost function, and a resource scheduling scheme basis is provided.
S222, establishing a deep learning model operation resource scheduling scheme, setting the lowest inference accuracy and processing delay according to different recognition scenes and targets, designing the parallel scene recognition application program scheduling scheme according to the set targets, minimizing the total cost of the application program of the concurrently operated deep learning model, and optimally restricting and distributing the operation resources of the parallel application program by the resource perception scheduler so as to balance the operation performance. The optimization constraint formula is as follows:
Figure BDA0002676237410000062
Figure BDA0002676237410000063
as shown in fig. 5 and 6, the method provided by the invention shows effectiveness by using ResNet _50, inclusion _ v3, VGG _16, MobileNet and NMT models to a vehicle-mounted end scene recognition experiment of the deep learning model based on the NestDNN framework. Specifically, the method comprises the following steps: when a NestDNN framework is used for establishing a resource-aware running scheduler to process a deep learning model, compared with a deep model without considering resource optimization allocation, the model performance changes remarkably: as shown in the following table, the model memory occupation and the model switching overhead are reduced by 41.7%; the accuracy is improved by 1.39%, and the processing time is improved by 155%.
TABLE 1 comparison of model memory occupancy and model switching overhead effect graphs
Model (model) Multi-volume model size (MB) Original model size (MB) Reducing memory occupation (MB)
ResNet_50 22.3 33.8 11.5
Inception_v3 81.4 234.5 153.1
VGG_16 127.0 261.2 134.2
MobileNet 6.7 9.7 3.0
NMT 293.0 532.7 139.7
Total of 530.4 1071.9 441.5
The foregoing is illustrative of the preferred embodiments of this invention, and it is to be understood that the invention is not limited to the precise form disclosed herein and that various other combinations, modifications, and environments may be resorted to, falling within the scope of the concept as disclosed herein, either as described above or as apparent to those skilled in the relevant art. And that modifications and variations may be effected by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (7)

1. The method for automatically driving the vehicle-mounted multi-target coupling recognition and tracking under heterogeneous traffic flows is characterized by comprising the following steps: the identification and tracking method comprises the steps of establishing a vehicle-mounted embedded application program resource scheduler, wherein the steps comprise:
performing model pruning and restoration processing on the driving scene recognition model, and establishing a resource allocation framework supporting the dynamic state of the deep learning model application program;
and establishing a deep learning model operation resource allocation scheduler, flexibly allocating resources for the deep learning models operated concurrently through the resource scheduler, and outputting an optimized scheduling scheme.
2. The method for automatically driving the vehicle-mounted multi-target coupling to recognize and track under the heterogeneous traffic flow according to claim 1, wherein: the identification and tracking method further comprises the step of establishing and training a driving scene identification model before establishing the vehicle-mounted embedded application program resource scheduler; the step of establishing and training a driving scene recognition model comprises the following steps: and establishing a driving scene sample data set and a recognition tracking model for training multiple targets in a driving scene.
3. The method for automatically driving the vehicle-mounted multi-target coupling to recognize and track under the heterogeneous traffic flow according to claim 1, wherein: the model pruning and restoration processing of the driving scene recognition model, and the establishment of a resource allocation framework supporting the deep learning model application program dynamics comprise:
pruning neurons of the deep learning model, and compressing the deep learning model;
restoring the deep learning model to generate a multi-capacity model;
and analyzing the given vehicle-mounted traffic scene recognition system model, selecting the optimal resource precision balance according to the requirement of each traffic scene recognition model, and optimizing the inference accuracy, the memory occupation and the processing delay of each derived model.
4. The method for automatically driving the vehicle-mounted multi-target coupling to recognize and track under the heterogeneous traffic flow according to claim 3, wherein: the method for pruning the neurons of the deep learning model, wherein compressing the deep learning model comprises:
setting the matrix of each layer of the convolution layer as | MijI, Mij represents a specific number in each convolution kernel, and the absolute values of the weight values of all convolution kernels are summed, i.e.
Figure FDA0002676237400000011
mi is the channel number of the convolution kernel;
according to importance to SiSorting, setting a minimum threshold and comparing S below the minimum thresholdiPruning is carried out, and the feature graph of the convolution kernel of the next layer corresponding to the current layer is deleted;
and (3) creating a new matrix and generating new weights for the next layer corresponding to the pruned convolutional layer, integrating the model structure to complete pruning, retraining the pruned model to complete model compression, and taking the minimum pruning model as a seed model.
5. The method for automatically driving the vehicle-mounted multi-target coupling to recognize and track under the heterogeneous traffic flow according to claim 3, wherein: the restoring the deep learning model and the generating the multi-capacity model comprise:
iteration is carried out by using the seed model to carry out model solidification on the filter parameters, and the existing model parameter data are saved;
according to the record of the neuron pruning path during model compression, on the basis of the reverse pruning direction, applying filter growth to supplement the pruned filter to increase the capacity of the model, and then increasing the accuracy of the model through training;
and continuously generating a new derivative model by repeatedly iterating the previous model, and finally generating a derivative model containing all the parameter capacities of the previous model to obtain the multi-capacity model.
6. The method for automatically driving the vehicle-mounted multi-target coupling to recognize and track under the heterogeneous traffic flow according to claim 1, wherein: the establishing of the deep learning model and the running of the resource allocation scheduler flexibly allocate resources for the deep learning model running concurrently through the resource scheduler, and the outputting of the optimized scheduling scheme comprises the following steps:
by setting a cost function in the vehicle-mounted driving scene recognition system
Figure FDA0002676237400000021
Figure FDA0002676237400000022
Analyzing all concurrently running application programs, and selecting an optimal derivative model for each application program;
the method comprises the steps of designing a scheduling scheme for the deep learning model in the parallel driving scene, minimizing the total cost of the application programs of the deep learning model running concurrently, carrying out optimization constraint on a resource perception scheduler and reasonably distributing the running resources of the parallel application programs so as to balance the running performance.
7. The method for automatically driving the vehicle-mounted multi-target coupling to recognize and track under the heterogeneous traffic flow according to claim 2, wherein: the identification tracking model for multiple targets in the training driving scene comprises the following steps:
training a driving scene multi-target recognition model based on a YOLOv3 algorithm by using a Dartnet deep learning framework, and importing a training set and a verification set;
setting training parameters, setting a learning weight parameter to be 0.0005, setting a momentum to be 0.9, setting a batch size to be 64, setting a learning rate to be 0.001 initially, and reducing the learning rate to be 0.0001 and 0.00001 in sequence, wherein the network iterates 25000, 20000 and 15000 times respectively corresponding to the learning rate of each stage, and the iteration is carried out for 60000 times in total;
the loss function adopts CIOU loss, when the descending gradient of the loss function tends to be smooth, the model training is finished, and the model parameters of the optimal training result are stored;
importing a Deepsort algorithm model into the trained model, modifying classes into the number of model identification classes 16, and setting filters as 3 x (5+ len (classes));
and training the model of YOLOv3 combined with Deepsort, and storing the model after finishing.
CN202010948875.XA 2020-09-10 2020-09-10 Vehicle-mounted multi-target coupling identification and tracking method for automatic driving under heterogeneous traffic flow Pending CN112215071A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010948875.XA CN112215071A (en) 2020-09-10 2020-09-10 Vehicle-mounted multi-target coupling identification and tracking method for automatic driving under heterogeneous traffic flow

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010948875.XA CN112215071A (en) 2020-09-10 2020-09-10 Vehicle-mounted multi-target coupling identification and tracking method for automatic driving under heterogeneous traffic flow

Publications (1)

Publication Number Publication Date
CN112215071A true CN112215071A (en) 2021-01-12

Family

ID=74049951

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010948875.XA Pending CN112215071A (en) 2020-09-10 2020-09-10 Vehicle-mounted multi-target coupling identification and tracking method for automatic driving under heterogeneous traffic flow

Country Status (1)

Country Link
CN (1) CN112215071A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112784748A (en) * 2021-01-22 2021-05-11 大连海事大学 Microalgae identification method based on improved YOLOv3
CN113470070A (en) * 2021-06-24 2021-10-01 国汽(北京)智能网联汽车研究院有限公司 Driving scene target tracking method, device, equipment and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190034235A1 (en) * 2017-12-28 2019-01-31 Shao-Wen Yang Privacy-preserving distributed visual data processing
CN109523017A (en) * 2018-11-27 2019-03-26 广州市百果园信息技术有限公司 Compression method, device, equipment and the storage medium of deep neural network
CN109934844A (en) * 2019-01-28 2019-06-25 中国人民解放军战略支援部队信息工程大学 A kind of multi-object tracking method and system merging geospatial information
WO2019217129A1 (en) * 2018-05-10 2019-11-14 Microsoft Technology Licensing, Llc Efficient data encoding for deep neural network training
CN111191772A (en) * 2020-01-02 2020-05-22 中国航空工业集团公司西安航空计算技术研究所 Intelligent computing general acceleration system facing embedded environment and construction method thereof
CN111476756A (en) * 2020-03-09 2020-07-31 重庆大学 Method for identifying casting DR image loose defects based on improved YO L Ov3 network model
CN111554105A (en) * 2020-05-29 2020-08-18 浙江科技学院 Intelligent traffic identification and statistics method for complex traffic intersection

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190034235A1 (en) * 2017-12-28 2019-01-31 Shao-Wen Yang Privacy-preserving distributed visual data processing
WO2019217129A1 (en) * 2018-05-10 2019-11-14 Microsoft Technology Licensing, Llc Efficient data encoding for deep neural network training
CN109523017A (en) * 2018-11-27 2019-03-26 广州市百果园信息技术有限公司 Compression method, device, equipment and the storage medium of deep neural network
CN109934844A (en) * 2019-01-28 2019-06-25 中国人民解放军战略支援部队信息工程大学 A kind of multi-object tracking method and system merging geospatial information
CN111191772A (en) * 2020-01-02 2020-05-22 中国航空工业集团公司西安航空计算技术研究所 Intelligent computing general acceleration system facing embedded environment and construction method thereof
CN111476756A (en) * 2020-03-09 2020-07-31 重庆大学 Method for identifying casting DR image loose defects based on improved YO L Ov3 network model
CN111554105A (en) * 2020-05-29 2020-08-18 浙江科技学院 Intelligent traffic identification and statistics method for complex traffic intersection

Non-Patent Citations (7)

* Cited by examiner, † Cited by third party
Title
AIHGF: "Github项目- 基于YOLOV3 和 DeepSort 的实时多人追踪", 《HTTPS://CLOUD.TENCENT.COM/DEVELOPER/ARTICLE/1396326》 *
BIYI FANG 等: "NestDNN: Resource-Aware Multi-Tenant On-Device Deep Learning for Continuous Mobile Vision", 《MOBILE COMPUTING AND NETWORKING》 *
HAO LI 等: "PRUNING FILTERS FOR EFFICIENT CONVNETS", 《ARXIV》 *
朱金凤: "基于YOLO-v3的高速公路交通事件检测系统研究", 《万方数据》 *
王磊 等: "面向嵌入式应用的深度神经网络模型压缩技术综述", 《北京交通大学学报》 *
秦严严 等: "混有协同自适应巡航控制车辆的异质交通流稳定性解析与基本图模型", 《物理学报》 *
鸡啄米的时光机: "使用YoloV3 + deep_sort实现多人目标跟踪", 《HTTPS://BLOG.CSDN.NET/QQ_33221533/ARTICLE/DETAILS/100849251》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112784748A (en) * 2021-01-22 2021-05-11 大连海事大学 Microalgae identification method based on improved YOLOv3
CN113470070A (en) * 2021-06-24 2021-10-01 国汽(北京)智能网联汽车研究院有限公司 Driving scene target tracking method, device, equipment and storage medium

Similar Documents

Publication Publication Date Title
CN112163465B (en) Fine-grained image classification method, fine-grained image classification system, computer equipment and storage medium
CN110321910B (en) Point cloud-oriented feature extraction method, device and equipment
CN109389244B (en) GRU-based multi-factor perception short-term scenic spot visitor number prediction method
WO2020042658A1 (en) Data processing method, device, apparatus, and system
CN112445823A (en) Searching method of neural network structure, image processing method and device
CN110942637B (en) SCATS system road traffic flow prediction method based on airspace map convolutional neural network
CN110956202B (en) Image training method, system, medium and intelligent device based on distributed learning
CN112215071A (en) Vehicle-mounted multi-target coupling identification and tracking method for automatic driving under heterogeneous traffic flow
US20230131518A1 (en) Model Generation Method and Apparatus, Object Detection Method and Apparatus, Device, and Storage Medium
CN117157678A (en) Method and system for graph-based panorama segmentation
WO2021174370A1 (en) Method and system for splitting and bit-width assignment of deep learning models for inference on distributed systems
CN113283404B (en) Pedestrian attribute identification method and device, electronic equipment and storage medium
CN112766578A (en) Vehicle use identification method and system based on vehicle network and storage medium
CN109982088B (en) Image processing method and device
CN112598062A (en) Image identification method and device
CN115272894A (en) Unmanned aerial vehicle-oriented image target detection method and device, electronic equipment and storage medium
CN115018039A (en) Neural network distillation method, target detection method and device
CN117056785A (en) Federal learning classification model training method based on self-adaptive model disturbance
CN113420651B (en) Light weight method, system and target detection method for deep convolutional neural network
CN112819157B (en) Neural network training method and device, intelligent driving control method and device
CN111932690B (en) Pruning method and device based on 3D point cloud neural network model
CN112132207A (en) Target detection neural network construction method based on multi-branch feature mapping
KR102143073B1 (en) Smart cctv apparatus for analysis of parking
EP4032028A1 (en) Efficient inferencing with fast pointwise convolution
CN111931768A (en) Vehicle identification method and system capable of self-adapting to sample distribution

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210112

RJ01 Rejection of invention patent application after publication