CN109993082A - The classification of convolutional neural networks road scene and lane segmentation method - Google Patents

The classification of convolutional neural networks road scene and lane segmentation method Download PDF

Info

Publication number
CN109993082A
CN109993082A CN201910211802.XA CN201910211802A CN109993082A CN 109993082 A CN109993082 A CN 109993082A CN 201910211802 A CN201910211802 A CN 201910211802A CN 109993082 A CN109993082 A CN 109993082A
Authority
CN
China
Prior art keywords
classification
road
neural networks
convolutional neural
scene
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910211802.XA
Other languages
Chinese (zh)
Other versions
CN109993082B (en
Inventor
应捷
刘逸
李建军
陈明玺
胡文凯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Shanghai for Science and Technology
Original Assignee
University of Shanghai for Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Shanghai for Science and Technology filed Critical University of Shanghai for Science and Technology
Priority to CN201910211802.XA priority Critical patent/CN109993082B/en
Publication of CN109993082A publication Critical patent/CN109993082A/en
Application granted granted Critical
Publication of CN109993082B publication Critical patent/CN109993082B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/588Recognition of the road, e.g. of lane markings; Recognition of the vehicle driving pattern in relation to the road

Abstract

The present invention relates to a kind of classification of convolutional neural networks road scene and lane segmentation method, establish double task cooperative structural models based on convolutional neural networks, it is made of an encoder, two decoders, pass through end-to-end training, realize that characteristic information is shared, the classification to the various roads scene such as urban road, backroad and highway is completed, and region can travel to the road in scene and be split, reaches the double goal of road scene classification and road area segmentation.Based on convolutional neural networks, by road scene perceive in scene classification in conjunction with the double task of road extraction, exported in real time by training mode end to end, be beneficial to improve intelligent driving auxiliary system function, effective reliable drive assistance function is provided.Wherein in model to the part of travelable extracted region using full convolutional network in such a way that condition random field combines, can generate high-definition picture optimization output as a result, reaching more accurate segmentation effect.

Description

The classification of convolutional neural networks road scene and lane segmentation method
Technical field
The present invention relates to a kind of image recognition technology, in particular to a kind of convolutional neural networks road scene classification and road Dividing method.
Background technique
The perception of road environment is the key technology in vehicle assistant drive technology, and the analysis and understanding of road scene are vehicles Carry the important component of intelligence system.In terms of automatic Pilot, the target of road scene identification is can to obtain road automatically The high-layer semantic information of image judges scene type belonging to image, and mentions to the travelable region in image It takes.Such as under the complex conditions background such as sleet, haze, to street, country road and expressway, these road scenes are divided Class identification makes the early warnings such as speed adjustment to the vehicle of traveling, it will help realize unmanned.In automobile assistant driving System aspects, carrying out profound semantic understanding to road scene facilitates navigation and the intelligent early-warning of automobile, uses tape label Road scene image can more describe scene locating for driver, ride safety of automobile is played a crucial role.
Realize scene classification research direction mainly include the following types: the new feature extraction of scene image, based on high level Semantic scene classification and the scene classification based on deep learning.When scene type reaches thousand classes or more, and database volume is prominent When breaking million images, traditional method based on low-level image feature and high-level semantic is difficult to handle mass data.And use depth The method of study then has good performance in this big data, and therefore, the present invention proposes the road field based on deep learning Scape perception and lane segmentation method.
In terms of deep learning road scene sort research, application No. is 201710894156.2 patents to propose one kind Driving Scene classification method based on convolutional neural networks.Using traditional convolution with feature extraction layer and Feature Mapping layer Neural network carries out feature extraction and training to image by convolution sum sub-sampling twice, pixel value is rasterized and connected into Output of one vector as network generates scene classifier, realizes the classification to variety classes Driving Scene.
Application No. is 201711174595.2 patents to propose a kind of visual scene identification side based on convolutional neural networks Method arbitrarily grabs two from the Same Scene of initial data respectively and finds similar image from different scenes, as scene The mode that perception variation and perception offset images pair occurs constructs test set, with convolutional neural networks to the feature of each image pair Vector makes the difference and is standardized, and obtains feature difference figure, when scene appearance acute variation, can obtain higher accuracy rate.
Application No. is 201610261849.3 patents to propose a kind of traffic scene classification based on convolutional neural networks Method, the original cut tree phase of characteristic information and image that the RGB-D image of traffic scene is extracted by convolutional neural networks In conjunction with forming optimal purity cost tree, and carry out the covering of optimal purity, obtain clearly objective contour, increase the standard of classification True property.
The patent of application number 201510080035.5 proposes a kind of deconvolution depth e-learning based on Weight Scene recognition method obtains the feature on road scene image different scale using the deconvolution depth network model of Weight Figure samples to characteristic pattern through spatial pyramid model, forms the feature expression of image, finally complete using supporting vector classifier Constituent class training and test, obtain the recognition result of different scenes.
Summary of the invention
The problem of the present invention be directed to the identifications of multiclass complexity road scene, proposes a kind of convolutional neural networks road scene Classify and lane segmentation method, establish double task cooperative structural models based on convolutional neural networks in method, is compiled by one Code device, two decoders composition realize that characteristic information is shared by end-to-end training, complete to urban road, backroad and The classification of the various roads scene such as highway, and region can travel to the road in scene and be split, reach road scene The double goal of classification and road area segmentation.
The technical solution of the present invention is as follows: a kind of convolutional neural networks road scene classification and lane segmentation method, specific to wrap Include following steps:
1) image in database is selected, randomly assigne is carried out to selection image, a part is used as training set image, another portion It is allocated as test set image;
2) training set image is sent into double task cooperative structure convolutional neural networks road scene classification and lane segmentation mould Type carries out convolutional neural networks training, convolutional neural networks after being trained;
Double task cooperative structure convolutional neural networks road scene classification are mentioned with lane segmentation model comprising input layer, feature Layer, information processing layer and output layer are taken, training set image is unified for an equal amount of RGB image and is sent into input layer, and input layer is defeated Image is sent into the feature coding device that feature extraction layer is generated by convolutional neural networks out, extracts lane segmentation in feature coding device The characteristics of image needed with scene classification generates characteristic pattern;Information processing layer is by decoder two solutions of classification decoder and segmentation Code device composition, the characteristic pattern of feature extraction layer output, which respectively enters two decoders, which carries out scene type judgement and road, can travel Extracted region;The result of two decoder outputs is sent into output layer, and through joint reasoning, joint reasoning refers to sorting code number device and segmentation Encoder shares identical characteristic information, and model only trains a weight network in training, and model is carrying out backpropagation ginseng When number adjustment, the network optimization is carried out by the way of superposition sorting code number device and the loss of partition encoding device, obtains training the road Hou Dui The convolutional neural networks of road scene perception, achievable scene classification determines and the task of lane segmentation;
3) convolutional neural networks identify after test set image being sent into step 2) training, and achievable scene classification determines With the task of lane segmentation.
Step 2) the feature coding device uses convolutional neural networks VGG-16 as model;Decoder of classifying uses convolution The full articulamentum of neural network VGG-16 generates one-dimensional characteristic vector, is then classified using Softmax classifier, completes more The task of class road scene classification;Divide decoder using full convolutional network in such a way that condition random field combines, completion road Divide task in road;Whole network is using training method end to end, and model only trains a weight network, in the reversed of inclusion layer In communication process by superposition classification decoder and divide decoder loss carry out the network optimization, be completed at the same time scene classification and The task of lane segmentation.
The classification performance of the step 2) classification decoder by calculate accuracy rate Precision and recall rate Recall this Two indexs are measured, and the formula difference of accuracy rate and recall rate is as follows:
Wherein, TP indicates that the sample size being correctly detecting, FP indicate that the correct sample being not detected, FN indicate wrong The sample that erroneous detection measures.
Step 2) the segmentation decoder is exactly the evaluation to Pixel-level classification, road point to the evaluation of segmentation result in fact The problem of cutting can be considered as two classification problems, and any pixel belongs to road surface or belongs to background;Divide decoder and carries out picture It when element classification, predicts that the classification results of output and mark image carry out Pixel-level comparison, evaluates the prediction output knot of each pixel Fruit, if it is " positive class " that pixel, which is judged to road surface pixel, it is " negative class " that pixel, which is judged to background pixel point, then the pixel classifications As a result true positives TP is taken1, false positive FP1, true negative TN1, false negative FN1One of four kinds of results;Wherein TP1Indicate that positive class is sentenced It is set to positive class, FP1Indicate that negative class determines the class that is positive, TN1Indicate that negative class determines the class that is negative, FN1Indicate that positive class determines the class that is negative.Into And can in the hope of the accurate rate TPR and recall rate FPR of each image on entire test set,
Segmentation effect is evaluated using the MaxF1 to classify based on pixel, F-Measure is that information retrieval field is common Evaluation criterion, offset parameter β, then the calculation formula of F-Measure is as follows:
Wherein, when taking β=1, F score is the harmonic-mean of accurate rate and recall rate.
The beneficial effects of the present invention are: convolutional neural networks road scene classification of the present invention and lane segmentation method, base In convolutional neural networks, by road scene perceive in scene classification in conjunction with the double task of road extraction, by end to end Training mode is exported in real time, is beneficial to improve intelligent driving auxiliary system function, is provided effectively reliable drive and is assisted function Energy.Full convolutional network (FCN, Fully Convolutional wherein is used to the part of travelable extracted region in model Networks) with CRF (the Conditional Random Field) mode that condition random field combines, high-resolution can be generated Rate image optimization exports as a result, reaching more accurate segmentation effect.
Detailed description of the invention
Fig. 1 is that the double task cooperative structure convolutional neural networks road scene classification of the present invention are illustrated with lane segmentation model Figure;
Fig. 2 is the structure chart for the convolutional neural networks VGG-16 that the present invention uses.
Specific embodiment
1, system forms
The present invention realizes that road scene classification is divided into two parts with road area, and first part is trained part, The training of network is carried out first.It is completed offline before system use.Second part is part of detecting, using having completed to train Network, the road image of input is differentiated.The processing step of two parts is identical, and specific steps see below (3. processing Process).Two-part difference is input data difference.The input data of first part is prior ready instruction as needed Practice data set.The input data of second part is test data set.Test data set can be preprepared indiscipline Image.Training data and test data can also be made of the real-time road image of collection in worksite.
2, data set
The data set that the present invention uses is by 4000 width road scene figures in SUN data set, KITTI data set and CamVid As composition, training and test are participated in using 4 class label images, classification is divided into desert road, avenue, highway and rural area 4 class road scene of path, each classification include 1000 images.
The ratio that training dataset and test data set is arranged is 3:1.A quarter is randomly selected from each classification Image forms test set image.Training dataset and test data set can also be as needed, by vehicle-mounted camera to road scene Video acquisition is carried out, then by Video Quality Metric is that image is constituted.In vehicle travel process, by the original graph of vehicle-mounted camera acquisition It is preprocessed as needing, then input convolutional neural networks model.
Input picture size is unified for 360 × 480 RGB image.Place is normalized to the contrast of image, brightness Reason.
3, process flow
Double task cooperative structure convolutional neural networks road scene classification of the invention and lane segmentation model include four layers of knot Structure: input layer, feature extraction layer, information processing layer and output layer.
Input layer is image input part.Described image refers to above-mentioned training set image or test set image.
Feature information extraction layer is the feature coding device generated by a convolutional neural networks.It is mentioned from the image of input layer It wins the confidence characteristics of image abundant needed for breath process layer.
Information processing layer is made of two decoders, not to be classification decoder and segmentation decoder.Two decoders are total The image feature information generated by feature extraction layer is enjoyed, road scene classification and lane segmentation task are respectively completed.
Output layer obtains the output result of convolutional neural networks.The result of two decoders is finally obtained through joint reasoning The result of road scene perception.
4, specific steps
Since target training dataset is smaller, in order to improve precision as far as possible in the case where low volume data, therefore uses and turn It moves learning method and carries out network training.Its basic thought be first pass through to be trained on another large-scale data set, and Then the model parameter that training is obtained is instructed again as the initial value of present networks parameter on target training dataset Practice, tuning further is carried out to original parameter, is also small parameter perturbations.The principle that shift learning is relied on is certain features similar General character having the same on training dataset.
By the method training network of shift learning, using the model parameter of the training on large-scale data set as this reality The problem of initial value for testing model, is tested on this basis, makes up data set too small bring accuracy decline.It will be 2012 It is trained in advance on ILSVRC (the ImageNet Large Scale Visual Recognition Challenge) data set in year Initial weight of the good VGG-16 network model parameter as feature coding device and classification decoder.It is used when training network Mini-batch gradient descent method optimizes target loss function J, and Study rate parameter (base_lr) is set as 0.01, moves Measuring parameter (momentum) is 0.9, and weight decaying (weight_decay) is set as 0.0005, is then constantly trained to network Until training loss function convergence.After training set picture is disturbed, extracted by the sequence that ten pictures are a batch, it is ensured that repeatedly The access times of every image are 1 during generation, and training VGGNet network is restrained until loss function.VGG mould is used in experiment Type realizes image characteristics extraction.
Model part parameter setting will be enumerated below, as shown in table 1.
Table 1
Parameter Default value Function
gpu-nu 0 Which GPU is selected to accelerate convolution algorithm
image_resize 360×480 Image size
weights Nothing The Model Weight of network
file_image Nothing The input file of image
In table 1, the effect of parameter gpu-nu is the sequence of decision model image processor GPU used when carrying out feature extraction Number, it, when default value is 0, is exactly GPU that when multiple GPU are acted on simultaneously, model can select corresponding GPU according to the parameter Number be 1.Parameter image_resize shows size when image input, in road scene perception task, image, The picture size that video detection arrives is generally unknown, and usual VGG model uses fixed size 224 × 224 as input, this hair The bright parameter value for defaulting picture size in an experiment is set as 360 × 480.Parameter weights is the power of convolutional neural networks model Weight, weight file is relevant with the network model used, when the number of network training is not identical, the updated power of program iteration Weight is also not quite similar.Initial weight of the invention uses ILSVRC data set training weight, is tested on this basis; Parameter file_image corresponds to the input address of experimental data set, while containing test set and training set.In training process, make The objective function for intersecting entropy loss (cross-entropy loss) function as training network is used, it is all in batchsize Image pixel is both participated in unwise calculation.When the pixel quantity between classification each in training set has very large deviation, based on true Classification need differentially to weigh loss, this process or processing be known as class between balance.
Specific processing step is as follows.
Step 1: input picture.Input size is unified for 360 × 480 RGB image.
Step 2: using convolutional neural networks VGG-16 as the encoder of model, extracts the feature of road image, generates Characteristic pattern.
The task of feature extraction layer is processing image and extracts lane segmentation and scene classification abundant that required necessity is schemed As information.Depth convolutional neural networks VGG16 model can fast and effeciently learn scene image feature, so that scene image is special Sign has more abstractness, and ability to express is more prominent.
The structure chart of the convolutional neural networks VGG-16 used as shown in Figure 2, VGG-16 include five sections of convolution, each section of volume Product includes 2-3 convolutional layer, while every segment trailer can connect a maximum pond layer (Maxpool) and be used to downscaled images size. The quantity of convolution kernel in every section is the same, and the convolution nuclear volume of section more rearward is more: for 64-128-256-512-512.
In cataloged procedure, retain the whole convolutional layers and pond layer of VGG-16, gives up and connect entirely for three of classification below Connect layer.It the use of size is 3*3, the convolution kernel that step-length is 1 can generate corresponding characteristic pattern when executing convolution operation each time, Extract the minutia of image.Nonlinear Mapping is carried out to its result with ReLU activation primitive, is then 2*2, step-length with size Pondization operation is carried out for 2 filter, the feature of image is screened.Each down-sampling becomes the length of characteristic pattern with width For original half.After multiple convolution, pondization operation, the spy of fixed resolution needed for finally obtaining information processing layer Sign figure.
Step 3: using dual-task decoder (respectively scene classification decoder and lane segmentation decoder) to step 2 The characteristic pattern of middle generation is decoded analysis, achievees the purpose that scene type determines and road can travel extracted region.
(1) classification decoder
Road scene includes desert road, four class of avenue, highway and country road, and when classification uses Softmax classifier.If m road scene data set sample is expressed as follows:
{(x1,y1),(x2,y2),……,(xm,ym)}
Wherein, training or test image x are given and extracts feature, x by feature extraction layeriIt is mentioned for i-th of sample through feature The feature vector exported after taking layer to handle, yiIt is the class label of i-th of sample, yi∈ { 1,2 ..., k }, data set include four altogether Class road scene image, then k=4.
For given test sample image x, indicate that sample x belongs to the general of a certain classification j with function p (y=j | x) Rate, wherein j ∈ { 1,2 ..., k }.Sample image x is divided into the probability of classification j in Softmax classifier are as follows:
Wherein, ω=(ω1, ω2..., ωk) be in full articulamentum neuron with export mind i-th of Softmax classifier The weight parameter being connected through member, ωTX is the weighted term of parameter ω and sample attribute.Pω(xi) it is a probability matrix.Probability With for 1, in particular to
It is that image is judged to the probability of each scene type to be indicated with the number between 0 to 1 that it, which is acted on, and with matrix Form output.
When classifier classifies to road image, the probability that image belongs to each classification is calculated, takes probability value maximum It is classified as the classification of the image.
(2) divide decoder
Full convolutional network (FCN, Fully Convolutional is used when travelable extracted region in road scene Networks) the mode combined with CRF (Conditional Random Field) condition random field.Original road scene figure As the size reduced through feature coding device Chi Huahou, low resolution characteristic pattern is generated, to obtain the segmentation figure big with original image etc., Traditional FCN (Fully Convolutional Networks) adopt by deconvolution using convolutional layer to output result Sample generates high-definition picture optimization output as a result, reaching segmentation effect in conjunction with jump link structure, but FCN is to increase image Size can introduce noise to the way that original image is filled, and the result divided is not fine enough, lack Space Consistency.And incite somebody to action When FCN is combined with CRF, training simultaneously generates a network end to end, promotes segmentation effect.In the segmentation of road area In task, first using full convolutional network FCN in input picture road surface and background carry out rough sort, obtain each pixel Belong to road surface or belongs to the probability of non-road surface (background).Then, use condition random field, to road surface in rough sort result with it is non- The marginal portion on road surface optimizes, and obtains that edge is more careful, more accurate lane segmentation result on the whole.Condition random field energy By the RGB feature of input picture and its spatial relation all as reference, maximum a posteriori is carried out in the prediction result of model Distributed operation, to enhance the segmentation effect of FCN.
Assuming that there are random areas X=(X1, X2..., Xn), n is the quantity of pixel, XiIt is the mark class of pixel i Not, wherein i ∈ { 1,2 ..., m }, all mark classifications constitute mark collection, can use L=(l1, l2..., lk) indicate, k is classification Number;There are another random areas M=(M1, M2..., Mn), MiIt is the color feature vector of pixel i.For condition random field (M, X) provides a probability function:
Wherein, E (X | M) is energy function, is simplified shown as E (X), and X belongs to mark collection L, and Z (M) is normalization factor.It is logical It crosses minimum energy function and obtains optimal pixel classifications result.The energy object function of condition random field are as follows:
E (X)=∑iψu(xi)+∑i,jψp(xi,xj) (3)
Wherein, i, j indicate pixel.The first item ψ of energy equationu(xi)=- logP (xi), referred to as unitary potential function should Pixel belongs to class label XiProbability, P (xi) indicate that pixel i belongs to the probability of a certain class mark.The second of energy equation Item ψp(xi,xj) it is pairs of potential function, belong to of a sort probability P (x for measuring two pixelsi,xj).And:
Wherein, μ (xi,xj) it is label comparison function, xi≠xjWhen, μ (xi,xjOtherwise)=1 is equal to 0, for judging difference Compatibility between label;P indicates location information;M indicates colouring information;θα、θγFor controlling the scale of location information;θβWith In the scale of control color similarity;ω is linear combination weight.
The prior probability of the unitary potential function item of condition random field is provided by FCN, primary Calculation go out pixel belong to it is each Then the probability of classification enhances segmentation effect by pairs of potential function.CRF is in image segmentation task by combining original image The mode of RGB feature improves the segmentation effect of network.When image edge area color difference is larger, the pixel quilt of both sides of edges It is larger to be divided into different classes of probability.
Step 4: the result that two kinds of decoders generate exported through joint reasoning to (joint reasoning refers to classification decoder With segmentation decoder share convolutional layer provide characteristic information, model training when only train a weight network, model into When row backpropagation parameter adjusts, the network optimization is carried out by the way of superposition classification decoder and segmentation decoder losses), it is right Road scene carries out effectively perceive.
Full convolutional network FCN and convolutional neural networks model VGG-16 convolutional layer having the same in structure, therefore Realize that road scene classification with when can exercise extracted region, is used for feature extraction using identical convolutional layer, the spy as model Encoder is levied, achievees the purpose that feature information extraction layer shares convolutional layer.Decoder of classifying is raw using the full articulamentum of VGG-16 At one-dimensional characteristic vector, then classified using Softmax classifier, completes the task of multiclass road scene classification.Segmentation Decoder in such a way that condition random field combines, completes lane segmentation task using full convolutional network.Whole network uses end To the training method at end, model only trains a weight network, passes through superposition classification solution in the back-propagation process of inclusion layer The loss of code device and segmentation decoder carries out the network optimization, is completed at the same time the task of scene classification and lane segmentation.
The classification performance of model is by calculating this two indexs progress of accuracy rate (Precision) and recall rate (Recall) It measures.In image classification task, it is the figure for really belonging to the category among the picture of certain classification that accuracy rate expression, which is predicted to be, Piece ratio.And recall rate indicates among the actual test sample of certain classification picture finally by model partition to such sample ratio The formula difference of example, accuracy rate and recall rate is as follows:
Wherein, TP indicates that the sample size being correctly detecting, FP indicate that the correct sample being not detected, FN indicate wrong The sample that erroneous detection measures.
Evaluation to segmentation result is exactly the evaluation to Pixel-level classification in fact.Lane segmentation problem can be considered as one two Classification problem, any pixel belong to road surface or belong to background.Classifier predicts the classification knot of output when carrying out pixel classifications Fruit and mark image carry out Pixel-level comparison, and the prediction for evaluating each pixel, which exports as a result, setting, is judged to road surface pixel for pixel For " positive class ", it is " negative class " that pixel, which is judged to background pixel point, then the pixel classifications result takes true positives (TP1), false positive (FP1), true negative (TN1), false negative (FN1One of) four kinds of results.Wherein TP1Indicate that positive class determines the class that is positive, FP1It indicates Negative class determines the class that is positive, TN1Indicate that negative class determines the class that is negative, FN1Indicate that positive class determines the class that is negative.And then it can be in the hope of entire The accurate rate (TPR) and recall rate (FPR) of each image on test set.
For the present invention using segmentation effect is evaluated based on the MaxF1 that pixel is classified, F-Measure is a kind of statistic, F-Measure is also known as F-Score, is the common evaluation criterion of information retrieval field.Offset parameter is β, then F-Measure Calculation formula is as follows:
Wherein, when taking β=1, F score is the harmonic-mean of accurate rate and recall rate.
Accuracy rate, recall rate and segmentation effect F value the heterogeneous networks recognition result pair as shown in table 2 that the method for the present invention obtains Than.
Table 2

Claims (4)

1. a kind of convolutional neural networks road scene classification and lane segmentation method, which is characterized in that specifically comprise the following steps:
1) image in database is selected, randomly assigne is carried out to selection image, a part is used as training set image, and another part is made For test set image;
2) by training set image be sent into double task cooperative structure convolutional neural networks road scenes classification and lane segmentation model into The training of row convolutional neural networks, convolutional neural networks after being trained;
Double task cooperative structure convolutional neural networks road scene classification and lane segmentation model include input layer, feature extraction Layer, information processing layer and output layer, training set image are unified for an equal amount of RGB image and are sent into input layer, input layer output Image is sent into the feature coding device that is generated by convolutional neural networks of feature extraction layer, extracted in feature coding device lane segmentation with The characteristics of image that scene classification needs generates characteristic pattern;Information processing layer is by decoder two decodings of classification decoder and segmentation Device composition, the characteristic pattern of feature extraction layer output respectively enter two decoders progress scene types and determine to can travel area with road It extracts in domain;The result of two decoder outputs is sent into output layer, and through joint reasoning, joint reasoning refers to that sorting code number device and segmentation are compiled Code device shares identical characteristic information, and model only trains a weight network in training, and model is carrying out backpropagation parameter When adjustment, the network optimization is carried out by the way of superposition sorting code number device and the loss of partition encoding device, to road after being trained The convolutional neural networks of scene perception, achievable scene classification determines and the task of lane segmentation;
3) convolutional neural networks identify after test set image being sent into step 2) training, and achievable scene classification determines and road The task of road segmentation.
2. the classification of convolutional neural networks road scene and lane segmentation method according to claim 1, which is characterized in that described Step 2) feature coding device uses convolutional neural networks VGG-16 as model;Decoder of classifying uses convolutional neural networks VGG- 16 full articulamentum generates one-dimensional characteristic vector, is then classified using Softmax classifier, completes multiclass road scene point The task of class;Divide decoder using full convolutional network in such a way that condition random field combines, completion lane segmentation task;It is whole A network only trains a weight network using training method end to end, model, leads in the back-propagation process of inclusion layer The loss progress network optimization for being superimposed classification decoder and dividing decoder is crossed, times of scene classification and lane segmentation is completed at the same time Business.
3. convolutional neural networks road scene classification according to claim 1 or claim 2 and lane segmentation method, which is characterized in that The classification performance of step 2) the classification decoder is by calculating this two indexs of accuracy rate Precision and recall rate Recall It is measured, the formula difference of accuracy rate and recall rate is as follows:
Wherein, TP indicates that the sample size being correctly detecting, FP indicate that the correct sample being not detected, FN indicate to be examined by mistake The sample measured.
4. convolutional neural networks road scene classification according to claim 1 or claim 2 and lane segmentation method, which is characterized in that Step 2) the segmentation decoder is exactly the evaluation to Pixel-level classification to the evaluation of segmentation result in fact, and lane segmentation problem can It is considered as two classification problems, any pixel belongs to road surface or belongs to background;Divide decoder when carrying out pixel classifications, The classification results and mark image for predicting output carry out Pixel-level comparison, and the prediction for evaluating each pixel exports as a result, setting picture It is " positive class " that element, which is judged to road surface pixel, and it is " negative class " that pixel, which is judged to background pixel point, then the pixel classifications result takes very Positive TP1, false positive FP1, true negative TN1, false negative FN1One of four kinds of results;Wherein TP1Indicate that positive class judgement is positive Class, FP1Indicate that negative class determines the class that is positive, TN1Indicate that negative class determines the class that is negative, FN1Indicate that positive class determines the class that is negative.And then it can be with The accurate rate TPR and recall rate FPR of each image on entire test set are acquired,
Segmentation effect is evaluated using the MaxF1 to classify based on pixel, F-Measure is that information retrieval field is commonly evaluated Standard, offset parameter β, then the calculation formula of F-Measure is as follows:
Wherein, when taking β=1, F score is the harmonic-mean of accurate rate and recall rate.
CN201910211802.XA 2019-03-20 2019-03-20 Convolutional neural network road scene classification and road segmentation method Expired - Fee Related CN109993082B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910211802.XA CN109993082B (en) 2019-03-20 2019-03-20 Convolutional neural network road scene classification and road segmentation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910211802.XA CN109993082B (en) 2019-03-20 2019-03-20 Convolutional neural network road scene classification and road segmentation method

Publications (2)

Publication Number Publication Date
CN109993082A true CN109993082A (en) 2019-07-09
CN109993082B CN109993082B (en) 2021-11-05

Family

ID=67130681

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910211802.XA Expired - Fee Related CN109993082B (en) 2019-03-20 2019-03-20 Convolutional neural network road scene classification and road segmentation method

Country Status (1)

Country Link
CN (1) CN109993082B (en)

Cited By (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109902600A (en) * 2019-02-01 2019-06-18 清华大学 A kind of road area detection method
CN110458047A (en) * 2019-07-23 2019-11-15 北京理工大学 A kind of country scene recognition method and system based on deep learning
CN110503637A (en) * 2019-08-13 2019-11-26 中山大学 A kind of crack on road automatic testing method based on convolutional neural networks
CN110598586A (en) * 2019-08-27 2019-12-20 电子科技大学 Target detection method and system
CN110599497A (en) * 2019-07-31 2019-12-20 中国地质大学(武汉) Drivable region segmentation method based on deep neural network
CN110717383A (en) * 2019-08-29 2020-01-21 阿里巴巴集团控股有限公司 Remote sensing detection method, device and system
CN110781773A (en) * 2019-10-10 2020-02-11 湖北工业大学 Road extraction method based on residual error neural network
CN111144418A (en) * 2019-12-31 2020-05-12 北京交通大学 Railway track area segmentation and extraction method
CN111160109A (en) * 2019-12-06 2020-05-15 北京联合大学 Road segmentation method and system based on deep neural network
CN111161282A (en) * 2019-12-30 2020-05-15 西南交通大学 Target scale selection method for image multi-level segmentation based on depth seeds
CN111191654A (en) * 2019-12-30 2020-05-22 重庆紫光华山智安科技有限公司 Road data generation method and device, electronic equipment and storage medium
CN111368846A (en) * 2020-03-19 2020-07-03 中国人民解放军国防科技大学 Road ponding identification method based on boundary semantic segmentation
CN111611919A (en) * 2020-05-20 2020-09-01 西安交通大学苏州研究院 Road scene layout analysis method based on structured learning
CN111898702A (en) * 2020-08-14 2020-11-06 海之韵(苏州)科技有限公司 Unmanned ship environment intelligent sensing method based on deep learning
CN111985324A (en) * 2020-07-14 2020-11-24 广西大学 Road detection method combining full convolution regression neural network and conditional random field
CN111985378A (en) * 2020-08-13 2020-11-24 中国第一汽车股份有限公司 Road target detection method, device and equipment and vehicle
CN112092827A (en) * 2020-09-23 2020-12-18 北京百度网讯科技有限公司 Automatic driving function control method, automatic driving function control device, electronic equipment and storage medium
CN112528500A (en) * 2020-12-11 2021-03-19 深兰科技(上海)有限公司 Evaluation method and evaluation equipment for scene graph construction model
CN112669316A (en) * 2021-01-29 2021-04-16 南方电网调峰调频发电有限公司 Power production abnormity monitoring method and device, computer equipment and storage medium
CN112766092A (en) * 2021-01-05 2021-05-07 北京航空航天大学 Method for quickly identifying background category based on brain-like neural network and application thereof
CN112949617A (en) * 2021-05-14 2021-06-11 江西农业大学 Rural road type identification method, system, terminal equipment and readable storage medium
CN113015887A (en) * 2019-10-15 2021-06-22 谷歌有限责任公司 Navigation directions based on weather and road surface type
CN113096001A (en) * 2021-04-01 2021-07-09 咪咕文化科技有限公司 Image processing method, electronic device and readable storage medium
CN113362332A (en) * 2021-06-08 2021-09-07 南京信息工程大学 Depth network segmentation method for coronary artery lumen contour under OCT image
CN113408633A (en) * 2021-06-29 2021-09-17 北京百度网讯科技有限公司 Method, apparatus, device and storage medium for outputting information
CN113449589A (en) * 2021-05-16 2021-09-28 桂林电子科技大学 Method for calculating driving strategy of unmanned automobile in urban traffic scene
EP3901822A1 (en) * 2020-04-24 2021-10-27 Stradvision, Inc. Method and device for on-vehicle active learning to be used for training perception network of autonomous vehicle
CN115223119A (en) * 2022-06-15 2022-10-21 广州汽车集团股份有限公司 Driving region detection method and system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106599773A (en) * 2016-10-31 2017-04-26 清华大学 Deep learning image identification method and deep learning image identification system used for intelligent driving, and terminal device
CN107609602A (en) * 2017-09-28 2018-01-19 吉林大学 A kind of Driving Scene sorting technique based on convolutional neural networks
CN108492272A (en) * 2018-03-26 2018-09-04 西安交通大学 Cardiovascular vulnerable plaque recognition methods based on attention model and multitask neural network and system
CN108876796A (en) * 2018-06-08 2018-11-23 长安大学 A kind of lane segmentation system and method based on full convolutional neural networks and condition random field

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106599773A (en) * 2016-10-31 2017-04-26 清华大学 Deep learning image identification method and deep learning image identification system used for intelligent driving, and terminal device
CN107609602A (en) * 2017-09-28 2018-01-19 吉林大学 A kind of Driving Scene sorting technique based on convolutional neural networks
CN108492272A (en) * 2018-03-26 2018-09-04 西安交通大学 Cardiovascular vulnerable plaque recognition methods based on attention model and multitask neural network and system
CN108876796A (en) * 2018-06-08 2018-11-23 长安大学 A kind of lane segmentation system and method based on full convolutional neural networks and condition random field

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
HAO ZHOU 等: "Image Semantic Segmentation Based on FCN-CRF Model", 《2016 INTERNATIONAL CONFERENCE ON IMAGE, VISION AND COMPUTING》 *
KIM-HAN THUNG 等: "A brief review on multi-task learning", 《MULTIMEDIA TOOLS AND APPLICATIONS》 *

Cited By (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109902600A (en) * 2019-02-01 2019-06-18 清华大学 A kind of road area detection method
CN109902600B (en) * 2019-02-01 2020-10-27 清华大学 Road area detection method
CN110458047A (en) * 2019-07-23 2019-11-15 北京理工大学 A kind of country scene recognition method and system based on deep learning
CN110599497A (en) * 2019-07-31 2019-12-20 中国地质大学(武汉) Drivable region segmentation method based on deep neural network
CN110503637A (en) * 2019-08-13 2019-11-26 中山大学 A kind of crack on road automatic testing method based on convolutional neural networks
CN110503637B (en) * 2019-08-13 2022-12-06 中山大学 Road crack automatic detection method based on convolutional neural network
CN110598586A (en) * 2019-08-27 2019-12-20 电子科技大学 Target detection method and system
CN110717383A (en) * 2019-08-29 2020-01-21 阿里巴巴集团控股有限公司 Remote sensing detection method, device and system
CN110717383B (en) * 2019-08-29 2023-05-02 阿里巴巴集团控股有限公司 Remote sensing detection method, device and system
CN110781773A (en) * 2019-10-10 2020-02-11 湖北工业大学 Road extraction method based on residual error neural network
CN113015887A (en) * 2019-10-15 2021-06-22 谷歌有限责任公司 Navigation directions based on weather and road surface type
CN111160109A (en) * 2019-12-06 2020-05-15 北京联合大学 Road segmentation method and system based on deep neural network
CN111160109B (en) * 2019-12-06 2023-08-18 北京联合大学 Road segmentation method and system based on deep neural network
CN111161282B (en) * 2019-12-30 2021-10-29 西南交通大学 Target scale selection method for image multi-level segmentation based on depth seeds
CN111191654A (en) * 2019-12-30 2020-05-22 重庆紫光华山智安科技有限公司 Road data generation method and device, electronic equipment and storage medium
CN111161282A (en) * 2019-12-30 2020-05-15 西南交通大学 Target scale selection method for image multi-level segmentation based on depth seeds
CN111144418B (en) * 2019-12-31 2022-12-02 北京交通大学 Railway track area segmentation and extraction method
CN111144418A (en) * 2019-12-31 2020-05-12 北京交通大学 Railway track area segmentation and extraction method
CN111368846A (en) * 2020-03-19 2020-07-03 中国人民解放军国防科技大学 Road ponding identification method based on boundary semantic segmentation
CN111368846B (en) * 2020-03-19 2022-09-09 中国人民解放军国防科技大学 Road ponding identification method based on boundary semantic segmentation
KR102589764B1 (en) 2020-04-24 2023-10-17 주식회사 스트라드비젼 On-vehicle active learning method and device for learning the perception network of an autonomous vehicle
KR20210152025A (en) * 2020-04-24 2021-12-14 주식회사 스트라드비젼 On-Vehicle Active Learning Method and Apparatus for Learning Perception Network of Autonomous Vehicle
EP3901822A1 (en) * 2020-04-24 2021-10-27 Stradvision, Inc. Method and device for on-vehicle active learning to be used for training perception network of autonomous vehicle
CN111611919A (en) * 2020-05-20 2020-09-01 西安交通大学苏州研究院 Road scene layout analysis method based on structured learning
CN111985324A (en) * 2020-07-14 2020-11-24 广西大学 Road detection method combining full convolution regression neural network and conditional random field
CN111985324B (en) * 2020-07-14 2022-10-28 广西大学 Road detection method combining full convolution regression neural network and conditional random field
CN111985378A (en) * 2020-08-13 2020-11-24 中国第一汽车股份有限公司 Road target detection method, device and equipment and vehicle
CN111898702A (en) * 2020-08-14 2020-11-06 海之韵(苏州)科技有限公司 Unmanned ship environment intelligent sensing method based on deep learning
CN112092827B (en) * 2020-09-23 2022-04-22 北京百度网讯科技有限公司 Automatic driving function control method, automatic driving function control device, electronic equipment and storage medium
CN112092827A (en) * 2020-09-23 2020-12-18 北京百度网讯科技有限公司 Automatic driving function control method, automatic driving function control device, electronic equipment and storage medium
CN112528500A (en) * 2020-12-11 2021-03-19 深兰科技(上海)有限公司 Evaluation method and evaluation equipment for scene graph construction model
CN112528500B (en) * 2020-12-11 2023-08-29 深兰人工智能应用研究院(山东)有限公司 Evaluation method and evaluation equipment for scene graph construction model
CN112766092A (en) * 2021-01-05 2021-05-07 北京航空航天大学 Method for quickly identifying background category based on brain-like neural network and application thereof
CN112669316A (en) * 2021-01-29 2021-04-16 南方电网调峰调频发电有限公司 Power production abnormity monitoring method and device, computer equipment and storage medium
CN113096001A (en) * 2021-04-01 2021-07-09 咪咕文化科技有限公司 Image processing method, electronic device and readable storage medium
CN112949617A (en) * 2021-05-14 2021-06-11 江西农业大学 Rural road type identification method, system, terminal equipment and readable storage medium
CN113449589A (en) * 2021-05-16 2021-09-28 桂林电子科技大学 Method for calculating driving strategy of unmanned automobile in urban traffic scene
CN113362332A (en) * 2021-06-08 2021-09-07 南京信息工程大学 Depth network segmentation method for coronary artery lumen contour under OCT image
CN113408633A (en) * 2021-06-29 2021-09-17 北京百度网讯科技有限公司 Method, apparatus, device and storage medium for outputting information
CN115223119A (en) * 2022-06-15 2022-10-21 广州汽车集团股份有限公司 Driving region detection method and system

Also Published As

Publication number Publication date
CN109993082B (en) 2021-11-05

Similar Documents

Publication Publication Date Title
CN109993082A (en) The classification of convolutional neural networks road scene and lane segmentation method
CN108985194B (en) Intelligent vehicle travelable area identification method based on image semantic segmentation
CN111612807B (en) Small target image segmentation method based on scale and edge information
CN113159051B (en) Remote sensing image lightweight semantic segmentation method based on edge decoupling
CN110263786B (en) Road multi-target identification system and method based on feature dimension fusion
CN108509978A (en) The multi-class targets detection method and model of multi-stage characteristics fusion based on CNN
US11651302B2 (en) Method and device for generating synthetic training data for an artificial-intelligence machine for assisting with landing an aircraft
CN110826638A (en) Zero sample image classification model based on repeated attention network and method thereof
CN107133974A (en) The vehicle type classification method that Gaussian Background modeling is combined with Recognition with Recurrent Neural Network
CN112464911A (en) Improved YOLOv 3-tiny-based traffic sign detection and identification method
CN112801270B (en) Automatic U-shaped network slot identification method integrating depth convolution and attention mechanism
CN112183635A (en) Method for realizing segmentation and identification of plant leaf lesions by multi-scale deconvolution network
CN111753787A (en) Separated traffic sign detection and identification method
CN109670392A (en) Based on mixing autocoder road image semantic segmentation method
CN110659601B (en) Depth full convolution network remote sensing image dense vehicle detection method based on central point
CN110281949B (en) Unified hierarchical decision-making method for automatic driving
CN115035361A (en) Target detection method and system based on attention mechanism and feature cross fusion
CN116110022B (en) Lightweight traffic sign detection method and system based on response knowledge distillation
CN110008899A (en) A kind of visible remote sensing image candidate target extracts and classification method
CN113298817A (en) High-accuracy semantic segmentation method for remote sensing image
CN111310820A (en) Foundation meteorological cloud chart classification method based on cross validation depth CNN feature integration
CN110310298A (en) A kind of road target real-time three-dimensional point cloud segmentation method based on cycling condition random field
CN113792631A (en) Aircraft detection and tracking method based on multi-scale self-adaption and side-domain attention
CN112084897A (en) Rapid traffic large-scene vehicle target detection method of GS-SSD
CN115205568B (en) Road traffic multi-element detection method based on multi-scale feature fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20211105