CN112215487B - Vehicle running risk prediction method based on neural network model - Google Patents
Vehicle running risk prediction method based on neural network model Download PDFInfo
- Publication number
- CN112215487B CN112215487B CN202011076552.2A CN202011076552A CN112215487B CN 112215487 B CN112215487 B CN 112215487B CN 202011076552 A CN202011076552 A CN 202011076552A CN 112215487 B CN112215487 B CN 112215487B
- Authority
- CN
- China
- Prior art keywords
- vehicle
- data
- neural network
- training
- value
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 41
- 238000003062 neural network model Methods 0.000 title claims abstract description 12
- 238000012549 training Methods 0.000 claims abstract description 32
- 238000013528 artificial neural network Methods 0.000 claims abstract description 17
- 238000012360 testing method Methods 0.000 claims abstract description 7
- 230000006870 function Effects 0.000 claims description 19
- 230000001133 acceleration Effects 0.000 claims description 12
- 238000002372 labelling Methods 0.000 claims description 7
- 230000008859 change Effects 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 4
- 230000008929 regeneration Effects 0.000 claims description 3
- 238000011069 regeneration method Methods 0.000 claims description 3
- 238000000605 extraction Methods 0.000 abstract description 2
- 230000008569 process Effects 0.000 description 10
- 238000013527 convolutional neural network Methods 0.000 description 9
- 238000012502 risk assessment Methods 0.000 description 8
- 230000006399 behavior Effects 0.000 description 6
- 238000013135 deep learning Methods 0.000 description 6
- 230000003993 interaction Effects 0.000 description 5
- 206010039203 Road traffic accident Diseases 0.000 description 4
- 230000004913 activation Effects 0.000 description 3
- 230000007423 decrease Effects 0.000 description 3
- 239000000284 extract Substances 0.000 description 3
- 238000011176 pooling Methods 0.000 description 3
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000008447 perception Effects 0.000 description 2
- 238000009966 trimming Methods 0.000 description 2
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 1
- 102100034112 Alkyldihydroxyacetonephosphate synthase, peroxisomal Human genes 0.000 description 1
- 101000799143 Homo sapiens Alkyldihydroxyacetonephosphate synthase, peroxisomal Proteins 0.000 description 1
- 206010033307 Overweight Diseases 0.000 description 1
- 230000003213 activating effect Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000000848 angular dependent Auger electron spectroscopy Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 238000012804 iterative process Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 235000020825 overweight Nutrition 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/06—Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
- G06Q10/063—Operations research, analysis or management
- G06Q10/0635—Risk analysis of enterprise or organisation activities
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/40—Business processes related to the transportation industry
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/01—Detecting movement of traffic to be counted or controlled
- G08G1/0104—Measuring and analyzing of parameters relative to traffic conditions
- G08G1/0125—Traffic data processing
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/16—Anti-collision systems
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Resources & Organizations (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Strategic Management (AREA)
- Economics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Computational Linguistics (AREA)
- General Business, Economics & Management (AREA)
- Tourism & Hospitality (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Marketing (AREA)
- Molecular Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Entrepreneurship & Innovation (AREA)
- Game Theory and Decision Science (AREA)
- Educational Administration (AREA)
- Development Economics (AREA)
- Operations Research (AREA)
- Quality & Reliability (AREA)
- Analytical Chemistry (AREA)
- Chemical & Material Sciences (AREA)
- Primary Health Care (AREA)
- Traffic Control Systems (AREA)
Abstract
The invention discloses a vehicle running risk prediction method based on a neural network model, which comprises the following steps: collecting vehicle running data to form vehicle history data; step two: extracting features of the vehicle history data by adopting a context time window to form statistical features; step three: extracting the statistical features, and step four: dividing the extraction result data in the third step, and in the fifth step: constructing a neural network; constructing an LSTM encoder-1 DCNN-LSTM decoder network architecture; step six: the unlabeled dataset is denoted as { X ] U -a }; dividing the labeled dataset into a training set and a test set, wherein the training set is denoted as S L ={X L ,Y L -a }; using labeled dataset S L Pre-training the neural network; then enter a self-learning stage for the set Y PL Predicting, and directly taking the predicted value as a real label; after the completion, all the labels without label data and the trained network model are obtained.
Description
Technical Field
The invention relates to the field of machine learning, in particular to a vehicle running risk prediction method based on a neural network model.
Background
According to statistics of the world health organization in 2018, 135 ten thousand people lose life due to traffic accidents each year, and the loss caused by road traffic collision can reach 3% of the total national production value of most countries. In addition, in traffic-related death accidents, about 94% are caused by drivers, and improper driving behavior of the drivers becomes a major factor in traffic accidents. These driving behaviors are often caused by poor perception of the surrounding environment by the driver, misleading or aggressive decisions and decisions, improper driving maneuvers of the vehicle. The driving risk assessment is based on analysis of various driving characteristics (including driver, vehicle and surrounding environment) at present and past time, and gives possibility of collision or other traffic accidents of the present vehicle. And (3) evaluating and predicting the running safety of the vehicle, and timely feeding back to a driver so as to improve the running safety of the vehicle and further reduce traffic accidents. Therefore, it is indispensable to evaluate and predict the running risk of the vehicle.
However, in driving risk assessment tasks, tagging driving data with risk is a challenging task. If the data is classified by using an unsupervised learning method, the obtained result may not be classified strictly according to the level of risk, and it is difficult to achieve a satisfactory effect in terms of accuracy. In addition, the driving risk assessment task needs to face massive high-dimensional, time-sequence-carrying and class-unbalanced vehicle driving data. Finally, the running risk assessment has a high requirement for accuracy, and it is difficult to accept the case where a high risk is determined as risk-free. In summary, there are great challenges in comprehensively, accurately, and efficiently assessing driving risk.
Through retrieval, chinese patent No. 201711234967.6 discloses a driving risk warning method and device, wherein a pre-established BP network is used for classifying corresponding road sections into high-risk road sections or low-risk road sections; sending out warning information to the vehicle so as to control the vehicle to send out warning when driving to a high-risk road section; CN201910574565.3 discloses a vehicle illegal running risk analysis method based on a Beidou positioning system, by accurately calculating the risk score of the illegal running of the vehicle, a user generates a risk analysis report of the illegal running according to the risk score of the illegal running of the vehicle, reminds and urges a driver to improve driving behaviors, and plays a role in early warning and checking the driving behaviors of the driver. However, the method is insufficient in consideration of massive high-dimensionality vehicle running data with time sequence and unbalanced categories in the vehicle running process, and is difficult to achieve fine running risk prediction, so that the accuracy is poor.
Disclosure of Invention
The invention designs and develops a vehicle driving risk prediction method based on a neural network model, which can only manually label a small part of data, automatically learn potential characteristics, build the neural network model and predict risk values in a future period.
Another object of the invention is a neural network model training method for vehicle driving risk prediction that can manually label only a small portion of data and automatically learn potential features to build a neural network model.
A vehicle running risk prediction method based on a neural network model,
step one: collecting vehicle running data to form vehicle history data;
step two: extracting features of the vehicle history data by adopting a context time window to form statistical features;
step three: extracting the statistical features, including: the type of vehicle, length and width of the vehicle; steering entropy value; parameter time to reverse collision TTC -1 Time interval THW of inverse head -1 Distance between head of inverse vehicle and DHW -1 The method comprises the steps of carrying out a first treatment on the surface of the The time length of the broken line is when the vehicle does not have the lane change intention, the time length of the solid line is when the vehicle is pressed, and the time length of the vehicle is when the vehicle is driven out of the solid line; local traffic flow density, local speed difference, and local acceleration difference.
Step four: dividing the extracted result data in the third step, randomly extracting no more than 5% of data from the data to carry out labeling, and forming a labeled data set; the rest data are label-free data sets, and are used for label-free training and testing of semi-supervised learning;
step five: constructing a neural network; constructing an LSTM encoder-1 DCNN-LSTM decoder network architecture;
step six: the unlabeled dataset is denoted as { X ] U -a }; dividing the labeled dataset into a training set and a test set, wherein the training set is denoted as S L ={X L ,Y L -a }; using labeled dataset S L Pre-training the neural network;
then enter a self-learning stage to make the unlabeled dataset { X } U Generating pseudo tags { Y } using a pre-trained network P -a }; for each generated pseudo tag, a certain confidence epsilon is carried out, and the confidence epsilon is matched with a threshold epsilon th Comparing, the set greater than the threshold is denoted as S P t ={X Uh t ,Y Ph t Sets smaller than the threshold are denoted Y PL t T is the iteration number; for the set { Y Ph t -considering the pseudo tag, i.e. the real tag, according to manifold assumptions; will be set S L And S is equal to P t Merging to form a new set S L t Then the training method is used for training the network; for { X ] U t Performing pseudo tag regeneration by using the retrained network; for the data set { X } which is not yet labeled in the last stage of self-learning U mst Predicting to take the predicted value as the real label; after the completion, all the labels without label data and the trained network model are obtained.
As a preference, the penalty function of the neural network is:
wherein the probability mass function f of the distribution P can be defined as
AOBC(0,k)=AOBC(k)
Wherein N (t, k) = |s in AOBC (k) L,k t N (0, k) = |s in|, OBC (k) L,k 0 I (I); n is the number of data in mini-batch, m is the category number, y ik To be a true value of the value,is a predicted value.
As a preference, a limit value penalty function is also included, as shown in the following formula:
where ev is the limit value.
As a preference, the steering entropy value SRE:
preferably, the reverse time to collision TTC -1
As one preferable, the specific calculation formula of the local traffic flow density is as follows:
wherein X is j =(x j ,y j ) T For a vehicle of interest, μ= (x, y) T As the center coordinates of the target vehicle,wherein sigma x And sigma (sigma) y Defined by the formula:
σ x =|v x |+k 1 L
σ y =|v y |+k 2 W
wherein v is x ,v y For transverse and longitudinal speed, k, of the vehicle 1 And k is equal to 2 Is a compensation factor.
Preferably, the local velocity difference is calculated as follows:
as one preferable, the local acceleration difference is calculated as follows:
the beneficial effects of the invention are as follows:
the method firstly extracts the characteristics of the running of the target vehicle, the interaction of the vehicle and the road, the local traffic condition and the like. And extracting features of massive vehicle driving data, and manually labeling a small part of the vehicle driving data, so that a small data set with labels and a large data set without labels are obtained. A convolutional neural network combining a one-dimensional convolutional neural network (1D-CNN) and a long and short time memory network (LSTM) is built, and a self-adaptive over-balanced cross entropy is adopted as a loss function of the neural network. Embedding the neural network into a semi-supervised learning framework, and pre-training, self-learning and fine-tuning the two data sets to obtain a final label result and a trained network. Finally, a limit penalty module is added to refine the model.
And a cost-sensitive semi-supervised deep learning method is adopted to analyze the vehicle driving data to obtain current and future driving risk scores, wherein the scores are continuous values ranging from 0 to 3, and the risk assessment is finer. The method can be applied to a driving risk warning system in the ADAS, so that timely feedback is given to a driver.
Good results can be obtained by only using a small amount of tag data, and the tag problem of a large amount of non-tag data is greatly solved. The self-adaptive over-balanced cross entropy loss function is adopted, so that the training performance of the semi-supervised deep learning with unbalanced categories is greatly improved. The loss function can enable the network to be in an over-balanced state in the whole training process, and the state can effectively improve the detection accuracy of high-risk data. The method can be applied to other related similar scenes.
A local traffic condition descriptor is provided for describing traffic conditions, speed and acceleration differences around a target vehicle, and a descriptor is provided, so that the description of surrounding scenes is simplified. The description may also be used in other similar fields.
Drawings
FIG. 1 is a schematic diagram of a contextual window of the present invention.
FIG. 2 is a schematic illustration of a vehicle of interest with respect to a target vehicle in accordance with the present invention.
FIG. 3 is a deep learning network model of the present invention.
Fig. 4 is a semi-supervised learning framework of the present invention.
Fig. 5 is a general framework of the present invention.
Detailed Description
The present invention is described in further detail below with reference to the drawings to enable those skilled in the art to practice the invention by referring to the description.
The invention provides a driving risk assessment and prediction method based on self-adaptive cost-sensitive semi-supervised deep learning. The method firstly extracts the characteristics of the running of the target vehicle, the interaction of the vehicle and the road, the local traffic condition and the like. And extracting features of massive vehicle driving data, and manually labeling a small part of the vehicle driving data, so that a small data set with labels and a large data set without labels are obtained. A convolutional neural network combining a one-dimensional convolutional neural network (1D-CNN) and a long and short time memory network (LSTM) is built, and a self-adaptive over-balanced cross entropy is adopted as a loss function of the neural network. Embedding the neural network into a semi-supervised learning framework, and pre-training, self-learning and fine-tuning the two data sets to obtain a final label result and a trained network. As a preference, finally, a limit value penalty module is added to refine the model. The method can only manually label a small part of data, automatically learn potential characteristics and predict risk values in a future period of time, so that timely feedback is provided for a driver.
Step 1: and converting and counting the vehicle driving history data. Because the original vehicle driving data is directly used as input, the characteristics are too sparse and the characteristic order is too low, so that the training difficulty is increased. A fixed window is extracted with a contextual time window (as shown in fig. 1). The width of the selection window is W w The sampling frequency per second is not less than 10Hz =3 seconds. A total of 5 seconds of history data is selected, and each time window overlap ratio ov=66.67%, i.e., 2 seconds. A total of 3 time windows are obtained and all data within the windows are counted.
Step 2: and extracting the statistical characteristics. The 47-dimensional features are extracted altogether, and comprise vehicle basic information, vehicle running and interaction information, vehicle surrounding environment information and the like. Each statistic has mean, maximum, minimum, variance statistics, so each data has 4 statistics.
A) Basic information of the vehicle:
the type of vehicle, the length L and the width W of the vehicle.
B) Target vehicle travel characteristics:
the longitudinal travel direction of the vehicle is defined as the x-axis and the lateral travel direction as the y-axis. And selecting the speed and acceleration data of the vehicle in the transverse direction and the longitudinal direction, extracting the average value, the maximum value, the minimum value and the variance of the speed and the acceleration data of the vehicle in the transverse direction and the longitudinal direction in each time window as input features, and summing up 16-dimensional features. The steering behavior of the driver is smoother during steady and normal driving, but the steering behavior of the driver may be confused under fatigue, distraction and the like. Steering entropy is used to quantitatively indicate the steering characteristics of the driver to reflect steering smoothness and driving safety.
Heading angle in timing window is theta= (theta) 1 ,θ 2 ,θ 3 ,L,θ m ) M is the number of data in the time window. The next corner is predicted using taylor second-order expansion for a given time, as shown in the following equation:
error function definition:
in the method, in the process of the invention,is the nth time theta n Is a predicted value of (a). Setting an alpha value which is preferably 0.035; the prediction error of 90% of the data falls between-alpha and alpha. Dividing the prediction error interval into 9 segments, i.e. 9 intervals, and then finding the distribution frequency of each segment, i.e. p i I=1, 2, ·9. The steering entropy value SRE is calculated using:
c) Interaction characteristics between vehicles:
the target vehicle speed is noted as v e Longitudinal coordinate x e The speed of the front vehicle is recorded as v p Longitudinal coordinate x p . The THW headway, i.e. the train of vehicles travelling on the same lane, is the passage of two successive vehicle headway endsA time interval of a certain section; the DHW head space refers to the distance interval between two continuous vehicle head ends passing through a certain section in a vehicle queue running on the same lane; TTC collision time, namely the collision time of two vehicles when the rear vehicle and the front vehicle both run according to the current speed. The calculation formula of THW, DHW, TTC is as follows.
DHW=(x p -x e )
However, taking TTC as an example, two cases may occur when TTC is directly applied: when the speed of the rear vehicle is lower than that of the front vehicle, namely the rear vehicle can never catch up with the front vehicle at the current relative speed, and TTC is a negative value; there is a case that if the rear vehicle speed is only a little faster than the front vehicle speed, i.e., it takes a long time for the rear vehicle to catch up with the front vehicle at the current relative speed, TTC is a very large positive value. The TTC range is theoretically (- +. ++ infinity a) of the above-mentioned components, while the truly high risk TTC interval is very small. Taking TTC directly as input may result in a decrease of the model accuracy, and furthermore, the high risk TTC interval may be compressed again by the subsequent feature normalization process. Therefore, the reverse time to collision TTC is adopted -1 As shown in the following formula.
Furthermore, for negative TTCs, all negative TTCs are assigned to one sufficiently large positive TTC in a unified manner for simplicity of the feature -1 max (typically 50 seconds) that the target vehicle will collide with the lead vehicle after a long time. The method for eliminating and replacing the irrelevant values can reduce negative confusion influence on the model and improve training accuracy. The expression is as follows:
By processing, the TTC value is compressed to a small value between 0 and a small value relative to the safety, and the TTC value with high risk is amplified, and the value range is also amplified, so that the accuracy of the model is improved. In the same way, converting THW and DHW into inverse headway THW -1 Distance between head of inverse vehicle and DHW -1 To amplify the range of high risk THW, DHW. And finally, selecting the maximum value, the mean value and the variance as output characteristics, and taking the maximum value, the mean value and the variance as 9 dimensions.
When the lane change intention occurs, the driver's perception of the lane to be changed is very important. If the lane is not well perceived with the lane to be changed, the lane changing operation is directly carried out, and higher running risk can be brought. Therefore, the maximum TTC between the vehicles running in parallel on the lane to be changed and the front and rear vehicles of the lane to be changed is selected when the lane to be changed is intended -1 And THW -1 This 3-dimensional feature serves as an input.
D) Interaction between a vehicle and a road
The time length of the broken line is when the vehicle does not have the lane change intention, the time length of the solid line is when the vehicle is pressed, and the time length of the vehicle is when the vehicle is driven out of the solid line.
E) Local traffic situation descriptor
In order to better describe other vehicles, roads, obstacles, etc. running during the running process of the vehicle, corresponding descriptors are needed for description. First, factors such as a vehicle traveling around a target vehicle, an obstacle, and the like are considered. Local traffic density descriptors (LTDs) based on gaussian weights are presented. Taking the target vehicle as a reference, taking 8 vehicles, namely a front vehicle, a left front vehicle, a right front vehicle, a left vehicle, a right vehicle, a left rear vehicle, a right rear vehicle and a rear vehicle, as interested vehicles (shown in fig. 2, if the vehicles are not used, the vehicles are marked as 0), and calculating the contribution degree of the interested vehicles to the vehicle flow density of the target vehicle.
The specific calculation formula is as follows:
wherein X is j =(x j ,y j ) T For a vehicle of interest, x j ,y j The abscissa and ordinate of the vehicle of interest, respectively; mu= (x, y) T As the center coordinates of the target vehicle,wherein sigma x And sigma (sigma) y Defined by the formula:
σ x =|v x |+k 1 L
σ y =|v y |+k 2 W
wherein v is x ,v y For transverse and longitudinal speed, k, of the vehicle 1 And k is equal to 2 As a compensation factor, k is preferably 1 =0.625,k 2 =1.25。
Thus, each piece of data in each time window forms a corresponding local traffic density descriptor. In addition, the local velocity difference is solved by taking the local traffic density as a weight, and a local velocity difference descriptor is obtained and used for describing the running velocity difference (LVD) of the surrounding vehicles taking the target vehicle as a reference. Similarly, the Local Acceleration Difference (LAD) is solved. The following formula is shown:
v e for target vehicle speed, let us say, v j For the speed of the vehicle of interest, a e For target vehicle acceleration, a j For vehicle acceleration of interest, N i For the number of vehicles of interest, the maximum value is 8, and the number is reduced if no corresponding vehicle exists, as shown in fig. 2, and the minimum value is 0, namely, no vehicle of interest exists in one vehicle, namely, no surrounding vehicles exist.
As for the obstacle, it can be regarded as a vehicle whose running speed is zero. Finally, the 3 descriptors are respectively calculated to obtain statistical indexes such as mean value, maximum value, minimum value, standard deviation and the like, and the total is 12-dimensional characteristics.
The following table is a feature statistic:
table 1 statistical input features
Step 3: all the counted data are divided, no more than 5% of the data are randomly extracted from the data for labeling, and the rest data are used for non-labeling training and testing of semi-supervised learning. The risk score at the current moment is evaluated according to an upper-level value of 2-5% (i.e. the data exceeds 95-98% of the values of all data), and the labeling values are 0 (good), 1 (general), 2 (bad) and 3 (very bad). The label method is to label the data, average the final result and round it into integer.
Step 4: and constructing a neural network. The LSTM encoder-1 DCNN-LSTM decoder network architecture is constructed as shown. Wherein Convolume is 1D-CNN, maxPooling is the largest pooling layer, dropout is the discarding layer, FC is the fully connected layer, and Softmax is the activating layer. All other convolutional and fully-connected layer activation functions except the last layer are ReLU.
The method comprises the following steps: and constructing a neural network. An LSTM encoder-1 DCNN-LSTM decoder network architecture is constructed as shown. First, the statistics are embedded and the data is mapped to the LSTM encoder by an embedding layer. The number of embedded layer input nodes is the characteristic dimension, and the output is 128. The LSTM encoder has 128 concealment units in total. And finally, taking the tensor of the last hidden unit for activation and deformation to obtain a one-dimensional tensor. The tensor is subjected to three-dimensional convolution, activation, one-dimensional pooling and random discarding, and the tensor after convolution is obtained. The convolution process mainly extracts deep patterns among hidden features under different time sequences, and the patterns can better reflect risk information. The number of convolution channels for the three convolution layers is 64, 128, 256, respectively. After the output of the last layer is unfolded, the first branch is connected with the two full-connection layers and the Softmax layer to obtain the current risk score, and the number of hidden units is 128, 64 and 4 respectively. And the other branch is connected with an LSTM decoder to decode the current hidden characteristic to obtain the prediction of the future risk value, and the number of hidden units is 128 LSTM units of the decoder and the full connection layers 128, 64 and 4 respectively. The environment adopted by the computer is Win10, the software name Python3.7 is used, the deep learning framework is Keras2.2.4, and the background is Tensorflow1.14
The LSTM coder encodes the input, and adopts a Dropout method, namely, in the process of training a deep learning network, neural network units are randomly discarded from the network according to a certain probability to reduce the overfitting, and the probability is preferably 0.2; the following is a 1DCNN (one-dimensional convolutional neural network) to reduce the error caused by the convolution between the 2DCNN pair features. Features that contribute significantly to risk values can be selected in conjunction with maximum pooling Maxpooling. After three convolutions-pooling-dropouts, the first branch is connected with two full connection layers and a Softmax layer to obtain the current risk score. The other branch is connected to an LSTM decoder to decode the current advanced feature to obtain the prediction of future risk value.
Step 5: the above network is embedded in a semi-supervised learning architecture as shown in fig. 4. The unlabeled dataset is denoted as { X ] U }. Dividing the labeled dataset into a training set and a test set, wherein the training set is denoted as S L ={X L ,Y L }. First, use is made of a labeled dataset S L The network is pre-trained. In the training process, an Early-stop method is adopted, namely, when a certain monitoring value is iterated for a plurality of times, the network performance is not improvedWhen the time rises, the network automatically terminates and returns the best performing network weight before stopping. The monitored value selects loss, i.e. loss does not drop after passing the parity times. The choice of parameters is relatively large here because the network is allowed to learn a small amount of data adequately.
Then enter a self-learning stage to make the unlabeled dataset { X } U Generating pseudo tags { Y } using a pre-trained network P }. For each pseudo tag generated, there is a certain confidence epsilon (the pre-trained network is obtained directly after passing through the Softmax layer). The confidence is calculated and the threshold epsilon is calculated th Comparing, the set greater than the threshold is denoted as S P t ={X Uh t ,Y Ph t The set with (t being the number of iterations) less than the threshold is denoted as { Y } PL t And t is the iteration number. For the set { Y Ph t Based on manifold assumptions, the pseudo tag is considered to be a real tag. Will be set S L And S is equal to P t Merging to form a new set S L t And then is used for training of the network. For { X ] U t Pseudo tag regeneration using a retrained network (unlabeled dataset not accommodated by an iterative process). The process is shown in the following formula:
s represents a training data set, L represents a labeled, t represents the number of iterations, X represents an input feature, Y represents a label, U represents an unlabeled, P represents a pseudo-label, and h represents a value greater than a threshold ε th L represents less than a threshold epsilon th ,m st Is the total number of iterations. Correspondingly, S L t Is the training data set with labels at the t-th iteration. The rest is the same.
The number of iterations of the threshold decreases progressively as the value of i is smaller, because the lower the confidence level, the more likely noise is introduced. Similarly, the number of choice of parameters decreases with decreasing confidence.
Preferably, fine tuning of the model is also possible. The trimming process sets all CNN and LSTM layers to be untrainable, trimming only the fully connected layers. The reason is that the feature extraction layer is trained well, and potential features of the unlabeled dataset are learned, only the weights of the full connection layer need to be adjusted. For the data set { X } which is not yet labeled in the last stage of self-learning U mst Prediction is performed with the predicted value directly as its real label. After the completion, all the labels without label data are obtained and the trained network model is obtained.
Step 6: a neural network loss function is set. The performance of the network is degraded due to class imbalance, i.e. high risk data is always far less than risk free data. The loss function is set to compensate for class imbalance. A multi-class cross entropy loss function (CE) is selected as the basic loss function as the multi-class loss function, as shown in the following equation:
wherein E is yi:P As random variable y with distribution P i At y i,k Mathematical expectation under distribution P; n is the number of data in mini-batch, m is the category number, y ik Is true value +.>Is a predicted value. Adding an over-weight (OBC) to the above-mentioned loss function, namely:
wherein, the liquid crystal display device comprises a liquid crystal display device,
however, in the course of constantly iterating semi-supervised learning, the amount of tagged data will constantly change, but each category will not change uniformly. It is necessary to adapt this to the following conditions:
N(t,k)=|S L t |
AOBC(0,k)=OBC(k)
wherein n= |s in OBC (k) L 0 |。
The expression I represents the number of elements in the collection; AOBC (t, k) is the weight lost by the kth class as a function of the number of iterations t.
Preferably, step 7: and adding a limit value punishment module to the network. In order to compensate for very specific severe classification errors of the neural network, a limit value penalty module needs to be introduced. The module employs fuzzy logic, with higher penalties being based as certain values are closer to the parameters of the moment of collision. The following formula is shown:
wherein ev is the limit value; data exceeding 99.99% of all values are generally taken as limit values.
Preferably, step 8: the network is pre-trained, self-learned and fine-tuned. The network optimizer is selected as an Adam optimizer, and the learning rate is 10 -3 Attenuation of 10 -6 ,ε th Selected to be 0.999999, 0.99999, 0.9999, 0.999, 0.99, 0.95, 0.9. And after fine tuning, obtaining all the data labels and the trained network. Each label output corresponds to a confidence level, and the current risk score can be obtained by multiplying the confidence level by the values of all labels. The trained network can then be used directly (i.e., to evaluate new data without semi-supervision). As shown in the overall frame of fig. 5.
Although embodiments of the present invention have been disclosed above, it is not limited to the details and embodiments shown and described, it is well suited to various fields of use for which the invention would be readily apparent to those skilled in the art, and accordingly, the invention is not limited to the specific details and illustrations shown and described herein, without departing from the general concepts defined in the claims and their equivalents.
Claims (3)
1. A vehicle running risk prediction method based on a neural network model is characterized in that,
step one: collecting vehicle running data to form vehicle history data;
step two: extracting features of the vehicle history data by adopting a context time window to form statistical features;
step three: extracting the statistical features, including: the type of vehicle, length and width of the vehicle; steering entropy value; parameter time to reverse collision TTC -1 Time interval THW of inverse head -1 Distance between head of inverse vehicle and DHW -1 The method comprises the steps of carrying out a first treatment on the surface of the The time length of the broken line is when the vehicle does not have the lane change intention, the time length of the solid line is when the vehicle is pressed, and the time length of the vehicle is when the vehicle is driven out of the solid line; local traffic flow density, local speed difference and local acceleration difference;
step four: dividing the extracted result data in the third step, randomly extracting no more than 5% of data from the data to carry out labeling, and forming a labeled data set; the rest data are label-free data sets, and are used for label-free training and testing of semi-supervised learning;
step five: constructing a neural network; constructing an LSTM encoder-1 DCNN-LSTM decoder network architecture;
step six: the unlabeled dataset is denoted as { X ] U -a }; dividing the labeled dataset into a training set and a test set, wherein the training set is denoted as S L ={X L ,Y L -a }; using labeled dataset S L Pre-training the neural network;
then enter a self-learning stage to make the unlabeled dataset { X } U Generating pseudo tags { Y } using a pre-trained network P -a }; for each generated pseudo tag, a certain confidence epsilon is carried out, and the confidence epsilon is matched with a threshold epsilon th Comparing, the set greater than the threshold is denoted as S P t ={X Uh t ,Y Ph t Sets smaller than the threshold are denoted Y Pl t T is the iteration number; for the set { Y Ph t -considering the pseudo tag, i.e. the real tag, according to manifold assumptions; will be set S L And S is equal to P t Merging to form a new set S L t Then the training method is used for training the network; for { X ] U t Performing pseudo tag regeneration by using the retrained network; for the data set { X } which is not yet labeled in the last stage of self-learning U mst Predicting to take the predicted value as the real label; after the completion, all labels without label data and a trained network model are obtained; predicting the running risk of the vehicle according to the trained network model; s represents a training data set, L represents a labeled, t represents the number of iterations, X represents an input feature, Y represents a label, U represents an unlabeled, P represents a pseudo-label, and h represents a value greater than a threshold ε th L represents less than a threshold epsilon th ,m st Is the total iteration number; correspondingly, S L t A training data set with a label at the t-th iteration;
the steering entropy value SRE:
dividing the prediction error interval into 9 segments, i.e. 9 intervals, and then obtaining the distribution frequency of each segment, i.e. p i ,i=1,2,..9;
The time to reverse collision TTC -1
TTC collision time, namely the collision time of two vehicles when the rear vehicle and the front vehicle both run at the current speed; the TTC of all negative values is uniformly assigned to a sufficiently large TTC of positive value -1 max ;TTC -1 max Taking 50 seconds;
The THW headway refers to the time interval that two continuous vehicle headway ends pass through a certain section in a vehicle queue running on the same lane; the DHW head space refers to the distance interval between two continuous vehicle head ends passing through a certain section in a vehicle queue running on the same lane;
the specific calculation formula of the local traffic flow density is as follows:
wherein X is j =(x j ,y j ) T For a vehicle of interest, μ= (x, y) T As the center coordinates of the target vehicle,wherein sigma x And sigma (sigma) y Defined by the formula:
σ x =|v x |+k 1 L
σ y =|v y |+k 2 W
wherein v is x ,v y For transverse and longitudinal speed, k, of the vehicle 1 And k is equal to 2 Is a compensation factor; length L and width W of the vehicle; n (N) i Is the number of vehicles of interest; x is x j ,y j The abscissa and ordinate of the vehicle of interest, respectively;
the local acceleration difference is calculated as follows:
the local velocity difference is calculated as follows:
2. The vehicle running risk prediction method based on a neural network model according to claim 1, wherein the penalty function of the neural network is:
wherein the probability mass function f of the distribution P can be defined as
AOBC(0,k)=OBC(k)
Wherein N (t, k) = |s in AOBC (k) L,k t I, since AOBC (0, k) =obc (k), N (0, k) = |s in OBC (k) L,k 0 I (I); n is the number of data in mini-batch, m is the category number, y ik To be a true value of the value,is a predicted value; iteration times t; in the above expression, || represents the number of elements in the collection.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011076552.2A CN112215487B (en) | 2020-10-10 | 2020-10-10 | Vehicle running risk prediction method based on neural network model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011076552.2A CN112215487B (en) | 2020-10-10 | 2020-10-10 | Vehicle running risk prediction method based on neural network model |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112215487A CN112215487A (en) | 2021-01-12 |
CN112215487B true CN112215487B (en) | 2023-05-23 |
Family
ID=74052983
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011076552.2A Active CN112215487B (en) | 2020-10-10 | 2020-10-10 | Vehicle running risk prediction method based on neural network model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112215487B (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113033902B (en) * | 2021-03-31 | 2024-03-19 | 中汽院智能网联科技有限公司 | Automatic driving lane change track planning method based on improved deep learning |
CN112849156B (en) * | 2021-04-25 | 2021-07-30 | 北京三快在线科技有限公司 | Driving risk identification method and device |
CN113222609B (en) * | 2021-05-07 | 2022-05-06 | 支付宝(杭州)信息技术有限公司 | Risk identification method and device |
CN113436432A (en) * | 2021-05-24 | 2021-09-24 | 武汉理工大学 | Method for predicting short-term traffic risk of road section by using road side observation data |
CN113326825A (en) * | 2021-08-03 | 2021-08-31 | 新石器慧通(北京)科技有限公司 | Pseudo tag generation method and device, electronic equipment and storage medium |
CN113326826A (en) * | 2021-08-03 | 2021-08-31 | 新石器慧通(北京)科技有限公司 | Network model training method and device, electronic equipment and storage medium |
CN114333416A (en) * | 2021-12-24 | 2022-04-12 | 阿波罗智能技术(北京)有限公司 | Vehicle risk early warning method and device based on neural network and automatic driving vehicle |
CN114397474B (en) * | 2022-01-17 | 2022-11-08 | 吉林大学 | FCN-MLP-based arc ultrasonic sensing array wind parameter measurement method |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107215335A (en) * | 2017-06-01 | 2017-09-29 | 北京交通大学 | Traffic safety risk feedback early warning system and method for early warning based on microcosmic driving |
CN109094482A (en) * | 2018-07-25 | 2018-12-28 | 哈尔滨工业大学 | Vehicle operation risk assessment information acquisition system, risk evaluating system and method |
CN109933619A (en) * | 2019-03-13 | 2019-06-25 | 西南交通大学 | A kind of semisupervised classification prediction technique |
CN109952583A (en) * | 2016-11-15 | 2019-06-28 | 谷歌有限责任公司 | The semi-supervised training of neural network |
CN110610271A (en) * | 2019-09-17 | 2019-12-24 | 北京理工大学 | Multi-vehicle track prediction method based on long and short memory network |
CN111368879A (en) * | 2020-02-17 | 2020-07-03 | 南京邮电大学 | Track data mining method based on deep semi-supervised neural network |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10423861B2 (en) * | 2017-10-16 | 2019-09-24 | Illumina, Inc. | Deep learning-based techniques for training deep convolutional neural networks |
-
2020
- 2020-10-10 CN CN202011076552.2A patent/CN112215487B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109952583A (en) * | 2016-11-15 | 2019-06-28 | 谷歌有限责任公司 | The semi-supervised training of neural network |
CN107215335A (en) * | 2017-06-01 | 2017-09-29 | 北京交通大学 | Traffic safety risk feedback early warning system and method for early warning based on microcosmic driving |
CN109094482A (en) * | 2018-07-25 | 2018-12-28 | 哈尔滨工业大学 | Vehicle operation risk assessment information acquisition system, risk evaluating system and method |
CN109933619A (en) * | 2019-03-13 | 2019-06-25 | 西南交通大学 | A kind of semisupervised classification prediction technique |
CN110610271A (en) * | 2019-09-17 | 2019-12-24 | 北京理工大学 | Multi-vehicle track prediction method based on long and short memory network |
CN111368879A (en) * | 2020-02-17 | 2020-07-03 | 南京邮电大学 | Track data mining method based on deep semi-supervised neural network |
Non-Patent Citations (2)
Title |
---|
Cost-sensitive semi-supervised deep learning to assess driving risk by application of naturalistic vehicle trajectories;Hongyu Hu等;《Expert Systems With Applications》;20210420;第1-13页 * |
Dong-Hyun Lee.Pseudo-Label : The Simple and Efficient Semi-Supervised Learning Method for Deep Neural Networks.《ICML 2013 Workshop : Challenges in Representation Learning(WREPL)》.2015, * |
Also Published As
Publication number | Publication date |
---|---|
CN112215487A (en) | 2021-01-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112215487B (en) | Vehicle running risk prediction method based on neural network model | |
Islam et al. | Crash data augmentation using variational autoencoder | |
CN110097755B (en) | Highway traffic flow state identification method based on deep neural network | |
Basavaraju et al. | A machine learning approach to road surface anomaly assessment using smartphone sensors | |
Das et al. | Detecting lane change maneuvers using SHRP2 naturalistic driving data: A comparative study machine learning techniques | |
Chen et al. | Dangerous driving behavior detection using video-extracted vehicle trajectory histograms | |
Lyu et al. | Using naturalistic driving data to identify driving style based on longitudinal driving operation conditions | |
CN113744526B (en) | Highway risk prediction method based on LSTM and BF | |
CN110120218A (en) | Expressway oversize vehicle recognition methods based on GMM-HMM | |
Hu et al. | Efficient mapping of crash risk at intersections with connected vehicle data and deep learning models | |
CN111114556A (en) | Lane change intention identification method based on LSTM under multi-source exponential weighting loss | |
Shi et al. | Real-time driving risk assessment using deep learning with XGBoost | |
Xue et al. | A context-aware framework for risky driving behavior evaluation based on trajectory data | |
Cummings et al. | Subjectivity in the creation of machine learning models | |
CN114882069A (en) | Taxi track abnormity detection method based on LSTM network and attention mechanism | |
Harkous et al. | A two-stage machine learning method for highly-accurate drunk driving detection | |
US20170316310A1 (en) | System and method for decoding spiking reservoirs with continuous synaptic plasticity | |
Jadhav et al. | Road accident analysis and prediction of accident severity using machine learning | |
Sun et al. | Vision-based traffic conflict detection using trajectory learning and prediction | |
Hao et al. | Aggressive lane-change analysis closing to intersection based on UAV video and deep learning | |
CN115981302A (en) | Vehicle following lane change behavior decision-making method and device and electronic equipment | |
Raju | EXPERIENCES OF SUPERVISED MACHINE LEARNING TO REPLICATE DRIVING 1 BEHAVIOR UNDER TRAFFIC ENVIRONMENT 2 | |
Raju et al. | Replicating driving behavior under heterogeneous non-lane-based traffic environment: A supervised machine learning based methodology | |
DARSONO et al. | Utilizing LSTM Networks for the Prediction of Driver Behavior. | |
CN114283575B (en) | Signal lamp intersection queuing length estimation method based on video monitoring data |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |