CN112622886B - Anti-collision early warning method for heavy operation vehicle comprehensively considering front and rear obstacles - Google Patents
Anti-collision early warning method for heavy operation vehicle comprehensively considering front and rear obstacles Download PDFInfo
- Publication number
- CN112622886B CN112622886B CN202011512720.8A CN202011512720A CN112622886B CN 112622886 B CN112622886 B CN 112622886B CN 202011512720 A CN202011512720 A CN 202011512720A CN 112622886 B CN112622886 B CN 112622886B
- Authority
- CN
- China
- Prior art keywords
- driving
- vehicle
- collision
- network
- function
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 67
- 230000006399 behavior Effects 0.000 claims abstract description 34
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 24
- 230000008569 process Effects 0.000 claims abstract description 17
- 230000002787 reinforcement Effects 0.000 claims abstract description 14
- 238000004088 simulation Methods 0.000 claims abstract description 12
- 230000006870 function Effects 0.000 claims description 123
- 230000009471 action Effects 0.000 claims description 34
- 230000033001 locomotion Effects 0.000 claims description 33
- 230000001133 acceleration Effects 0.000 claims description 21
- 238000012549 training Methods 0.000 claims description 15
- 238000005457 optimization Methods 0.000 claims description 12
- 210000002569 neuron Anatomy 0.000 claims description 9
- 230000004913 activation Effects 0.000 claims description 6
- 238000013528 artificial neural network Methods 0.000 claims description 6
- 238000005259 measurement Methods 0.000 claims description 6
- 238000010606 normalization Methods 0.000 claims description 6
- 238000005192 partition Methods 0.000 claims description 6
- 238000011156 evaluation Methods 0.000 claims description 4
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 claims description 3
- 238000013461 design Methods 0.000 claims description 3
- 238000011478 gradient descent method Methods 0.000 claims description 3
- 239000011159 matrix material Substances 0.000 claims description 3
- 239000011541 reaction mixture Substances 0.000 claims description 3
- 238000012360 testing method Methods 0.000 claims description 3
- 230000003044 adaptive effect Effects 0.000 abstract description 3
- 230000006378 damage Effects 0.000 description 5
- 238000011160 research Methods 0.000 description 5
- 206010039203 Road traffic accident Diseases 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000002485 combustion reaction Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000004880 explosion Methods 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 208000027418 Wounds and injury Diseases 0.000 description 1
- 230000002411 adverse Effects 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000003912 environmental pollution Methods 0.000 description 1
- 208000014674 injury Diseases 0.000 description 1
- 239000007788 liquid Substances 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 238000011158 quantitative evaluation Methods 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
- B60W30/08—Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/02—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to ambient conditions
- B60W40/06—Road conditions
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/08—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
- B60W40/09—Driving style or behaviour
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/10—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to vehicle motion
Landscapes
- Engineering & Computer Science (AREA)
- Automation & Control Theory (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Control Of Driving Devices And Active Controlling Of Vehicle (AREA)
Abstract
The invention discloses an anti-collision early warning method for a heavy operation vehicle, which comprehensively considers front and rear obstacles. Firstly, aiming at the road running environment of China, a driving simulation platform is built, and typical driving behaviors of excellent drivers under various running conditions are collected. Secondly, a reverse reinforcement learning algorithm based on the maximum entropy is introduced to learn the driving behavior of a good driver. And finally, describing the anti-collision early warning strategy problem as a Markov decision process, and establishing an anti-collision driving decision model based on forward reinforcement learning to obtain an accurate, reliable and adaptive anti-collision early warning strategy. The method provided by the invention comprehensively considers the influence of forward and backward obstacles on vehicle collision, provides accurate and quantized driving suggestions such as throttle opening, steering wheel angle control quantity and the like for a driver, can adapt to different driving conditions and driver operation, and solves the problem that the existing anti-collision early warning method for heavy commercial vehicles is lack of accuracy and adaptability.
Description
Technical Field
The invention relates to a vehicle anti-collision early warning strategy, in particular to a heavy-duty operation vehicle anti-collision early warning method comprehensively considering front and rear obstacles, and belongs to the technical field of automobile safety.
Background
The safety condition of the commercial vehicle, which is a main undertaker of road transportation, directly influences the safety of road transportation. Different from small passenger vehicles, most of the vehicles for operation, transport and transportation are large and medium-sized vehicles, and the vehicle has the characteristics of high centroid position, large overall dimension, large total mass and the like, and has the advantages of high vehicle operation intensity, long operation time and complex operation environment. In case of traffic accidents in the transportation process, serious consequences such as group death and group injury, cargo falling, combustion, explosion and the like are easily caused, and adverse effects such as property loss, environmental pollution, ecological damage and the like are caused.
Relevant researches show that the collision is the most main accident form in the road transportation process, the proportion of the forward collision in the collision accident is the largest, and particularly, the collision accidents on the expressway are mostly forward collisions. Although the occurrence frequency of the backward collision is relatively low, for heavy operation vehicles represented by dangerous goods transport tank cars, the backward collision can easily cause the damage of the tank body, further cause the leakage, even the combustion and the explosion of dangerous goods in the tank, and generate secondary damage far exceeding the damage caused by the accident itself, so that the vehicle has higher danger. Relevant statistical data from the U.S. highway traffic safety administration indicate that vehicle collision events can be reduced by about 30% to 60% if early warning prompts can be given to the driver and 0.5 second preprocessing time is added before the collision event occurs. Therefore, the accurate and reliable front and back anti-collision early warning strategy of the heavy operation vehicle is researched, and the method has important effects on improving the transportation safety guarantee capability of dangerous goods and improving the road traffic safety.
At present, many patents and documents are provided for studying collision avoidance warning strategies for vehicles, but most of them are directed to small passenger vehicles. Compared with a passenger vehicle, the heavy commercial vehicle has the characteristics of higher centroid position, larger load capacity and the like, so that the braking distance is longer, the side-tipping stability is poorer, in the emergency braking or lane-changing process, the instability of the vehicle can be further increased due to the liquid in the tank or the cargo on the trailer shaking, and the vehicle is extremely easy to tip over due to instability. Therefore, the anti-collision early warning strategy for the passenger vehicle is difficult to be applied to heavy commercial vehicles.
In the research of anti-collision early warning strategies for heavy-duty operating vehicles, the classification early warning prompt is only carried out aiming at the collision danger degree in a single direction such as the front direction or the rear direction at present, and the influence of factors such as driver operation and running conditions on vehicle collision is not considered. Although the existing method can play a certain early warning role, the problems of poor adaptability to different driving conditions and inaccurate early warning exist, and the method is difficult to adapt to complicated and variable traffic environments and vehicle driving conditions with fluctuation differences. In addition, the existing method mainly adopts the forms of sound, light and the like to carry out early warning prompt, and does not relate to the research of anti-collision early warning strategies for providing specific driving suggestions such as driving speed, driving tracks and the like, and is lack of accuracy and reliability.
Generally, the current anti-collision early warning strategy research aiming at the heavy operation vehicle still has great defects in the aspects of accuracy, adaptability and the like, and the anti-collision early warning strategy research of the heavy operation vehicle which is accurate, reliable and self-adaptive to the operation and running working conditions of a driver is lacked.
Disclosure of Invention
The purpose of the invention is as follows: the invention discloses an anti-collision early warning method for a heavy commercial vehicle, which comprehensively considers front and rear obstacles and aims at solving the problem that the anti-collision early warning method for the heavy commercial vehicle lacks accuracy and adaptability. The method can provide accurate and quantized driving suggestions such as the opening degree of a throttle valve and the steering wheel angle control quantity for a driver, can adapt to different driving conditions and driver operation, and improves the accuracy and the adaptability of the anti-collision early warning method for the heavy-duty commercial vehicle.
The technical scheme is as follows: the invention provides an anti-collision early warning strategy comprehensively considering front and rear obstacles aiming at heavy operation vehicles such as a semi-trailer tank car and a semi-trailer train. Firstly, aiming at the road running environment of China, a driving simulation platform is built, and typical driving behaviors of excellent drivers under various running conditions are collected. Secondly, a reverse reinforcement learning algorithm based on the maximum entropy is introduced to learn the driving behavior of a good driver. And finally, describing the anti-collision early warning strategy problem as a Markov decision process, and establishing an anti-collision driving decision model based on forward reinforcement learning to obtain an accurate, reliable and adaptive anti-collision early warning method. The method comprises the following steps:
the method comprises the following steps: building driving simulation platform
In order to reduce the occurrence frequency of traffic accidents caused by vehicle collision and improve the safety of heavy commercial vehicles, the invention provides an anti-collision early warning strategy, which is applicable to the following scenes: in the process of running of a heavy-duty operation vehicle, obstacles exist in front of and behind the vehicle, and in order to prevent collision with surrounding vehicles, decision strategies such as acceleration, deceleration, steering and the like are effectively and timely provided for a driver so as to avoid collision accidents.
According to the scene described above, a driving simulation platform is built, and the driving behavior of an excellent driver in a real driving environment is collected. The method specifically comprises the following steps:
firstly, a Prescan-based driving simulation platform is built, a town virtual environment model comprising a straight road and a curve road is built according to the Chinese road driving environment, and a driver controls a heavy operation vehicle to move through a driving simulator.
Secondly, a centimeter-level high-precision differential GPS, an inertia measurement unit and a millimeter wave radar are installed on the heavy operation vehicle to obtain accurate motion state information and relative motion state information of the vehicle, wherein the accurate motion state information and the relative motion state information specifically comprise position, speed, yaw angle, acceleration, relative speed and relative distance. Meanwhile, the control information of the driver is obtained by utilizing a vehicle body CAN bus, and the control information comprises the pressure of a brake pedal, the steering wheel angle and the opening degree of a throttle valve.
And finally, 6 driving conditions of lane changing, lane keeping, vehicle following, constant speed, acceleration and deceleration are designed, 30 excellent drivers with different ages and driving styles are selected to perform a data acquisition test, data acquisition of various typical driving behaviors of the excellent drivers is realized under a space-time global unified coordinate system, and a driving database of the excellent drivers is constructed.
In the present invention, the front vehicle means a vehicle located in front of the road on which the heavy-duty vehicle travels, located within the same lane line, and having the same traveling direction. The rear vehicle is a vehicle which is positioned behind the driving road of the heavy operation vehicle, is positioned in the same lane line and has the same driving direction.
Step two: learning driving behavior of human excellent driver
In order to improve the adaptability of the anti-collision early warning strategy, the invention introduces a reverse reinforcement learning algorithm based on the maximum entropy to learn the driving behaviors of the excellent driver collected in the step one under different driving conditions.
In an actual traffic scene, the driving behavior of an excellent driver is not easy to express explicitly, but it is relatively easy to acquire a driving track generated by the excellent driving behavior. Considering that the driving track of the excellent driver has the maximum reward value in all possible tracks, the driving behavior of the excellent driver is represented by the reward function.
First, a reward function for the excellent driver's driving trajectory is established:
in the formula (1), xiiRepresents the travel locus of the ith excellent driver, and xii={(S1,A1),(S2,A2),...,(Sm,Am) M represents the number of driving tracks of excellent drivers collected, rθ(ξi) Feature vector representing the ith excellent driver's driving track, i.e. reward function for this driving track, rθ(Si,Ai) Reward value, S, representing the ith "state-action" in this trackiIndicating the state at time i, AiIndicating the operation at time i.
Considering that an excellent driver often makes driving decisions according to variables such as running speed, yaw angle, distance from a lane line, distance from front and rear obstacles, and the like, the present invention linearly fits a reward value using longitudinal speed, lateral speed, yaw angle, and distance from front and rear obstacles.
rθ(Si,Ai)=rθ(φ1,φ2,φ3,φ4)=θrT·φ (2)
In the formula (2), the characteristic value phi1=vsx cosψs,φ2=vsy sinψs,φ3=dsf-d0,φ4=dsr-d0,vsx,vsyRespectively, the lateral and longitudinal speeds of a heavy commercial vehicle in meters per second, psisIs the yaw angle in degrees, dsf,dsrRespectively represents the relative distance between the heavy operation vehicle and the front vehicle and the rear vehicle, and the unit is meter and thetarTPhi represents the fitted eigenvalue for the coefficient matrix.
The probability of a trajectory having the maximum entropy can be expressed as:
in the formula (3), p (xi)i| θ) represents the probability of the trace having the maximum entropy, Z (θ) is a partition function, and representing a policy nt-1And n represents the number of sampling tracks in the current strategy.
Secondly, establishing a probability model of the driving track of the excellent driver, and solving the driving track with the maximum entropy by using the maximum information entropy principle, wherein the formula is shown as (4):
in the formula (4), the reaction mixture is,representing the collected driving track of the excellent driver.
Converting equation (4) into a product by using a Lagrange multiplier method:
in the formula (5), J (θ) is a loss function.
Considering that the greater the probability of occurrence of the driving trajectory of the excellent driver, the more the reward function expresses the driving behavior of the excellent driver, equation (5) is described as:
minimizing the reward function by utilizing a gradient descent method to obtain a global optimal solution of the reward function:
and finally, optimizing the parameters of the reward function by using a gradient descent algorithm, and further learning the global optimal solution of the reward function. According to the optimized parameter thetarThe current reward function r can be outputθ(Si,Ai) I.e. a function characterizing the excellent driver driving behavior.
Step three: establishing an anti-collision driving decision model
According to the invention, a DDPG algorithm is adopted, and an anti-collision driving decision model is established based on the driving behavior of the excellent driver collected in the step one and the excellent driving strategy obtained in the step two, so that anti-collision early warning strategies under different driver operation and driving conditions are researched. The method specifically comprises the following 4 sub-steps:
substep 1: defining basic parameters for an anti-collision driving decision model
Considering that the future motion state of the heavy-duty operating vehicle is influenced by the current motion state and the current action at the same time, the anti-collision driving decision problem is modeled into a Markov decision process, and the basic parameters of the model are defined: state S at time ttState S at time t +1t+1Action A at time ttAnd action AtCorresponding return value Rt(ii) a In particular to:
(1) Defining a state space
The running safety of a heavy-duty vehicle is related to not only the motion state of the vehicle itself but also the relative motion state of the front and rear obstacles. Therefore, using the motion state information obtained in step one, a state space is defined:
St=(vsx,vsy,vsf,vsr,asx,asy,dsf,dsr,ωs,θs,δbr,δthr) (8)
in the formula (8), vsf,vsrRespectively representing the relative speeds of the heavy commercial vehicle, a front vehicle and a rear vehicle, and the unit is meter per second; a issx,asyRespectively representing the transverse acceleration and the longitudinal acceleration of the heavy commercial vehicle, and the unit is meter per second of square; omegasThe yaw rate of the vehicle is expressed in radians per second; thetasIs the steering wheel angle of the vehicle, in degrees, deltabr,δthrRespectively represents the opening degree of a brake pedal and the opening degree of a throttle valve of the vehicle, and the unit is percentage.
(2) Defining action decisions
In order to establish a more accurate and reliable anti-collision early warning strategy, the invention considers the transverse motion and the longitudinal motion of the vehicle, simultaneously considers that the control quantities of a throttle valve and a brake pedal of the vehicle can not appear simultaneously, takes the steering wheel angle and the accelerating/braking normalization quantity as the control quantities, and defines the early warning strategy output by a decision model, namely an action decision At=[θstr_out,δs_out]。
Wherein A istFor action decision at time t, θstr_outRepresents the normalized steering wheel angle control quantity in the range of [ -1,1],δs_outRepresents an acceleration/braking normalization quantity in the range of-1, 1]. When deltatsuo_When the value is 0, the vehicle is moving at a constant speed, and when the value is deltas_outWhen-1, the vehicle is braked at the maximum deceleration, and when δs_outWhen 1, the vehicle is accelerated at the maximum acceleration.
(3) Defining a reward function
Defining the reward function as:
Rt=r1+r2+r3 (9)
in the formula (9), RtFor a reward function at time t, r1For a safety distance reward function, r2As a comfort reward function, r3Is a penalty function.
First, in order to prevent a collision of a vehicle, a safe distance reward function r is designed1:
In the formula (10), d0A safe distance threshold.
Secondly, in order to ensure the driving comfort of the vehicle, the excessive impact degree should be avoided as much as possible, and a comfort rewarding function r is designed2=|asy(t+1)-asy(t)|。
Finally, in order to judge the error action of the vehicle, a penalty function r is designed3:
In the formula (11), SpenFor penalty, in the present invention, take SpenThe decision model will get a penalty of-100 when the vehicle crashes or rolls over.
Substep 2: network architecture for building anti-collision decision model
And (3) constructing an anti-collision driving decision network by using a strategy-evaluation network framework, wherein the anti-collision driving decision network comprises a strategy network and a value function network. Wherein a policy network is used for the pair state StAnd regressing the features to output a continuous action at(ii) a Value function network for receiving state StAnd action AtTo evaluate the value of the current "state-action". Specifically, the method comprises the following steps:
(1) designing a policy network
Establishing a strategy network by utilizing a plurality of neural networks with full connection layer structures; the normalized state space StIn turn with the full-link layer F1Full connection layer F2And a full joint layer F3Connected to obtain an output O1Immediate action decision At;
Considering that the dimension of the state space is 12, the number of neurons of the state input layer is set to 12. The activation function of each fully-connected layer is a Linear rectification Unit (ReLU) with the expression F (x) max (0, x), and the fully-connected layer F1、F2、F3The number of neurons in (A) is 20, 20, 10, respectively.
(2) Design value function network
Establishing a value function network by utilizing a plurality of neural networks with full connection layer structures; the normalized state quantity StAnd action AtIn turn with the full-link layer F4Full connection layer F5And a full joint layer F6Connecting to obtain an output O2, namely a Q value; the activation function of each fully-connected layer is ReLU, and the fully-connected layer F4、F5、F6The number of neurons in (A) is 20, 20, 10, respectively.
Substep 3: training strategy network and value function network
The strategy network and the value function network have respective network parameters, and the network parameters of the two parts are updated during training iteration, so that the network converges to obtain a better result. The specific training updating step comprises the following steps:
Substep 3.2: establishing a reward function using equation (2) and initializing a value function network parameter θQPolicy network parameter θμAnd a parameter thetar;
Substep 3.3: taking the formula (9) as an initial strategy optimization target, and performing strategy optimization by using a DDPG algorithm to obtain an initial strategyπ0;
Substep 3.4: performing iterative solution, each iteration comprising substep 3.41 to substep 3.45, in particular:
substep 3.43: optimizing reward function parameters using stochastic gradient descent algorithm minimization equation (7)
Substep 3.44: the optimized reward function rθ(Si,Ai) As an optimization target, the DDPG algorithm is used for strategy optimization, and the value function network parameter theta is updatedQAnd a policy network parameter θμ;
Substep 3.45: and calculating the updating amplitude of the reward function, wherein when the updating amplitude of the reward function is smaller than a given threshold, the reward function at the moment is the optimal reward function.
Substep 3.5: and (4) performing iterative updating according to the method provided by the substep 3.4, so that the strategy network and the value function network are gradually converged. In the training process, if the vehicle collides or turns over, the current round is stopped and a new round is started for training. And when the heavy-duty operation vehicle stably and effectively avoids vehicle collision by using the decision strategy output by the model, the iteration is finished.
Substep 4: outputting an anti-collision early warning strategy by using an anti-collision driving decision model
The information collected by the centimeter-level high-precision differential GPS, the inertial measurement unit, the millimeter wave radar and other sensors is input into the trained anti-collision driving decision network, so that reasonable steering wheel turning angle and throttle opening degree commands can be output in real time, accurate, quantitative and reliable driving suggestions are provided for drivers, and the anti-collision early warning strategy output of the heavy-duty operation vehicle, which is accurate, reliable and self-adaptive to driver operation and driving conditions, is realized.
Has the advantages that: compared with a general vehicle anti-collision early warning strategy, the method provided by the invention has the characteristics of more accuracy, reliability and self-adaption, and is specifically embodied as follows:
(1) the method provided by the invention comprehensively considers the influence of forward and backward obstacles on vehicle collision, accurately quantifies driving strategies such as driving speed, steering of a steering wheel and the like in a numerical form, and realizes accurate and reliable anti-collision early warning decision of heavy commercial vehicles.
(2) The method provided by the invention can adapt to different driver operations and driving conditions, the output driving strategy can be adaptively adjusted according to the driver operations and the driving condition changes, and the problem that the existing anti-collision early warning strategy for heavy-duty operation vehicles is lack of accuracy and adaptability is solved.
(3) The method provided by the invention does not need complex vehicle dynamics modeling, and the calculation method is simple and clear.
Drawings
FIG. 1 is a schematic diagram of a technical route of the present invention;
fig. 2 is a schematic diagram of a network architecture of an anti-collision driving decision model established by the present invention.
Detailed Description
The technical scheme of the invention is further explained by combining the attached drawings.
In order to establish an accurate, reliable and self-adaptive anti-collision early warning strategy for driver operation and driving conditions, the invention provides an anti-collision early warning strategy which comprehensively considers front and rear obstacles aiming at heavy operation vehicles such as semi-trailer trains and semi-trailer tank trucks. Firstly, aiming at the road running environment of China, a driving simulation platform is built, and typical driving behaviors of excellent drivers under various running conditions are collected. Secondly, a reverse reinforcement learning algorithm based on the maximum entropy is introduced to learn the driving behavior of a good driver. And finally, describing the anti-collision early warning strategy problem as a Markov decision process, and establishing an anti-collision driving decision model based on forward reinforcement learning to obtain an accurate, reliable and adaptive anti-collision early warning strategy. The technical route of the invention is shown in figure 1, and the specific steps are as follows:
the method comprises the following steps: building driving simulation platform
In order to reduce the occurrence frequency of traffic accidents caused by vehicle collision and improve the safety of heavy commercial vehicles, the invention provides an anti-collision early warning strategy, which is applicable to the following scenes: in the process of running of a heavy-duty operation vehicle, obstacles exist in front of and behind the vehicle, and in order to prevent collision with surrounding vehicles, decision strategies such as acceleration, deceleration, steering and the like are effectively and timely provided for a driver so as to avoid collision accidents.
According to the scene described above, a driving simulation platform is built, and the driving behavior of an excellent driver in a real driving environment is collected. The method specifically comprises the following steps:
firstly, a Prescan-based driving simulation platform is built, a town virtual environment model comprising a straight road and a curve road is built according to the Chinese road driving environment, and a driver controls a heavy operation vehicle to move through a driving simulator.
Secondly, a centimeter-level high-precision differential GPS, an inertia measurement unit and a millimeter wave radar are installed on the heavy operation vehicle to obtain accurate motion state information and relative motion state information of the vehicle, wherein the accurate motion state information and the relative motion state information specifically comprise position, speed, yaw angle, acceleration, relative speed and relative distance. Meanwhile, the control information of the driver is obtained by utilizing a vehicle body CAN bus, and the control information comprises the pressure of a brake pedal, the steering wheel angle and the opening degree of a throttle valve.
And finally, 6 driving conditions of lane changing, lane keeping, vehicle following, constant speed, acceleration and deceleration are designed, 30 excellent drivers with different ages and driving styles are selected to perform a data acquisition test, data acquisition of various typical driving behaviors of the excellent drivers is realized under a space-time global unified coordinate system, and a driving database of the excellent drivers is constructed.
In the present invention, the front vehicle means a vehicle located in front of the road on which the heavy-duty vehicle travels, located within the same lane line, and having the same traveling direction. The rear vehicle is a vehicle which is positioned behind the driving road of the heavy operation vehicle, is positioned in the same lane line and has the same driving direction.
Step two: learning driving behavior of human excellent driver
In order to improve the adaptability of the anti-collision early warning strategy, the invention introduces a reverse reinforcement learning algorithm based on the maximum entropy to learn the driving behaviors of the excellent driver collected in the step one under different driving conditions.
In an actual traffic scene, the driving behavior of an excellent driver is not easy to express explicitly, but it is relatively easy to acquire a driving track generated by the excellent driving behavior. Considering that the driving track of the excellent driver has the maximum reward value in all possible tracks, the driving behavior of the excellent driver is represented by the reward function.
First, a reward function for the excellent driver's driving trajectory is established:
in the expression (1), xi i represents the traveling locus of the i-th excellent driver, and xii={(S1,A1),(S2,A2),...,(Sm,Am) M represents the number of driving tracks of excellent drivers collected, rθ(ξi) Feature vector representing the ith excellent driver's driving track, i.e. reward function for this driving track, rθ(Si,Ai) Reward value, S, representing the ith "state-action" in this trackiIndicating the state at time i, AiIndicating the operation at time i.
Considering that an excellent driver often makes driving decisions according to variables such as running speed, yaw angle, distance from a lane line, distance from front and rear obstacles, and the like, the present invention linearly fits a reward value using longitudinal speed, lateral speed, yaw angle, and distance from front and rear obstacles.
rθ(Si,Ai)=rθ(φ1,φ2,φ3,φ4)=θrT·φ (2)
In the formula (2), the characteristic value phi1=vsx cosψs,φ2=vsy sinψs,φ3=dsf-d0,φ4=dsr-d0,vsx,vsyRespectively, the lateral and longitudinal speeds of a heavy commercial vehicle in meters per second, psisIs the yaw angle in degrees, dsf,dsrRespectively represents the relative distance between the heavy operation vehicle and the front vehicle and the rear vehicle, and the unit is meter and thetarTPhi represents the fitted eigenvalue for the coefficient matrix.
The probability of a trajectory having the maximum entropy can be expressed as:
in the formula (3), p (xi)i| θ) represents the probability of the trace having the maximum entropy, Z (θ) is a partition function, and representing a policy nt-1And n represents the number of sampling tracks in the current strategy.
Secondly, establishing a probability model of the driving track of the excellent driver, and solving the driving track with the maximum entropy by using the maximum information entropy principle, wherein the formula is shown as (4):
in the formula (4), the reaction mixture is,representing the collected driving track of the excellent driver.
Converting equation (4) into a product by using a Lagrange multiplier method:
in the formula (5), J (θ) is a loss function.
Considering that the greater the probability of occurrence of the driving trajectory of the excellent driver, the more the reward function expresses the driving behavior of the excellent driver, equation (5) is described as:
minimizing the reward function by utilizing a gradient descent method to obtain a global optimal solution of the reward function:
and finally, optimizing the parameters of the reward function by using a gradient descent algorithm, and further learning the global optimal solution of the reward function. According to the optimized parameter thetarThe current reward function r can be outputθ(Si,Ai) I.e. a function characterizing the excellent driver driving behavior.
Step three: establishing an anti-collision driving decision model
Common anti-collision early warning strategies mainly comprise a method based on a system physical model and a data driving method. The anti-collision early warning strategy based on the system physical model compares an actual value representing collision danger with a set alarm threshold value, and performs collision early warning when the actual value exceeds the threshold value, however, in the vehicle movement process, uncertainty exists in vehicle movement parameters, road conditions and rear traffic states, so that the methods lack accuracy and environmental adaptability. In the data-driven-based method, the deep reinforcement learning method combines the perception capability of deep learning and the decision capability of reinforcement learning, and has the characteristic of adaptability to the uncertainty problem. Therefore, the anti-collision driving decision model of the heavy-duty operation vehicle is established by adopting a deep reinforcement learning algorithm and comprehensively considering the influence of the forward and backward barriers on the vehicle collision.
The decision method based on deep reinforcement learning mainly comprises the following steps: and the decision-making method is based on a value function, strategy search and an Actor-Critic framework. The value-based deep reinforcement learning algorithm cannot handle the problem of continuous output and cannot meet the requirement of continuously outputting a driving strategy in an anti-collision decision. Compared with a method based on strategy search, the decision method based on the Actor-Critic architecture combines value function estimation and strategy search, has a high updating speed, and obtains a good effect in the aspect of outputting a continuous action space by using a Deep Q Network (DQN) experience playback thought as a Deep Deterministic Policy Gradient (DDPG) algorithm. Therefore, the anti-collision driving decision model is established by adopting the DDPG algorithm and based on the driving behaviors of the excellent drivers collected in the step one and the excellent driving strategies obtained in the step two, and anti-collision early warning strategies under different driver operations and driving conditions are researched. The method specifically comprises the following 4 sub-steps:
substep 1: defining basic parameters for an anti-collision driving decision model
Considering that the future motion state of the heavy-duty operating vehicle is influenced by the current motion state and the current action at the same time, the anti-collision driving decision problem is modeled into a Markov decision process, and the basic parameters of the model are defined: state S at time ttState S at time t +1t+1Action A at time ttAnd action AtCorresponding return value Rt(ii) a Specifically, the method comprises the following steps:
(1) defining a state space
The running safety of a heavy-duty vehicle is related to not only the motion state of the vehicle itself but also the relative motion state of the front and rear obstacles. Therefore, using the motion state information obtained in step one, a state space is defined:
St=(vsx,vsy,vsf,vsr,asx,asy,dsf,dsr,ωs,θs,δbr,δthr) (8)
in the formula (8), vsf,vsrRespectively representing the relative speeds of the heavy commercial vehicle, a front vehicle and a rear vehicle, and the unit is meter per second; a issx,asyRespectively representing the transverse acceleration and the longitudinal acceleration of the heavy commercial vehicle, and the unit is meter per second of square; omegasThe yaw velocity of the heavy commercial vehicle is expressed in radian per second; thetasFor the steering wheel angle of heavy commercial vehicles in degrees, deltabr,δthrRespectively represents the opening of a brake pedal and the opening of a throttle valve of the heavy commercial vehicle, and the unit is percentage.
(2) Defining action decisions
In order to establish a more accurate and reliable anti-collision early warning strategy, the invention considers the transverse motion and the longitudinal motion of the vehicle, simultaneously considers that the control quantities of a throttle valve and a brake pedal of the vehicle can not appear simultaneously, takes the steering wheel angle and the accelerating/braking normalization quantity as the control quantities, and defines the early warning strategy output by a decision model, namely an action decision At=[θstr_out,δs_out]。
Wherein A istFor action decision at time t, θstr_outRepresents the normalized steering wheel angle control quantity in the range of [ -1,1],δs_outRepresents an acceleration/braking normalization quantity in the range of-1, 1]. When deltatsuo_When the value is 0, the vehicle is moving at a constant speed, and when the value is deltas_outWhen-1, the vehicle is braked at the maximum deceleration, and when δs_outWhen 1, the vehicle is accelerated at the maximum acceleration.
(3) Defining a reward function
To realize action blockPolicy AtAnd (4) quantitative evaluation of the advantages and the disadvantages, namely materializing and digitizing the evaluation in a mode of establishing a return function. If the action A is executedtAnd then, the running state of the heavy commercial vehicle can be safer, the return value is reward, otherwise, the return value is punishment, and the anti-collision driving decision model can judge the last executed error action to a certain extent.
Different from passenger vehicles, heavy commercial vehicles have the characteristics of higher mass center position, larger load capacity and the like, and are easy to rollover in the processes of emergency braking, steering and lane changing. Therefore, when an anti-collision early warning strategy is established, the occurrence of vehicle collision and rollover needs to be considered at the same time. Defining the reward function as:
Rt=r1+r2+r3 (9)
in the formula (9), RtFor a reward function at time t, r1For a safety distance reward function, r2As a comfort reward function, r3Is a penalty function.
First, in order to prevent a collision of a vehicle, a safe distance reward function r is designed1:
In the formula (10), d0A safe distance threshold.
Secondly, in order to ensure the driving comfort of the vehicle, the excessive impact degree should be avoided as much as possible, and a comfort rewarding function r is designed2=|asy(t+1)-asy(t)|。
Finally, in order to judge the error action of the vehicle, a penalty function r is designed3:
In the formula (11), SpenFor penalty, in the present invention, take Spen-100, when the vehicle crashes or rolls overThe decision model will get a penalty of-100.
Substep 2: network architecture for building anti-collision decision model
And (3) constructing an anti-collision driving decision network by using a strategy-evaluation network framework, wherein the anti-collision driving decision network comprises a strategy network and a value function network. Wherein a policy network is used for the pair state StAnd regressing the features to output a continuous action at(ii) a Value function network for receiving state StAnd action AtTo evaluate the value of the current "state-action". The network architecture is shown in fig. 2, specifically:
(1) designing a policy network
Establishing a strategy network by utilizing a plurality of neural networks with full connection layer structures; the normalized state space StIn turn with the full-link layer F1Full connection layer F2And a full joint layer F3Connected to obtain an output O1Immediate action decision At;
Considering that the dimension of the state space is 12, the number of neurons of the state input layer is set to 12. The activation function of each fully-connected layer is a Linear rectification Unit (ReLU) with the expression F (x) max (0, x), and the fully-connected layer F1、F2、F3The number of neurons in (A) is 20, 20, 10, respectively.
(2) Design value function network
Establishing a value function network by utilizing a plurality of neural networks with full connection layer structures; the normalized state quantity StAnd action AtIn turn with the full-link layer F4Full connection layer F5And a full joint layer F6Connecting to obtain an output O2, namely a Q value; the activation function of each fully-connected layer is ReLU, and the fully-connected layer F4、F5、F6The number of neurons in (A) is 20, 20, 10, respectively.
Substep 3: training strategy network and value function network
The strategy network and the value function network have respective network parameters, and the network parameters of the two parts are updated during training iteration, so that the network converges to obtain a better result. The specific training updating step comprises the following steps:
Substep 3.2: establishing a reward function using equation (2) and initializing a value function network parameter θQPolicy network parameter θμAnd a parameter thetar;
Substep 3.3: taking the formula (9) as an initial strategy optimization target, and performing strategy optimization by using a DDPG algorithm to obtain an initial strategy pi0;
Substep 3.4: performing iterative solution, each iteration comprising substep 3.41 to substep 3.45, in particular:
substep 3.43: optimizing reward function parameters using stochastic gradient descent algorithm minimization equation (7)
Substep 3.44: the optimized reward function rθ(Si,Ai) As an optimization target, the DDPG algorithm is used for strategy optimization, and the value function network parameter theta is updatedQAnd a policy network parameter θμ;
Substep 3.45: and calculating the updating amplitude of the reward function, wherein when the updating amplitude of the reward function is smaller than a given threshold, the reward function at the moment is the optimal reward function.
Substep 3.5: and (4) performing iterative updating according to the method provided by the substep 3.4, so that the strategy network and the value function network are gradually converged. In the training process, if the vehicle collides or turns over, the current round is stopped and a new round is started for training. And when the heavy-duty operation vehicle stably and effectively avoids vehicle collision by using the decision strategy output by the model, the iteration is finished.
Substep 4: outputting an anti-collision early warning strategy by using an anti-collision driving decision model
The information collected by the centimeter-level high-precision differential GPS, the inertial measurement unit, the millimeter wave radar and other sensors is input into the trained anti-collision driving decision network, so that reasonable steering wheel turning angle and throttle opening degree commands can be output in real time, accurate, quantitative and reliable driving suggestions are provided for drivers, and the anti-collision early warning strategy output of the heavy-duty operation vehicle, which is accurate, reliable and self-adaptive to driver operation and driving conditions, is realized.
Claims (1)
1. The utility model provides a heavy type operation vehicle anticollision early warning method of obstacle before the comprehensive consideration, its characterized in that: the method comprises the following steps:
step one, building a driving simulation platform:
the method comprises the steps of constructing a driving simulation platform with obstacles in front and at the back of a heavy operation vehicle in the driving process of the heavy operation vehicle, and collecting driving behaviors of excellent drivers in a real driving environment; the method specifically comprises the following steps:
firstly, a driving simulation platform based on Prescan is built, a town virtual environment model comprising a straight road and a curve is built, and a driver controls a heavy operation vehicle to move through a driving simulator;
secondly, a centimeter-level high-precision differential GPS, an inertia measurement unit and a millimeter wave radar are installed on the heavy operation vehicle to obtain accurate motion state information and relative motion state information of the vehicle, wherein the accurate motion state information and the relative motion state information specifically comprise position, speed, yaw angle, acceleration, relative speed and relative distance; meanwhile, control information of a driver is obtained by utilizing a vehicle body CAN bus, wherein the control information comprises brake pedal pressure, steering wheel turning angle and throttle opening;
finally, 6 driving working conditions of lane changing, lane keeping, vehicle following, constant speed, acceleration and deceleration are designed, 30 excellent drivers with different ages and driving styles are selected to perform data acquisition tests, data acquisition of various typical driving behaviors of the excellent drivers is achieved under a space-time global unified coordinate system, and a driving database of the excellent drivers is constructed;
the definition of the front vehicle refers to a vehicle which is positioned in front of a running road of a heavy operation vehicle, positioned in the same lane line and has the same running direction; the rear vehicle is a vehicle which is positioned behind the driving road of the heavy operation vehicle, is positioned in the same lane line and has the same driving direction;
step two: learning driving behavior of human excellent driver
A reverse reinforcement learning algorithm based on the maximum entropy is introduced, and driving behaviors of the excellent driver collected in the step one under different driving conditions are learned;
representing the driving behavior of a human excellent driver by using a reward function;
first, a reward function for the excellent driver's driving trajectory is established:
in the formula (1), xiiRepresents the travel locus of the ith excellent driver, and xii={(S1,A1),(S2,A2),...,(Sm,Am) M represents the number of driving tracks of excellent drivers collected, rθ(ξi) Feature vector representing the ith excellent driver's driving track, i.e. reward function for this driving track, rθ(Si,Ai) Reward value, S, representing the ith "state-action" in this trackiIndicating the state at time i, AiAn action indicating time i;
linear fitting is carried out on the reward value by utilizing the longitudinal speed, the transverse speed, the yaw angle and the distance between the front obstacle and the rear obstacle;
rθ(Si,Ai)=rθ(φ1,φ2,φ3,φ4)=θrT·φ (2)
in the formula (2), the characteristic value phi1=vsxcosψs,φ2=vsysinψs,φ3=dsf-d0,φ4=dsr-d0,vsx,vsyRespectively, the lateral and longitudinal speeds of a heavy commercial vehicle in meters per second, psisIs the yaw angle in degrees, dsf,dsrRespectively represents the relative distance between the heavy operation vehicle and the front vehicle and the rear vehicle, and the unit is meter and thetarTIs a coefficient matrix, phi represents the characteristic value after fitting;
the probability of a trajectory having the maximum entropy can be expressed as:
in the formula (3), p (xi)i| θ) represents the probability of the trace having the maximum entropy, Z (θ) is a partition function, and representing a policy nt-1The number of the lower tracks is n, and the number of the lower tracks in the current strategy is n;
secondly, establishing a probability model of the driving track of the excellent driver, and solving the driving track with the maximum entropy by using the maximum information entropy principle, wherein the formula is shown as (4):
in the formula (4), the reaction mixture is,representing the collected driving track of the excellent driver;
converting equation (4) into a product by using a Lagrange multiplier method:
in the formula (5), J (theta) is a loss function;
considering that the greater the probability of occurrence of the driving trajectory of the excellent driver, the more the reward function expresses the driving behavior of the excellent driver, equation (5) is described as:
minimizing the reward function by utilizing a gradient descent method to obtain a global optimal solution of the reward function:
finally, parameters of the reward function are optimized by using a gradient descent algorithm, and then the global optimal solution of the reward function is learned; according to the optimized parameter thetarThe current reward function r can be outputθ(Si,Ai) I.e. a function characterizing excellent driver driving behavior;
step three: establishing an anti-collision driving decision model
Establishing an anti-collision driving decision model by adopting a DDPG algorithm based on the driving behavior of the excellent driver collected in the step one and the excellent driving strategy obtained in the step two, and researching anti-collision early warning strategies under different driver operation and driving conditions; the method specifically comprises the following 4 sub-steps:
substep 1: defining basic parameters for an anti-collision driving decision model
Modeling the anti-collision driving decision problem as a Markov decision process, and defining basic parameters of the model: state S at time ttState S at time t +1t+1Action A at time ttAnd action AtCorresponding return value Rt(ii) a Specifically, the method comprises the following steps:
(1) defining a state space
The running safety of the heavy operation vehicle is not only related to the motion state of the vehicle, but also related to the relative motion state of front and rear obstacles; therefore, using the motion state information obtained in step one, a state space is defined:
St=(vsx,vsy,vsf,vsr,asx,asy,dsf,dsr,ωs,θs,δbr,δthr) (8)
in the formula (8), vsf,vsrRespectively representing the relative speeds of the heavy commercial vehicle, a front vehicle and a rear vehicle, and the unit is meter per second; a issx,asyRespectively representing the transverse acceleration and the longitudinal acceleration of the heavy commercial vehicle, and the unit is meter per second of square; omegasThe yaw rate of the vehicle is expressed in radians per second; thetasIs the steering wheel angle of the vehicle, in degrees, deltabr,δthrRespectively representing the opening of a brake pedal and the opening of a throttle valve of the vehicle, and the unit is percentage;
(2) defining action decisions
Considering both the transverse motion and the longitudinal motion of the vehicle and considering that the control quantity of a throttle valve and a brake pedal of the vehicle cannot appear simultaneously, the steering wheel angle and the accelerating/braking normalization quantity are used as the control quantity, and an early warning strategy output by a decision model is defined, namely an action decision At=[θstr_out,δs_out];
Wherein A istFor action decision at time t, θstr_outRepresents the normalized steering wheel angle control quantity in the range of [ -1,1],δs_outRepresents an acceleration/braking normalization quantity in the range of-1, 1](ii) a When deltatsuo_When the value is 0, the vehicle is moving at a constant speed, and when the value is deltas_outWhen-1, the vehicle is braked at the maximum deceleration, and when δs_outWhen the acceleration is 1, the vehicle is accelerated at the maximum acceleration;
(3) defining a reward function
Defining the reward function as:
Rt=r1+r2+r3 (9)
in the formula (9), RtFor a reward function at time t, r1For a safety distance reward function, r2As a comfort reward function, r3Is a penalty function;
first, in order to prevent a collision of a vehicle, a safe distance reward function r is designed1:
In the formula (10), d0A safe distance threshold;
secondly, in order to ensure the driving comfort of the vehicle, the excessive impact degree should be avoided as much as possible, and a comfort rewarding function r is designed2=|asy(t+1)-asy(t)|;
Finally, in order to judge the error action of the vehicle, a penalty function r is designed3:
In the formula (11), SpenIs a penalty item;
substep 2: network architecture for building anti-collision decision model
Constructing an anti-collision driving decision network by utilizing a strategy-evaluation network framework, wherein the anti-collision driving decision network comprises a strategy network and a value function network; wherein a policy network is used for the pair state StAnd regressing the features to output a continuous action at(ii) a Value function networkFor receiving state StAnd action AtTo evaluate the value of the current "state-action"; specifically, the method comprises the following steps:
(1) designing a policy network
Establishing a strategy network by utilizing a plurality of neural networks with full connection layer structures; the normalized state space StIn turn with the full-link layer F1Full connection layer F2And a full joint layer F3Connected to obtain an output O1Immediate action decision At;
Setting the number of neurons of the state input layer to 12 in consideration of the dimension of the state space to 12; the activation function of each fully-connected layer is a Linear rectification Unit (ReLU) with the expression F (x) max (0, x), and the fully-connected layer F1、F2、F3The number of the neurons is 20, 20 and 10;
(2) design value function network
Establishing a value function network by utilizing a plurality of neural networks with full connection layer structures; the normalized state quantity StAnd action AtIn turn with the full-link layer F4Full connection layer F5And a full joint layer F6Connecting to obtain an output O2, namely a Q value; the activation function of each fully-connected layer is ReLU, and the fully-connected layer F4、F5、F6The number of the neurons is 20, 20 and 10;
substep 3: training strategy network and value function network
The strategy network and the value function network have respective network parameters, and the network parameters of the two parts are updated during training iteration, so that the network is converged to obtain a better result; the specific training updating step comprises the following steps:
Substep 3.2: establishing a reward function using equation (2) and initializing a value function network parameter θQPolicy network parameter θμAnd a parameter thetar;
Substep 3.3: taking the formula (9) as an initial strategy optimization target, and performing strategy optimization by using a DDPG algorithm to obtain an initial strategy pi0;
Substep 3.4: performing iterative solution, each iteration comprising substep 3.41 to substep 3.45, in particular:
substep 3.43: optimizing reward function parameters using stochastic gradient descent algorithm minimization equation (7)
Substep 3.44: the optimized reward function rθ(Si,Ai) As an optimization target, the DDPG algorithm is used for strategy optimization, and the value function network parameter theta is updatedQAnd a policy network parameter θμ;
Substep 3.45: calculating the updating amplitude of the reward function, wherein when the updating amplitude of the reward function is smaller than a given threshold value, the reward function at the moment is the optimal reward function;
substep 3.5: performing iterative updating according to the method provided by the substep 3.4 to gradually converge the policy network and the value function network; in the training process, if the vehicle is collided or turned over, the current round is stopped and a new round is started for training; when the heavy-duty operation vehicle stably and effectively avoids vehicle collision by using a decision strategy output by the model, the iteration is completed;
substep 4: outputting an anti-collision early warning strategy by using an anti-collision driving decision model
The information collected by the centimeter-level high-precision differential GPS, the inertial measurement unit, the millimeter wave radar and other sensors is input into the trained anti-collision driving decision network, so that reasonable steering wheel turning angle and throttle opening degree commands can be output in real time, accurate, quantitative and reliable driving suggestions are provided for drivers, and the anti-collision early warning strategy output of the heavy-duty operation vehicle, which is accurate, reliable and self-adaptive to driver operation and driving conditions, is realized.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011512720.8A CN112622886B (en) | 2020-12-20 | 2020-12-20 | Anti-collision early warning method for heavy operation vehicle comprehensively considering front and rear obstacles |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011512720.8A CN112622886B (en) | 2020-12-20 | 2020-12-20 | Anti-collision early warning method for heavy operation vehicle comprehensively considering front and rear obstacles |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112622886A CN112622886A (en) | 2021-04-09 |
CN112622886B true CN112622886B (en) | 2022-02-15 |
Family
ID=75317806
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011512720.8A Active CN112622886B (en) | 2020-12-20 | 2020-12-20 | Anti-collision early warning method for heavy operation vehicle comprehensively considering front and rear obstacles |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112622886B (en) |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113184040B (en) * | 2021-06-03 | 2023-04-07 | 长安大学 | Unmanned vehicle line-controlled steering control method and system based on steering intention of driver |
CN113753034B (en) * | 2021-10-21 | 2022-08-02 | 东南大学 | Large-scale commercial vehicle anti-collision decision method considering road adhesion conditions |
CN113954837B (en) * | 2021-11-06 | 2023-03-14 | 交通运输部公路科学研究所 | Deep learning-based lane change decision-making method for large-scale commercial vehicle |
CN114407925B (en) * | 2022-01-20 | 2024-05-14 | 江苏大学 | Automatic driving track planning system and method based on space-time aerial view and strategy gradient algorithm |
CN114407931B (en) * | 2022-02-21 | 2024-05-03 | 东南大学 | Safe driving decision method for automatic driving operation vehicle of high class person |
CN114379540B (en) * | 2022-02-21 | 2024-04-30 | 东南大学 | Rollover-prevention driving decision method for large-sized operation vehicle considering influence of front obstacle |
CN114863708B (en) * | 2022-05-09 | 2023-04-18 | 东南大学 | Road confluence area roadside real-time accurate induction method for commercial vehicles |
CN116946162B (en) * | 2023-09-19 | 2023-12-15 | 东南大学 | Intelligent network combined commercial vehicle safe driving decision-making method considering road surface attachment condition |
CN116959260B (en) * | 2023-09-20 | 2023-12-05 | 东南大学 | Multi-vehicle driving behavior prediction method based on graph neural network |
CN117348415B (en) * | 2023-11-08 | 2024-06-04 | 重庆邮电大学 | Automatic driving decision method based on finite state machine |
CN117456753B (en) * | 2023-12-26 | 2024-03-08 | 山东高速信息集团有限公司 | Safety early warning method and system between vehicles |
CN117912259B (en) * | 2024-03-19 | 2024-06-21 | 中汽数据有限公司 | Traffic accident reproduction method and device based on automobile electronic data, electronic equipment and storage medium |
CN118153212B (en) * | 2024-05-11 | 2024-07-05 | 长春设备工艺研究所 | Digital heterogeneous model generation system and method based on multi-scale fusion |
CN118312750A (en) * | 2024-06-13 | 2024-07-09 | 鹰驾科技(深圳)有限公司 | Vehicle-mounted chip-based driving auxiliary decision-making method and system |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110027553A (en) * | 2019-04-10 | 2019-07-19 | 湖南大学 | A kind of anti-collision control method based on deeply study |
WO2020109547A1 (en) * | 2018-11-29 | 2020-06-04 | Valeo Schalter Und Sensoren Gmbh | Advanced highway assist scenario |
CN111696387A (en) * | 2020-05-21 | 2020-09-22 | 东南大学 | Self-adaptive anti-collision grading early warning method based on forward obstacle identification |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1717778B1 (en) * | 2005-04-29 | 2007-07-11 | Ford Global Technologies, LLC | Method and system for forward collision avoidance in an automotive vehicle |
-
2020
- 2020-12-20 CN CN202011512720.8A patent/CN112622886B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020109547A1 (en) * | 2018-11-29 | 2020-06-04 | Valeo Schalter Und Sensoren Gmbh | Advanced highway assist scenario |
CN110027553A (en) * | 2019-04-10 | 2019-07-19 | 湖南大学 | A kind of anti-collision control method based on deeply study |
CN111696387A (en) * | 2020-05-21 | 2020-09-22 | 东南大学 | Self-adaptive anti-collision grading early warning method based on forward obstacle identification |
Non-Patent Citations (2)
Title |
---|
UAV environmental perception and autonomous obstacle avoidance:A deep learning and depth camera combined solution;Wei Li等;《Computers and Electronics in Agriculture》;20200831;全文 * |
基于两级滤波的车辆相对加速度估计;宋翔等;《东南大学学报》;20150131;全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN112622886A (en) | 2021-04-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112622886B (en) | Anti-collision early warning method for heavy operation vehicle comprehensively considering front and rear obstacles | |
CN112633474B (en) | Backward collision avoidance driving decision method for heavy commercial vehicle | |
CN114407931B (en) | Safe driving decision method for automatic driving operation vehicle of high class person | |
CN112580148B (en) | Heavy-duty operation vehicle rollover prevention driving decision method based on deep reinforcement learning | |
CN114312830B (en) | Intelligent vehicle coupling decision model and method considering dangerous driving conditions | |
CN113954837B (en) | Deep learning-based lane change decision-making method for large-scale commercial vehicle | |
CN114564016A (en) | Navigation obstacle avoidance control method, system and model combining path planning and reinforcement learning | |
CN113753034B (en) | Large-scale commercial vehicle anti-collision decision method considering road adhesion conditions | |
CN114379540B (en) | Rollover-prevention driving decision method for large-sized operation vehicle considering influence of front obstacle | |
CN109283843A (en) | A kind of lane-change method for planning track merged based on multinomial with particle swarm algorithm | |
CN114580302A (en) | Decision planning method for automatic driving automobile based on maximum entropy reinforcement learning | |
CN114987461A (en) | Intelligent passenger car dynamic lane change trajectory planning method under multi-car complex traffic environment | |
CN115257789A (en) | Decision-making method for side anti-collision driving of commercial vehicle in urban low-speed environment | |
CN115593433A (en) | Remote take-over method for automatic driving vehicle | |
CN114889589A (en) | Intelligent automobile steering and braking cooperative collision avoidance control system and method | |
Zhao et al. | Adaptive drift control of autonomous electric vehicles after brake system failures | |
Zhan et al. | Risk-aware lane-change trajectory planning with rollover prevention for autonomous light trucks on curved roads | |
CN115454065B (en) | Transverse and longitudinal coupling vehicle formation prediction control method based on reinforcement learning algorithm | |
Zhang et al. | Minimum time lane changing problem of vehicle handling inverse dynamics considering the driver’s intention | |
Shadrin | Affordable and efficient autonomous driving in all weather conditions | |
CN115257820A (en) | Open interference scene-oriented forward collision avoidance driving decision method for commercial vehicle | |
Wang et al. | A double-layered nonlinear model predictive control based control algorithm for local trajectory planning for automated trucks under uncertain road adhesion coefficient conditions | |
CN115292671A (en) | Driver horizontal-vertical coupling behavior model | |
Narendran et al. | Autonomous lateral control of vehicles in an automated highway system | |
CN114925461A (en) | Emergency steering control strategy network model, training method, modeling method and simulation method for automatic driving commercial vehicle |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |