CN114035575B - Unmanned vehicle motion planning method and system based on semantic segmentation - Google Patents
Unmanned vehicle motion planning method and system based on semantic segmentation Download PDFInfo
- Publication number
- CN114035575B CN114035575B CN202111301912.9A CN202111301912A CN114035575B CN 114035575 B CN114035575 B CN 114035575B CN 202111301912 A CN202111301912 A CN 202111301912A CN 114035575 B CN114035575 B CN 114035575B
- Authority
- CN
- China
- Prior art keywords
- network
- semantic segmentation
- image
- images
- speed
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000011218 segmentation Effects 0.000 title claims abstract description 100
- 238000000034 method Methods 0.000 title claims abstract description 34
- 239000013598 vector Substances 0.000 claims abstract description 58
- 230000004438 eyesight Effects 0.000 claims abstract description 30
- 238000004088 simulation Methods 0.000 claims abstract description 27
- 238000012549 training Methods 0.000 claims abstract description 9
- 230000006870 function Effects 0.000 claims description 21
- 230000007613 environmental effect Effects 0.000 claims description 20
- 230000004927 fusion Effects 0.000 claims description 15
- 238000013528 artificial neural network Methods 0.000 claims description 12
- 239000010410 layer Substances 0.000 claims description 12
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 8
- 238000011176 pooling Methods 0.000 claims description 8
- 230000008569 process Effects 0.000 claims description 8
- 230000000007 visual effect Effects 0.000 claims description 8
- 239000002356 single layer Substances 0.000 claims description 6
- 238000000605 extraction Methods 0.000 claims description 5
- 238000004364 calculation method Methods 0.000 claims description 4
- 238000004422 calculation algorithm Methods 0.000 claims description 3
- 230000015654 memory Effects 0.000 claims description 3
- 239000011800 void material Substances 0.000 claims description 3
- 230000004913 activation Effects 0.000 claims description 2
- 239000011796 hollow space material Substances 0.000 claims description 2
- 239000011159 matrix material Substances 0.000 claims description 2
- 230000006399 behavior Effects 0.000 abstract description 3
- 239000002352 surface water Substances 0.000 abstract 1
- 238000010586 diagram Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 230000008447 perception Effects 0.000 description 4
- 101100295091 Arabidopsis thaliana NUDT14 gene Proteins 0.000 description 3
- 230000004888 barrier function Effects 0.000 description 3
- 238000013507 mapping Methods 0.000 description 3
- 230000016776 visual perception Effects 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 2
- 210000004027 cell Anatomy 0.000 description 2
- 238000005094 computer simulation Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000007787 long-term memory Effects 0.000 description 2
- 238000003062 neural network model Methods 0.000 description 2
- 230000006403 short-term memory Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000013210 evaluation model Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 230000003278 mimic effect Effects 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0246—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using a video camera in combination with image processing means
- G05D1/0253—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using a video camera in combination with image processing means extracting relative motion information from a plurality of images taken successively, e.g. visual odometry, optical flow
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0212—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
- G05D1/0221—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory involving a learning process
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0212—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
- G05D1/0223—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory involving speed control of the vehicle
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0276—Control of position or course in two dimensions specially adapted to land vehicles using signals provided by a source external to the vehicle
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Aviation & Aerospace Engineering (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- General Physics & Mathematics (AREA)
- Automation & Control Theory (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Electromagnetism (AREA)
- Image Analysis (AREA)
- Traffic Control Systems (AREA)
Abstract
The invention discloses a semantic segmentation-based unmanned vehicle motion planning method and system, wherein a view is obtained from a vehicle vision sensor, the semantically segmented view and a corresponding weight are obtained through a semantic segmentation network and a category weight network, an expert driving experience base training simulation learning module bound with the corresponding view is used for fusing the semantically segmented view and the corresponding weight on a pixel level, a feature vector is extracted from a fused image, after training is completed, speed and steering are obtained end to end by directly inputting a current view, passing through the semantic segmentation module and the simulation learning module, and motion planning is carried out based on the speed and the steering. The invention can well carry out unmanned motion planning on the vehicle, and the behavior planning accuracy under the simulation environment Carla is improved by 27.8 percent compared with the method which is not suitable for semantic segmentation under some specific scenes, such as the reflection obstacle of the surface water in rainy days.
Description
Technical Field
The invention relates to the field of unmanned driving, in particular to a semantic segmentation-based unmanned vehicle motion planning method and system.
Background
In recent years, a deep neural network model based on end-to-end learning is an emerging research hotspot in the field of automatic driving. Different from the traditional method, the method can overcome the problem that the traditional method relies on prior environment modeling, and can directly emit control quantity from the perception mapping, so that the system is simplified. In 2016, NVIDIA demonstrated the feasibility of the end-to-end control approach by training the neural network by collecting real vehicle driving data. An end-to-end automatic driving model CIL is proposed by Dosovitskiy and the like in 2017, and experiments show that the CIL can effectively utilize navigation information to conduct navigation. In 2019, codevilla et al propose that the CILRS model uses ResNet34 to realize environment perception, so that a good effect is achieved, and the problem of abnormal parking of the CIL model is eliminated to a certain extent.
The traditional end-to-end model only takes the current environment information as network input, only depends on the current image to obtain the road track and the obstacle information at the moment, and a human driver judges the movement trend and the movement speed of the dynamic obstacle based on the visual information of a period of time in the past so as to determine the driving strategy. Compared to a human driver, the conventional model will certainly lose the motion information of the dynamic obstacle.
In order to acquire dynamic information of the surrounding environment, environmental information at successive times must be added as input of the neural network, which undoubtedly increases the data processing amount of the network, thereby delaying the network from feeding back the vehicle driving information.
Disclosure of Invention
The invention aims to provide a semantic segmentation-based unmanned vehicle motion planning method and system, which can reduce the data processing burden of a learning network while increasing the data input amount, and more accurately control the speed and direction of an unmanned vehicle. Thereby effectively determining the driving strategy.
The technical solution for realizing the purpose of the invention is as follows:
a semantic segmentation based unmanned vehicle motion planning method comprises the following steps:
periodically collecting continuous frame surrounding views of the left direction, the front direction and the right direction of the vehicle;
based on the environmental information classification, respectively carrying out semantic segmentation on the left and right direction current view data and the continuous n frames of forward view data through a semantic segmentation network to obtain a semantic segmentation image;
based on the environmental information classification, acquiring category weight information corresponding to the continuous n frames of forward images through a category weight network;
acquiring the speed and direction corresponding to each frame of view based on an expert database;
taking the segmented images, the category weight information and the speed and direction corresponding to the view as training data, training and learning the dynamic surround vision imitating learning network based on the neural network, and repeating the steps until a convergent dynamic surround vision imitating learning model is obtained;
collecting surrounding views in the left direction, the front direction and the right direction of a vehicle, acquiring semantic segmentation images of the surrounding views in the left direction, the front direction and the right direction and category weight information corresponding to continuous n frames of forward images, and inputting the semantic segmentation images and the category weight information into a dynamic surrounding vision simulation learning model to acquire a planning speed and a planning direction;
and carrying out motion planning on the unmanned vehicle based on the planned speed and the planned direction.
Further, the semantic segmentation network comprises a residual error network ResNet, a full convolution network FCN, a cavity space convolution pooling pyramid and a decoder; the method for obtaining the semantic segmentation image by respectively performing semantic segmentation on the left-direction current view data, the right-direction current view data and the continuous n frames of forward view data through the semantic segmentation network specifically comprises the following steps of:
performing semantic segmentation on the current left and right view data through a residual error network ResNet and a full convolution network FCN, and performing semantic segmentation on the current and previous n-1 frame forward view data;
based on the environmental information classification, extracting and classifying semantic segmentation image characteristic information by adopting a cavity space convolution pooling pyramid;
and recovering the details and the spatial dimension of the image features through a decoder based on the classified image feature information to obtain a final semantic segmentation image.
Further, the category weight network specifically includes: and forming a residual unit by four residual blocks, copying the fourth residual block for three times, connecting the copied fourth residual block to the residual unit in a cascading manner, and simultaneously performing parallel hollow space convolution pooling pyramid structure.
Further, the dynamic surround vision simulation learning network comprises a feature extraction fusion network and a branch decision network, and an L1 loss function is adopted; wherein:
the feature extraction fusion network fuses category weight information and semantic segmentation images of forward images, n 512-dimensional feature vectors are obtained from the fused images through a ResNet34 residual error network, the n 512-dimensional feature vectors obtain a 512-dimensional feature vector through a single-layer LSTM network, 64-dimensional feature vectors are obtained from the semantic segmentation images in the left direction and the semantic segmentation images in the right direction through the residual error network respectively, and one 512-dimensional feature vector and two 64-dimensional feature vectors are spliced to obtain a 640-dimensional combined feature vector;
the branch decision network activates the corresponding branch network according to navigation information, and the navigation information c comprises four states of road following, road walking, left turning and right turning;
the L1 loss function is:
wherein,vehicle speed v at time t +1 t+1 Is predicted value of (4), (v) is greater than or equal to>Is the predicted value of the square disk angle at time t, a t Is the steering wheel angle at time t, v t+1 Is the vehicle speed at time t + 1.
Further, the number of hidden nodes of the LSTM network is 128, and the updating process is as follows:
external state S based on last moment t-1 And input x of the current time t Input gate i is calculated t And an output gate o t Forget door t And candidate statesThe calculation formula is as follows:
i t =σ(M i x t +N i S t-1 +b i )
o t =σ(M o x t +N o S t-1 +b o )
l t =σ(M l x t +N l S t-1 +b l )
wherein, tanh is an activation function, M i And N i Is a parameter matrix, with a value interval of (0, 1), a being a logistic function, a value range of (0,1);
passing forgetting door t And an input gate i t To update the memory cell I t The update formula is:
passing the internal state to the external state S t External state S t Comprises the following steps:
S t =o t tanh(I t )
further, the 640-dimensional joint feature vector is:
wherein e is t 512-dimensional feature vectors and two 64-dimensional feature vectors are spliced to obtain a 640-dimensional joint feature vector,for corresponding 512-dimensional feature vectors for n successive frames of forward images>640-dimensional associated feature vectors corresponding to right and left images at time t, respectively>Semantic segmentation images R corresponding to the forward image, the right image and the left image at the time t respectively 34 Is a 34 layer residual network function, R 18 Is a layer 18 residual network function and L is an LSTM network function.
Further, the dynamic surround vision simulation learning model is as follows:
(v t+1 ,a t )=F(e t ,c)
(v t+1 ,a t )=A c (e t )
wherein v is t+1 Is the vehicle speed at time t +1, a t T is the steering wheel angle, c is the navigation command, and F is the vehicle speed v t+1 And steering wheel angle a t And e t And navigation instructions c.
Further, the collection period is 0.3s, n =4, and the environmental information classification includes: empty, buildings, fences, pedestrians, pillars, lane dividing lines, roads, sidewalks, vegetation, vehicles, walls, signal lamps and other unidentified types.
Further, a PID controller is adopted to carry out motion planning on the unmanned vehicle, and the PID controller adopts a PI control algorithm, and specifically comprises the following steps:
wherein:vehicle speed v at time t +1 t+1 V is predicted value of t Is the vehicle speed at time t, V t Throttle size at time t, b t Braking at time t, m t Is a longitudinal control quantity.
An unmanned vehicle motion planning system based on the unmanned vehicle motion planning method comprises a visual sensor, a category weight module, a semantic segmentation module, a simulation learning module, a speed control module and a steering control module, wherein:
the vision sensor is used for periodically acquiring surrounding views of the left direction, the middle direction and the right direction of the vehicle;
the category weight module is used for acquiring category weight information corresponding to the forward image through a category weight network;
the semantic segmentation module is used for performing semantic segmentation on left-direction current view data and right-direction current view data and continuous n frames of forward view data respectively by a semantic segmentation network to obtain a semantic segmentation image;
the simulation learning module carries out network learning based on neural network learning to obtain a converged dynamic surrounding vision simulation learning model; the simulation learning module comprises an image fusion module and a branch decision module, wherein the image fusion module is used for fusing category weight information with semantic segmentation images of each frame of forward image, the fused image is subjected to ResNet34 residual error network to obtain n 512-dimensional feature vectors, the n 512-dimensional feature vectors are subjected to single-layer LSTM network to obtain a 512-dimensional feature vector, the semantic segmentation images in the left direction and the semantic segmentation images in the right direction are subjected to residual error network to respectively obtain 64-dimensional feature vectors, and one 512-dimensional feature vector and two 64-dimensional feature vectors are spliced to obtain a 640-dimensional combined feature vector; and the branch decision module activates a corresponding branch network according to the navigation information, and the navigation information c comprises four states of road following, line, left turn and right turn.
The speed control module controls the speed of the unmanned vehicle by simulating the planned speed output by the learning module;
the steering control module controls steering of the unmanned vehicle by mimicking the planned direction output by the learning module.
Compared with the prior art, the method has the beneficial effects that: the invention determines the motion track and the speed of a dynamic barrier through continuous multi-frame visual information, reduces the processing amount of input data by a neural network through semantic segmentation, inputs the multi-frame visual information into the semantic segmentation network for processing so as to simplify invalid information of the environment, and the processed image can completely retain effective environmental information of roads, barriers, pedestrians and the like and is more close to the environmental information sensed by a human driver in the driving process; according to the invention, the multi-frame semantic segmentation image is input into the end-to-end neural network model, so that vehicle control information is obtained, unmanned driving behavior planning can be well performed on the vehicle, and the accuracy of the speed and direction of the unmanned vehicle is improved.
Drawings
Fig. 1 is a model architecture diagram of the unmanned vehicle motion planning system of the present invention.
Fig. 2 is a schematic view of an operating environment based on carra unmanned vehicle simulation platform version 8.04 of the present invention.
FIG. 3 is a diagram illustrating semantic segmentation.
Fig. 4 is a diagram of an LSTM network architecture.
Fig. 5 is a diagram of a dynamic surround vision mock learning network architecture.
Fig. 6 is a schematic diagram of proportional-integral control.
FIG. 7 is a graph of throttle/brake and control quantities for a vehicle.
FIG. 8 is a schematic diagram of the convolution of holes at different expansion ratios.
Detailed Description
In order to enable environmental information acquired by an unmanned vehicle to be closer to environmental information perceived by a human driver in the driving process, and reduce the data processing burden of a neural network while increasing the data input quantity, the invention provides a dynamic simulation learning algorithm based on the combination of surround vision and semantic segmentation, wherein multiframe visual information is firstly input into a trained semantic segmentation network for processing so as to simplify invalid information of the environment, and the processed image can completely retain effective environmental information of roads, obstacles, pedestrians and the like; and inputting the processed multi-frame semantic segmentation image into an end-to-end learning network model so as to obtain vehicle control information. The method is the same as the logic of the perception environment of a human driver, the motion trail and the speed of the dynamic barrier are determined through historical visual information, the processing amount of input data of a neural network is reduced through semantic segmentation, and therefore the driving strategy can be effectively determined.
The model idea provided by the method is that visual perception, a category weight network, a semantic segmentation network, image fusion, a Dynamic Surround vision simulation Learning network and speed and steering control of an unmanned vehicle are adopted, wherein the Dynamic Surround vision simulation Learning network is based on a neural network and is named as DSCIL (Dynamic Surround-view simulation Learning). Firstly, a left, middle and right surrounding view is obtained through visual perception installed on a vehicle, front images of 4 continuous frames are input into a category weight network, and all 13 semantic segmentation category pixel-level weights are output for a current frame; then inputting the continuous 4 frames of forward images and the images of the current frames at the left and right sides into a semantic segmentation network to obtain corresponding semantic images; and then inputting the obtained semantic segmentation image and the class weight into a DSCIL model, multiplying the obtained semantic segmentation image and the class weight on a pixel level on the basis of the class to obtain the importance representation of each class on the spatial information after semantic segmentation, outputting the expected vehicle speed, and finally controlling the speed and the steering according to the error between the expected vehicle speed and the current vehicle speed.
The invention provides a semantic segmentation-based unmanned vehicle motion planning method, which comprises the following steps of:
periodically acquiring continuous frame surrounding views in the left direction, the front direction and the right direction of a vehicle;
based on the environmental information classification, performing semantic segmentation on each frame of view data in the left direction, the front direction and the right direction through a semantic segmentation network to obtain a semantic segmentation image, wherein the segmentation image is as shown in FIG. 3;
based on the environmental information classification, obtaining the category weight information corresponding to each frame of forward image through a category weight network;
based on the environmental information classification, acquiring category weight information corresponding to the continuous n frames of forward images through a category weight network;
acquiring the speed and direction corresponding to each frame of view based on an expert database;
taking the segmented images, the category weight information and the speed and the direction corresponding to the view as training data, training and learning the dynamic surround vision imitation learning network based on the neural network, and repeating the steps until a convergent dynamic surround vision imitation learning model is obtained;
collecting surrounding views in the left direction, the front direction and the right direction of a vehicle, acquiring semantic segmentation images of the surrounding views in the left direction, the front direction and the right direction and category weight information corresponding to continuous n frames of forward images, and inputting the semantic segmentation images and the category weight information into a dynamic surrounding vision simulation learning model to acquire a planning speed and a planning direction;
and based on the planning speed and the planning direction, a PID controller is adopted to plan the motion of the unmanned vehicle.
Direct semantic segmentation usually loses information on partial space and is not beneficial to end-to-end behavior planning of unmanned driving, so that a category weight network is designed, resNet is used as a backbone to construct, and finally, a space importance weight corresponding to each category is output, namely, the output of 13 × 200 × 88 is output;
ResNet has four residual blocks in total, and the fourth residual block is copied three times and then connected behind the residual units in a cascade mode. Meanwhile, a parallel ASPP structure is adopted to process the feature graph output in the front, 4 hole convolutions with different expansion rates (6, 12, 18 and 24) are adopted at the top of the feature graph, context information is captured through a global average pooling layer, and then 1 x 1 convolution is used for fusing the features processed by the ASPP structure branches.
The semantic segmentation network is a semantic segmentation network which is usually constructed by taking ResNet as a backbone. After the FCN is used for realizing image segmentation, hole convolution is introduced to obtain multi-scale feature information, and more semantic feature information is mined through a pyramid structure to improve the classification effect.
The hole convolution is to sample the original image by setting a rate (rate). As shown in fig. 8, when rate =1, it is not different from the standard convolution; when rate is greater than 1, the convolution kernel is expanded through the expansion rate, sampling is carried out on the original image every (rate-1) pixels, the scope of the receptive field is expanded, and the semantic segmentation features with a larger scope are extracted under the condition that the number of parameters and the calculation amount are not increased.
The Spatial convolution Pooling Pyramid (ASPP) of the void space means that the void convolution with different expansion rates is adopted to learn multi-scale features, and the region with any scale can be classified more accurately.
The semantic segmentation network employs an encoder-decoder module that recovers the details and spatial dimensions of image features using 16 times bilinear upsampling to obtain segmentation results. The invention aims at classifying the environment information, replaces a decoder module for segmentation into a full connection layer result, takes softmax as an output layer, connects each neuron with output and realizes the classification of the input image. The classification classes include 0 for empty (None), 1 for Buildings (Buildings), 2 for Fences (Fences), 3 for Other unspecified (Other), 4 for Pedestrians (Pedestrians), 5 for pillars (Poles), 6 for lane lines (Roadlines), 7 for Roads (Roads), 8 for Sidewalks (sideroads), 9 for vegetation (vegentations), 10 for Vehicles (Vehicles), 11 for Walls (Walls), and 12 for traffic lights (traffisings).
In order to consider the influence of dynamic obstacles, the perception mode of a human driver is simulated, and more comprehensive environmental information is obtained by increasing the visual information of historical multiframes. The DSCIL model adopts 4 frames of forward images, and the interval between each frame is 0.3s, namely the forward image at the current moment and the forward images at the first 0.3s, the first 0.6s and the first 0.9s respectively. Compared with the single-frame image, the driving speed and the driving track of the automobile can be judged more effectively.
With reference to fig. 5, the dynamic surround vision simulation learning network includes a feature extraction fusion network and a branch decision network, and adopts an L1 loss function; specifically, the method comprises the following steps:
referring to FIG. 4, the number of hidden nodes of the LSTM network is 128, the structure of the LSTM network is shown in FIG. 5, wherein S t ,I t ,x t Respectively a state at the time t, an internal state, a candidate state and a network input; and i t ,o t ,l t The input gate, the output gate and the forgetting gate at the moment t respectively have the value range of (0, 1), and allow the information to pass through in a certain proportion. Where σ represents the logistic function and the range is (0, 1). The state S is updated at every moment, the state S is short-term memory, the network parameters can be regarded as long-term memory, and the updating period is far slower than that of the state S. The internal state I may hold critical information at a certain moment and for a period of time, which is longer than short-term memory and shorter than long-term memory.
The updating process of the LSTM network is as follows: (1) First of all, the external state S of the last moment is used t-1 And input x of the current time t And calculating an input gate, an output gate, a forgetting gate and a candidate state, wherein a calculation formula is shown as formula 1.1-1.4, wherein M and N are parameter matrixes, and b is a bias parameter: (2) Use forgetting door t And an input gate i t To update the memory cell I t The update formula is shown in formula 1.5: (3) Using output gates o t Passing the internal state to an external state S t The formula is shown as formula 1.6.
i t =σ(M i x t +N i S t-1 +b i ) (1.1)
o t =σ(M o x t +N o S t-1 +b o ) (1.2)
l t =σ(M l x t +N l S t-1 +b l ) (1.3)
S t =o t tanh(I t ) (1.6)
512-dimensional fusion characteristics obtained by four frames of forward images are spliced with two 64-dimensional image characteristics obtained by the left and right images, so that the description of a semantic segmentation network and a DSCIL model on a dynamic environment is obtained.
The dynamic surround vision simulation learning network is mainly trained by fusing images and corresponding expert libraries, so that end-to-end unmanned vehicle speed and steering can be obtained by surrounding views in the current state, a speed and steering control module is input to make vehicle driving decisions, and the dynamic surround vision simulation learning network performs speed and direction planning based on a branch decision network.
The vehicle can face four conditions in the driving process, namely the condition that the vehicle follows a road and turns left, runs straight or turns right when meeting an intersection. For different navigation instructions, the DSCIL model uses different branch decision network correspondences, i.e. each navigation instruction only activates its corresponding branch network.
The branch network is selectively activated according to the information c, and different navigation information activates the corresponding branch network to predict the speed and the steering wheel angle of the vehicle, similar to a single-pole multi-throw switch. The navigation information c comprises four states of road following, straight going, left turning and right turning. The speed is also used as an input to predict the accelerator and the brake without a CILRS model (dynamic surround vision imitation learning network) in the network, so that the mapping from network learning to the ground speed and low acceleration with logic errors in data is avoided. This mapping may result in the vehicle possibly stopping when the target is not reached and the road ahead is accessible.
The four branch networks have the same structure, the input of the network is 640-dimensional environment characteristic vectors fused by the visual perception network, and the 2-dimensional vectors are finally output through the full-connection network with 256 nodes, wherein the two layers of nodes are respectively the predicted vehicle speed and the angle of the steering wheel. The vehicle speed is a real number larger than 0, and the value range of the angle is (-1, 1). The last layer of fully connected network uses dropout with probability of 0.5 to avoid the over-fitting phenomenon.
Suppose thatRespectively expressed as a forward image, a right image and a left image obtained by semantic segmentation at the time t, and the fusion characteristic of the forward image at the time can be obtained by the formula 1.7, wherein R 34 Is a function for a 34 layer residual network, L is a function for an LSTM network,
feature vectors of right and left images at time tAnd &>This can be obtained from equations 1.8 and 1.9, respectively, where R18 is the corresponding function of the 18-layer residual network. Dynamic environmental characteristics e at time t t Is when the forward image fusion feature>Right image feature->And the left image feature>As shown in formula 1.10.
As shown in equation 1.11, assume the vehicle speed v of the network at time t +1 t+1 And steering wheel angle a at time t t Is a dynamic environmental feature e t And a function F of the navigation instruction c, and because different navigation instructions correspond to different decision branch networks, the decision branch network corresponding to the navigation instruction c is marked as A c Then, formula 1.12 can be obtained. By combining equations 1.10,1.11 and 1.12, the final mathematical description of equation 1.13, i.e., the DSCIL network (dynamic simulation learning model), can be obtained.
(v t+1 ,a t )=F(e t ,c) (1.11)
(v t+1 ,a t )=A c (e t ) (1.12)
Codevilla et al have experimentally found that L1 loss is more suitable for autonomous driving missions than L2 loss, so the DSCIL model uses the L1 loss, and the loss function is shown in equation 1.14. WhereinIs the square angle of the disk at time t t Is predicted value of (4), (v) is greater than or equal to>Vehicle speed v at time t + 1 t+1 The predicted value of (2). So the first term of the loss function is the disc angle error term and the second term is the vehicle speed error term, both error terms are equal in weight and are 0.5.
The branch decision network only solves the problem of lateral control of the vehicle, and for the problem of longitudinal control of the vehicle, the branch decision network uses PID (proportional-integral-derivative) control to solve the problem according to the current vehicle speed and the next expected vehicle speed.
To solve the longitudinal control problem of the vehicle, proportional-integral (PI) control is introduced to control the vehicle speed. The principle diagram of proportional-integral control is shown in FIG. 6, where a t Is the set value of the system at time t, b t Is the output value of the system at time t, deviation c t Is the difference between the given value and the output value, as shown in equation 1.15. As shown in formula 1.16, m t Is the controlled quantity of the controlled object, is determined by the deviation c t Proportional and integral terms of, wherein k s Is the proportionality coefficient, k i Is an integral coefficient. The proportional term is adjusted based on the deviation, the proportional term increases when the deviation increases, the proportional term decreases when the deviation decreases, and the non-backlash control cannot be realized only by the proportional term. The integral term can effectively eliminate the static error by integrating the system deviation.
c t =a t -b t (1.15)
Empirically, we set the proportional coefficient of the control speed to 0.25 and the integral coefficient to 0.2, and thus we can obtain the control quantity m as shown in the formula 1.17 t Wherein v is t Is the vehicle speed at the time point of t,is the expected value of the speed at time t. Vehicle longitudinal control quantity m at time t t And t moment accelerator V t When m is as shown in formula 1.18 t When the throttle is less than or equal to 0, the throttle size is 0: when m is t When the throttle is larger than 0, the throttle size is equal to m t 1.3 times and a minimum of 0.75. Vehicle longitudinal control quantity m at time t t And brake b at time t t When m is as shown in formula 1.19 t When the brake force is larger than 0, the brake force is 0; when m is t When the brake force is less than or equal to 0, the brake size is equal to m t Minus 0.35 times and 1.
FIG. 7 shows the vehicle longitudinal control quantity m at time t t Throttle and V t And a brake b t In the relationship of (1), the abscissa axis of the coordinate axis represents the control quantity m t The ordinate of the coordinate axis is accelerator/brake, the solid line in the figure represents the variation relation of the accelerator size with the control quantity, and the dotted line represents the variation relation of the brake size with the control quantity. Longitudinal control quantity m t When the timing is positive, the brake size is 0, the accelerator size is within a certain range and m t Is a direct proportional relationship and is limited by a threshold of 0.75 after exceeding the range: when the longitudinal control quantity is negative, the throttle size is 0, the brake size is within a certain range and m t Is in direct proportion, and is limited by a threshold of 1.0 after exceeding the range.
A system based on unmanned vehicle motion planning, includes a vision sensor, a category weight module, a semantic segmentation module, a mimic learning module, a speed control module, and a steering control module, wherein:
the vision sensor is used for periodically acquiring surrounding views of the left direction, the middle direction and the right direction of the vehicle;
the category weight module is used for acquiring category weight information corresponding to the forward image through a category weight network;
the semantic segmentation module is used for performing semantic segmentation on left-direction current view data and right-direction current view data and continuous n frames of forward view data respectively by a semantic segmentation network to obtain a semantic segmentation image;
the simulation learning module carries out network learning based on neural network learning to obtain a converged dynamic surrounding vision simulation learning model; the simulation learning module comprises an image fusion module and a branch decision module, wherein the image fusion module is used for fusing category weight information with semantic segmentation images of each frame of forward images, acquiring n 512-dimensional feature vectors from the fused images through a ResNet34 residual error network, acquiring a 512-dimensional feature vector from the n 512-dimensional feature vectors through a single-layer LSTM network, respectively acquiring 64-dimensional feature vectors from the semantic segmentation images in the left direction and the semantic segmentation images in the right direction through the residual error network, and splicing the 512-dimensional feature vectors with the two 64-dimensional feature vectors to acquire a 640-dimensional combined feature vector; and the branch decision module activates a corresponding branch network according to the navigation information, and the navigation information c comprises four states of road following, line, left turn and right turn.
The speed control module controls the speed of the unmanned vehicle by simulating the planned speed output by the learning module;
the steering control module controls steering of the unmanned vehicle by mimicking the planned direction output by the learning module.
Example 1
With reference to fig. 1, in the unmanned vehicle motion planning system provided in this embodiment, a current left, middle, and right surrounding view is obtained through a vision sensor of the unmanned vehicle, where 4 frames of forward images are input to a category weighting module to obtain spatial category weighting information 13 × 200 × 88 corresponding to a current frame; meanwhile, the 4 frames of forward images and the left and right current frames of images are input into a semantic segmentation module to obtain corresponding 200 × 88 semantic images. 4 frames of forward images after the category weight information and the corresponding semantic segmentation images are combined and processed are subjected to ResNet34 network to obtain 4 512-dimensional feature vectors, 4 feature vectors are subjected to single-layer LSTM network to obtain 512-dimensional feature vectors, the processed left and right side current frame images are respectively subjected to 64-dimensional feature vectors through the same ResNet18 network, and 3 feature vectors are spliced to obtain 640-dimensional combined feature vectors. The joint feature vector predicts the vehicle speed and the steering wheel angle through a 3-layer fully-connected imitation learning module.
The driving evaluation models of the CARLA automatic driving platforms, namely CARLAbenchmark and NoCrash benchmark, are tested and compared with other five classical end-to-end driving models.
With reference to fig. 2, carra benchmark includes 4 different driving conditions, straight, one Turn, navigation, nav. Particularly in a dynamic environment, the test result is obviously superior to other models, which fully illustrates the advantages of the models in the dynamic environment. When the test is carried out under the New Weather & town environment, the task completion rate of the model in the Straight driving condition is 100%, the task completion rate in the other three driving conditions is 98%, and the task completion rate is greatly improved compared with the task completion rate of the other five classical models, for example, compared with the CAL model, the task completion rate of the method is respectively improved by 6%,26%,30% and 34% in the four driving conditions.
The NoCrash bench is more strict than the caralabelchmark in task requirement, and if a stronger collision occurs in the driving process, the task failure is judged, and the model is tested and compared with other five classical end-to-end models. In a New Weather environment, for tasks under three different traffic conditions, namely Empty, regular and Dense, the task completion rate of a model is 100 +/-0%, 85 +/-3%, 36 +/-2%, the CAL model completion rate is 85 +/-2%, 68 +/-5%, 33 +/-2%, and the CILRS model task completion rate is 96 +/-1%, 77 +/-1% and 47 +/-1%. Under a New Town environment, the task completion rate of the model is 67 +/-2%, 56 +/-1%, 26 +/-1%, the completion rate of the CAL model is 36 +/-6%, 26 +/-2%, 9 +/-1%, and the task completion rate of the CILRS model is 66+2%,49 +/-5%, 23 +/-1%. For different end-to-end models, the model provided by the method achieves the best performance in most driving tasks, and particularly in a dynamic environment, the reliability of the model in the dynamic environment is fully explained.
Claims (6)
1. A semantic segmentation based unmanned vehicle motion planning method is characterized by comprising the following steps:
periodically acquiring continuous frame surrounding views in the left direction, the front direction and the right direction of a vehicle;
based on environmental information classification, respectively performing semantic segmentation on left and right direction current view data and continuous n frames of forward view data through a semantic segmentation network to obtain semantic segmentation images;
based on the environmental information classification, acquiring category weight information corresponding to the continuous n frames of forward images through a category weight network;
acquiring the speed and direction corresponding to each frame of view based on an expert database;
taking the segmented images, the category weight information and the speed and direction corresponding to the view as training data, training and learning the dynamic surround vision imitating learning network based on the neural network, and repeating the steps until a convergent dynamic surround vision imitating learning model is obtained; the dynamic surround vision simulation learning network comprises a feature extraction fusion network and a branch decision network, and adopts an L1 loss function; wherein:
the feature extraction fusion network fuses category weight information and a semantic segmentation image of a forward image, n 512-dimensional feature vectors are obtained from the fused image through a ResNet34 residual error network, the n 512-dimensional feature vectors obtain a 512-dimensional feature vector through a single-layer LSTM network, 64-dimensional feature vectors are obtained from the semantic segmentation image in the left direction and the semantic segmentation image in the right direction through the residual error network respectively, and a 640-dimensional combined feature vector is obtained by splicing the 512-dimensional feature vector and the two 64-dimensional feature vectors;
the branch decision network activates the corresponding branch network according to navigation information, and the navigation information c comprises four states of road following, road walking, left turning and right turning;
the L1 loss function is:
wherein,vehicle speed v at time t +1 t+1 Is predicted value of (4), (v) is greater than or equal to>Is the predicted value of the square disk angle at time t, a t Is the steering wheel angle at time t, v t+1 Vehicle speed at time t + 1;
the number of hidden nodes of the LSTM network is 128, and the updating process is as follows:
external state S based on last moment t-1 And input x of the current time t Calculating input gate i t And an output gate o t Forget door t And candidate statesThe calculation formula is as follows:
i t =σ(M i x t +N i S t-1 +b i )
o t =σ(M o x t +N o S t-1 +b o )
l t =σ(M l x t +N l S t-1 +b l )
wherein tanh is the activation function, M i And N i Is a parameter matrix, the value range is (0, 1), sigma is a logistic function, and the value range is (0, 1);
through forgetting door l t And an input gate i t To update the memory cell I t The update formula is:
passing the internal state to the external state S t External state S t Comprises the following steps:
S t =o t tanh(I t )
the 640-dimensional joint feature vector is:
wherein e is t 512-dimensional feature vectors and two 64-dimensional feature vectors are spliced to obtain a 640-dimensional joint feature vector,for corresponding 512-dimensional feature vectors for n successive frames of forward images>640-dimensional associated feature vectors corresponding to right and left images at time t, respectively>Semantic segmentation images R corresponding to the forward image, the right image and the left image at the time t respectively 34 Is 34 layer residual networkFunction, R 18 Is a 18-layer residual network function, L is an LSTM network function;
the dynamic surround vision imitation learning model is as follows:
(v t+1 ,a t )=F(e t ,c)
(v t+1 ,a t )=A c (e t )
wherein v is t+1 Is the vehicle speed at time t +1, a t T is the steering wheel angle, c is the navigation command, and F is the vehicle speed v t+1 And steering wheel angle a t And e t And navigation instructions c, A c A decision branch network corresponding to the navigation instruction c;
surrounding views in the left direction, the front direction and the right direction of the vehicle are collected, semantic segmentation images of the surrounding views in the left direction, the front direction and the right direction and category weight information corresponding to continuous n frames of forward images are obtained and input to a dynamic surrounding vision simulation learning model to obtain a planning speed and a planning direction;
and carrying out motion planning on the unmanned vehicle based on the planning speed and the planning direction.
2. The unmanned vehicle motion planning method of claim 1, wherein the semantic segmentation network comprises a residual network ResNet, a full convolutional network FCN, a void space convolutional pooling pyramid, and a decoder; the method for obtaining the semantic segmentation image by respectively performing semantic segmentation on the left-direction current view data, the right-direction current view data and the continuous n frames of forward view data through the semantic segmentation network specifically comprises the following steps of:
performing semantic segmentation on the current left and right view data through a residual error network ResNet and a full convolution network FCN, and performing semantic segmentation on the current and previous n-1 frame forward view data;
based on the environmental information classification, extracting and classifying semantic segmentation image characteristic information by adopting a cavity space convolution pooling pyramid;
and recovering the details and the space dimensionality of the image features through a decoder based on the classified image feature information to obtain a final semantic segmentation image.
3. The unmanned vehicle motion planning method of claim 1, wherein the category weight network specifically is: and forming a residual unit by four residual blocks, copying the fourth residual block for three times, connecting the copied fourth residual block to the residual unit in a cascading manner, and simultaneously performing parallel hollow space convolution pooling pyramid structure.
4. The unmanned vehicle motion planning method of claim 1, wherein the collection period is 0.3s, n =4, and the classification of the environmental information comprises: empty, buildings, fences, pedestrians, pillars, lane dividing lines, roads, sidewalks, vegetation, vehicles, walls, signal lamps and other unidentified types.
5. The unmanned vehicle motion planning method according to claim 1, wherein a PID controller is used to plan the motion of the unmanned vehicle, and the PID controller uses a PI control algorithm, specifically:
6. An unmanned vehicle motion planning system based on the unmanned vehicle motion planning method based on semantic segmentation according to any one of claims 1 to 5, which is characterized by comprising a visual sensor, a category weight module, a semantic segmentation module, a simulation learning module, a speed control module and a steering control module, wherein:
the vision sensor is used for periodically acquiring surrounding views of the left direction, the middle direction and the right direction of the vehicle;
the category weight module is used for acquiring category weight information corresponding to the forward image through a category weight network;
the semantic segmentation module is used for performing semantic segmentation on left-direction current view data and right-direction current view data and continuous n frames of forward view data respectively by a semantic segmentation network to obtain a semantic segmentation image;
the simulation learning module performs network learning based on neural network learning to obtain a converged dynamic surround vision simulation learning model; the simulation learning module comprises an image fusion module and a branch decision module, wherein the image fusion module is used for fusing category weight information with semantic segmentation images of each frame of forward images, acquiring n 512-dimensional feature vectors from the fused images through a ResNet34 residual error network, acquiring a 512-dimensional feature vector from the n 512-dimensional feature vectors through a single-layer LSTM network, respectively acquiring 64-dimensional feature vectors from the semantic segmentation images in the left direction and the semantic segmentation images in the right direction through the residual error network, and splicing the 512-dimensional feature vectors with the two 64-dimensional feature vectors to acquire a 640-dimensional combined feature vector; the branch decision module activates a corresponding branch network according to navigation information, wherein the navigation information c comprises four states of road following, line, left turn and right turn;
the speed control module controls the speed of the unmanned vehicle by simulating the planned speed output by the learning module;
the steering control module controls steering of the unmanned vehicle by simulating the planned direction output by the learning module.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111301912.9A CN114035575B (en) | 2021-11-04 | 2021-11-04 | Unmanned vehicle motion planning method and system based on semantic segmentation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111301912.9A CN114035575B (en) | 2021-11-04 | 2021-11-04 | Unmanned vehicle motion planning method and system based on semantic segmentation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114035575A CN114035575A (en) | 2022-02-11 |
CN114035575B true CN114035575B (en) | 2023-03-31 |
Family
ID=80142861
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111301912.9A Active CN114035575B (en) | 2021-11-04 | 2021-11-04 | Unmanned vehicle motion planning method and system based on semantic segmentation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114035575B (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114035575B (en) * | 2021-11-04 | 2023-03-31 | 南京理工大学 | Unmanned vehicle motion planning method and system based on semantic segmentation |
CN116048096B (en) * | 2023-02-23 | 2024-04-30 | 南京理工大学 | Unmanned vehicle movement planning method based on hierarchical depth perception |
CN117078923B (en) * | 2023-07-19 | 2024-07-16 | 苏州大学 | Automatic driving environment-oriented semantic segmentation automation method, system and medium |
CN117184105B (en) * | 2023-07-20 | 2024-03-26 | 清华大学 | Steering angular velocity prediction method and device based on multi-mode data fusion |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109410307A (en) * | 2018-10-16 | 2019-03-01 | 大连理工大学 | A kind of scene point cloud semantic segmentation method |
CN110189337A (en) * | 2019-05-31 | 2019-08-30 | 广东工业大学 | A kind of automatic Pilot image, semantic dividing method |
DE102018128531A1 (en) * | 2018-11-14 | 2020-05-14 | Valeo Schalter Und Sensoren Gmbh | System and method for analyzing a three-dimensional environment represented by a point cloud through deep learning |
CN111210435A (en) * | 2019-12-24 | 2020-05-29 | 重庆邮电大学 | Image semantic segmentation method based on local and global feature enhancement module |
CN111443701A (en) * | 2018-12-29 | 2020-07-24 | 南京理工大学 | Unmanned vehicle/robot behavior planning method based on heterogeneous deep learning |
CN111832453A (en) * | 2020-06-30 | 2020-10-27 | 杭州电子科技大学 | Unmanned scene real-time semantic segmentation method based on double-path deep neural network |
CN113486568A (en) * | 2021-04-06 | 2021-10-08 | 南京晓庄学院 | Vehicle control dynamic simulation learning algorithm based on surround vision |
CN114035575A (en) * | 2021-11-04 | 2022-02-11 | 南京理工大学 | Unmanned vehicle motion planning method and system based on semantic segmentation |
-
2021
- 2021-11-04 CN CN202111301912.9A patent/CN114035575B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109410307A (en) * | 2018-10-16 | 2019-03-01 | 大连理工大学 | A kind of scene point cloud semantic segmentation method |
DE102018128531A1 (en) * | 2018-11-14 | 2020-05-14 | Valeo Schalter Und Sensoren Gmbh | System and method for analyzing a three-dimensional environment represented by a point cloud through deep learning |
CN111443701A (en) * | 2018-12-29 | 2020-07-24 | 南京理工大学 | Unmanned vehicle/robot behavior planning method based on heterogeneous deep learning |
CN110189337A (en) * | 2019-05-31 | 2019-08-30 | 广东工业大学 | A kind of automatic Pilot image, semantic dividing method |
CN111210435A (en) * | 2019-12-24 | 2020-05-29 | 重庆邮电大学 | Image semantic segmentation method based on local and global feature enhancement module |
CN111832453A (en) * | 2020-06-30 | 2020-10-27 | 杭州电子科技大学 | Unmanned scene real-time semantic segmentation method based on double-path deep neural network |
CN113486568A (en) * | 2021-04-06 | 2021-10-08 | 南京晓庄学院 | Vehicle control dynamic simulation learning algorithm based on surround vision |
CN114035575A (en) * | 2021-11-04 | 2022-02-11 | 南京理工大学 | Unmanned vehicle motion planning method and system based on semantic segmentation |
Non-Patent Citations (1)
Title |
---|
无人驾驶中的场景实时语义分割方法;秦飞巍;《计算机辅助设计与图形学学报》;20210730;第33卷(第7期);第1026-1038页 * |
Also Published As
Publication number | Publication date |
---|---|
CN114035575A (en) | 2022-02-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114035575B (en) | Unmanned vehicle motion planning method and system based on semantic segmentation | |
CN112215337B (en) | Vehicle track prediction method based on environment attention neural network model | |
CN112099496B (en) | Automatic driving training method, device, equipment and medium | |
Fernando et al. | Deep inverse reinforcement learning for behavior prediction in autonomous driving: Accurate forecasts of vehicle motion | |
CN112703459B (en) | Iterative generation of confrontational scenarios | |
CN112888612A (en) | Autonomous vehicle planning | |
CN113044064B (en) | Vehicle self-adaptive automatic driving decision method and system based on meta reinforcement learning | |
CN113715842B (en) | High-speed moving vehicle control method based on imitation learning and reinforcement learning | |
CN112232490A (en) | Deep simulation reinforcement learning driving strategy training method based on vision | |
CN112784485B (en) | Automatic driving key scene generation method based on reinforcement learning | |
Qiao et al. | Behavior planning at urban intersections through hierarchical reinforcement learning | |
CN113255054A (en) | Reinforcement learning automatic driving method based on heterogeneous fusion characteristics | |
CN113552883B (en) | Ground unmanned vehicle autonomous driving method and system based on deep reinforcement learning | |
CN110281949A (en) | A kind of automatic Pilot unifies hierarchical decision making method | |
Sun et al. | Human-like highway trajectory modeling based on inverse reinforcement learning | |
CN118097989B (en) | Multi-agent traffic area signal control method based on digital twin | |
CN114620059A (en) | Automatic driving method and system thereof, and computer readable storage medium | |
US20240208546A1 (en) | Predictive models for autonomous vehicles based on object interactions | |
CN111552294A (en) | Outdoor robot path-finding simulation system and method based on time dependence | |
CN113486568A (en) | Vehicle control dynamic simulation learning algorithm based on surround vision | |
CN111443701A (en) | Unmanned vehicle/robot behavior planning method based on heterogeneous deep learning | |
CN114779764A (en) | Vehicle reinforcement learning motion planning method based on driving risk analysis | |
CN116048096B (en) | Unmanned vehicle movement planning method based on hierarchical depth perception | |
US11808582B1 (en) | System processing scenario objects during simulation | |
CN111508256A (en) | Traffic information reconstruction method based on regional time-space domain and intelligent traffic system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |