CN115782835A - Automatic parking remote driving control method for passenger boarding vehicle - Google Patents
Automatic parking remote driving control method for passenger boarding vehicle Download PDFInfo
- Publication number
- CN115782835A CN115782835A CN202310084318.1A CN202310084318A CN115782835A CN 115782835 A CN115782835 A CN 115782835A CN 202310084318 A CN202310084318 A CN 202310084318A CN 115782835 A CN115782835 A CN 115782835A
- Authority
- CN
- China
- Prior art keywords
- image
- information
- vector
- driving control
- boarding vehicle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 36
- 238000004458 analytical method Methods 0.000 claims abstract description 48
- 239000013598 vector Substances 0.000 claims description 72
- 230000006870 function Effects 0.000 claims description 43
- 230000007246 mechanism Effects 0.000 claims description 24
- 238000011176 pooling Methods 0.000 claims description 23
- 239000011159 matrix material Substances 0.000 claims description 18
- 238000012549 training Methods 0.000 claims description 14
- 238000012545 processing Methods 0.000 claims description 11
- 238000001514 detection method Methods 0.000 claims description 10
- 238000000605 extraction Methods 0.000 claims description 9
- 230000008569 process Effects 0.000 claims description 8
- 238000013527 convolutional neural network Methods 0.000 claims description 6
- 238000012935 Averaging Methods 0.000 claims description 3
- 238000004364 calculation method Methods 0.000 claims description 3
- 238000010586 diagram Methods 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 238000010223 real-time analysis Methods 0.000 claims 1
- 230000000694 effects Effects 0.000 abstract description 2
- 238000010191 image analysis Methods 0.000 abstract description 2
- 230000004913 activation Effects 0.000 description 6
- 230000009466 transformation Effects 0.000 description 6
- 230000008859 change Effects 0.000 description 5
- 230000001133 acceleration Effects 0.000 description 4
- 230000003993 interaction Effects 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 3
- 230000003137 locomotive effect Effects 0.000 description 2
- 210000002569 neuron Anatomy 0.000 description 2
- 238000007781 pre-processing Methods 0.000 description 2
- 238000012216 screening Methods 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000011217 control strategy Methods 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000008034 disappearance Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
Images
Classifications
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Image Analysis (AREA)
Abstract
The invention provides a remote driving control method for automatic parking of a passenger boarding vehicle, which belongs to the field of intelligent driving, adopts a brand-new analysis technology framework combining local image analysis with global image and voice analysis to realize the remote driving control method for the boarding vehicle, carries out cooperative analysis in real time according to passenger voice signals in the boarding vehicle, and greatly improves the safety and the accuracy of the remote driving control method for the boarding vehicle; the attention cooperative analysis model is further pertinently improved, so that the analysis result is remarkably improved, and the control effect of boarding of passengers is further improved.
Description
Technical Field
The invention belongs to the field of intelligent driving, and particularly relates to an automatic parking remote driving control method for a passenger boarding vehicle.
Background
At present, in the current situation of automatic control of boarding vehicles, due to poor optimization of a control strategy, the existing automatic driving method does not consider sudden change factors of temporary plans of boarding vehicles during driving under the condition that passengers have emergency events in the process. The existing remote automatic parking driving control method has a deep learning strategy of adopting a convolutional neural network structure, the deep learning method mostly adopts a convolutional neural network to realize road image signals, and obstacle avoidance, path planning and the like are carried out, but the identification precision is limited, and the extraction capability of a neural network algorithm on large-view global correlation information is poor. In addition, the conventional method is used for a deep learning method of automatic driving, and a single convolutional neural network is mostly adopted to analyze image signals in road conditions, so that sudden conditions in the actual process of the vehicle and the requirements of passengers in the riding process are not considered. Therefore, the technical problem which needs to be solved at present is to carry out automatic parking remote driving control on passengers boarding a locomotive by carrying out collaborative analysis in real time according to the demands of the passengers in the locomotive.
Disclosure of Invention
The invention aims to solve the problems in the prior art and provides an automatic parking remote driving control method for a passenger boarding vehicle, which realizes automatic driving of the passenger boarding vehicle and sends the passenger to a designated boarding place.
The invention is realized by the following technical scheme:
step S100: acquiring signals, namely acquiring local image signals and global image signals of the surrounding environment through camera equipment arranged on the boarding roof, and acquiring real-time voice signals of passengers through voice acquisition equipment in a passenger boarding vehicle;
step S110: the images obtained by the camera comprise normal images and wide-angle high-resolution images with different magnification factors and are used for capturing local image signals and global image signals in road conditionsZ a Wherein the local image signal is represented asZ a1 The global image signal is represented asZ a2 Voice signal acquired by voice signal acquisition moduleZ b 。
Step S200: based on the image signal and the voice signal obtained in S100, an image signal processing module and a voice signal coding module are constructed to preprocess signals in different modes;
the signal preprocessing of the invention comprises the following steps: for image signal, adopting value normalization method, and for input signal vectorZ a The preprocessed image signal isX z = (Z a -Z min )/(Z max -Z min ) Wherein Z is min Represents the sameZ a Minimum value in signal, Z max Is composed ofZ a Maximum signal value in the signal respectively obtaining the preprocessed local image signalXAnd a global image signalX 1 。
For speech signalsZ b The collected voice signal is vector-coded by adopting a voice vector coding algorithm to obtain a voice signalX 2 ;
Step S300: training a local information analysis model based on the preprocessed local image signals, and training an image and voice collaborative analysis model based on the preprocessed global image signals and voice signals;
the local information analysis model realizes road identification, pedestrian identification, vehicle identification, signal lamp identification and dynamic obstacle identification in the driving process, and the specific steps are as follows:
step S311: the input image signal is a preprocessed partial image signalWhereinC×HWhich represents the size of the image,Wrepresenting the number of images;
step S312: to pairXFeature extraction is carried out through a deep convolution feature module to obtain a feature mapLocal key information is selected by adopting multi-region pooling operation, and the specific calculation steps are as follows:
step S313: for characteristic diagramX C Randomly dividing N image blocks with different sizes, and calculating by maximum poolingWThe image block with the largest pixel value is contained in the same position of the image, and the maximum value solving function of the image block isWhereinIs a function of the maximum value of the signal,all image block vectors representing the kth position in the W images;
finally, the image blocks with the maximum pixel value in each corresponding position are spliced into an image feature mapWhereinIn the formulaRepresenting the characteristics of the spliced image, k is the same as (1,N),is represented at a positionkOn the upper partWThe largest image block of a picture is,splicing functions for image block features;
step S314: image block vector using average pooling operationProcessing, averaging pooled operating functionsWhereinRepresents the average pooling function and is the average pooling function,is represented at a positionkOn the upper partWThe image block characteristics of the mean value in an image,all image block vectors representing the kth position in the W images; and the image block features after the average pooling processing are spliced into an image feature mapWherein,k∈(1,N),Splicing functions for image block features;
step S315: for the spliced image feature mapAndthe vector is processed into a one-dimensional vector through a convolution layer, a maximum pooling layer and an average pooling layer and is input into a full-link layer, and finally, a nonlinear function is adopted for processingBy usingsoftmaxThe function, its expression is as follows:in the formula a i 、a j Weight, x, representing input vector i And x j As input variables, E 1 Is the output class number.
In addition, the generation of training samples in the model training in the embodiment is well known to those skilled in the art.
After the local image signals are processed, in order to fuse voice interaction information and carry out cooperative analysis by combining a global image signal and a voice signal of the position of the boarding vehicle in real time, information extraction is carried out on a global wide-angle image shot by a camera system through a convolutional neural network, and then the voice signal and the image signal obtained after vector coding are processed through an automatic supervision coding module at the same time. The method comprises the following specific steps:
step S320: in the collaborative analysis model, assumptionsZ a Global image signal inX 1 The vector processed by the feature extraction module isX a1 ,Speech signalX 2 The coded vector isX a2 And adopting a vector sequential splicing mode to carry out image signal processingX a1 And coded speech signalX a2 Are combined into a vectorX a The module comprises four self-learning matrixes which are respectively used for representing position information, depth information, content information and relevance informationA,Q、K、VVector representation with dimensions of then vectorX a Multiplication by a weight matrix (linear transformation), the corresponding matrix of which is defined as W a 、W q 、W k 、W v And obtaining a corresponding self-learning input vector, wherein the self-supervision learning strategy based on the four vectors is as follows:position information vector in formula A = X a ×W a Depth information vector Q = X a ×W q In, innerVolume information vector K = X a ×W k And a relevancy information vector V = X a ×W v ,d k Is a vectorKDimension of (2), self-attention mechanism functionThe output value represents the degree of association between the information with high degree of association between the image and the voice signal and the detection object and between the information when the model decodes the vector information.
Step S321: the invention can adopt sine coding mode and other coding modes aiming at the coding modes of the image block and the voice signal. The coding mode is changed, so that the relative relevance between different signals can be ensured to be learned by the constructed model.
In the self-attention module. Slave typeAs can be seen in (1), attention mechanism function output values and vectorsVAnd AQK T Is proportional, i.e. the signal processed by the module is determined by the associated information learned from the signal itself.
Step S322: vector transformation matrix W for position information, depth information, content information and information associated with voice a ,W q 、W k And W v And the model is continuously optimized in the training process of the model, so that the model is ensured to realize the global image signal learning. Based on the attention mechanism module, the self-attention mechanism module is improved in the following mode: in the formulaIs a splicing function with a total dimension after splicing ofd model Then, the corresponding parameter matrices are:andwherein R is the output variable domain of the self-attention mechanism module,respectively represent the number of columns of the matrix,for the number of rows in the matrix,Focus() is a function of the self-attention mechanism,H i is shown asThe attention mechanism sub-module outputs.
Step S323: also, the self-attention module transforms the matrix through self-learning according to the principle of the self-attention mechanismRespectively calculating the attention degree of a single self-attention mechanism, wherein the output value of the attention degree corresponds to the emphasis degree of the attention message; all the output values of the self-attention module are spliced through a total coefficient matrixThe attention degree of the image signal and the voice signal is output, so that the interaction and the collaborative learning of the image signal and the voice signal are realized.
Step S324: in a collaborative analysis model, the feature dimensionality is reduced through an improved attention mechanism, the network layer can automatically screen out a key feature set through continuous training, and then the combination of features is realized through a nonlinear function, so that the relevance is further enhanced.
Step S325: in a network, different activation functions can be used, and the invention adopts a nonlinear function:in the formulaW z This can be achieved by fully connecting layers, representing two vectors multiplied by corresponding elements,b 0 in order to be a term of the offset,xrepresenting input variables, notLinear piecewise exponential functionNLEThe expression of (c) is as follows:the piecewise activation function can improve the model speed and inhibit irrelevant information on one hand, and on the other hand, the function retains the attribute of nonlinear transformation, the piecewise activation function is different from linearity, and the information obtained after screening is further strengthened by adopting exponential segmentation. The piecewise function is applied to the input variablexWhen the slope is negative, a smaller slope is adopted, and the phenomenon that network neurons do not work due to the fact that gradients in the network disappear is avoided.
And outputting a global detection and analysis result through the collaborative analysis model, wherein the output result comprises road information, distance information of vehicles before and after the distance, the number information of vehicles in other lanes, parking point pedestrian condition information and emergency parking information.
Step S400: selecting a corresponding executed state instruction as a remote driving control instruction based on two outputs of the S300 local information analysis model and the collaborative analysis model;
the module mainly completes simple task analysis in road conditions based on local attention, including traffic light identification and pedestrian and obstacle detection. And outputting corresponding waiting and parking state instructions.
The cooperative analysis model learns the global correlation information, controls and infers based on the detection and analysis results, and outputs driving state instructions mainly comprising road switching, acceleration, deceleration, left and right turning and parking point intelligent identification parking points. If the cooperative analysis model detects that the number of the road vehicles is less, switching to a road model with better road condition execution according to the model lane change instruction; if the distance between the vehicle and the front vehicle is detected to be close, the signal corresponds to a deceleration signal, and if the distance between the vehicle and the front vehicle is detected to be far, the signal is an acceleration signal. The information obtained by sorting the different lane information corresponds to lane change information. If the information of the parking point passengers is detected, a parking instruction is sent out; and if the detected passenger information comprises the information of seeking help, decelerating and parking, giving a parking instruction by combining the road condition.
Compared with the prior art, the invention has the beneficial effects that: the invention provides a brand-new boarding vehicle remote driving control method based on local image analysis and global voice analysis, which can carry out cooperative analysis in real time according to passenger voice signals in a vehicle, thereby greatly improving the safety and the accuracy of the boarding vehicle remote driving control method; and secondly, the attention cooperative analysis model is improved in a targeted manner, so that the analysis result is obviously improved, and the control effect of boarding of passengers is further improved.
Drawings
FIG. 1 is a flow chart of the passenger boarding vehicle automatic parking remote driving control method.
Detailed Description
The present invention is described in further detail below with reference to FIG. 1:
step S100: acquiring signals, namely acquiring local image signals and global image signals of the surrounding environment through camera equipment arranged on the boarding vehicle roof, and acquiring real-time voice signals of passengers through voice acquisition equipment in a passenger boarding vehicle;
step S100: acquiring signals, namely acquiring local image signals and global image signals of the surrounding environment through camera equipment arranged on the boarding vehicle roof, and acquiring real-time voice signals of passengers through voice acquisition equipment in a passenger boarding vehicle;
step S110: the images obtained by the camera comprise normal images and wide-angle high-resolution images with different magnification factors and are used for capturing local image signals and global image signals in road conditionsZ a Wherein the local image signal is represented asZ a1 The global image signal is represented asZ a2 Voice signal obtained by voice signal acquisition moduleZ b 。
Step S200: based on the image signal and the voice signal obtained in S100, an image signal processing module and a voice signal coding module are constructed to preprocess signals in different modes;
the signal preprocessing of the invention comprises the following steps: for image signal, adopting value normalization method, and for input signal vectorZ a The pre-processed image signal isX z = (Z a -Z min )/(Z max -Z min ) Wherein Z is min Represents the sameZ a Minimum value in signal, Z max Is composed ofZ a Maximum signal value in the signal respectively obtaining the preprocessed local image signalXAnd a global image signalX 1 。
For speech signalsZ b The collected voice signal is vector-coded by adopting a voice vector coding algorithm to obtain a voice signalX 2 ;
Step S300: training a local information analysis model based on the preprocessed local image signals, and training an image and voice collaborative analysis model based on the preprocessed global image signals and voice signals;
the local information analysis model realizes road identification, pedestrian identification, vehicle identification, signal lamp identification and dynamic obstacle identification in the driving process, and the specific steps are as follows:
step S311: the input image signal is a preprocessed partial image signalIn whichC×HWhich represents the size of the image,Wrepresenting the number of images;
step S312: for is toXFeature extraction is carried out through a deep convolution feature module to obtain a feature mapLocal key information is selected by adopting multi-region pooling operation, and the specific calculation steps are as follows:
step S313: for characteristic diagramX C Randomly dividing N image blocks with different sizes, and calculating by maximum poolingWThe image block with the largest pixel value is contained in the same position of the image, and the maximum value solving function of the image block isIn whichIs a function of the maximum value of the signal,all image block vectors representing the kth position in the W images;
finally, the image blocks with the maximum pixel value in each corresponding position are spliced into an image feature mapIn whichIn the formulaRepresenting the characteristics of the spliced image, k is the same as (1,N),is represented at a positionkOn the upper partWThe largest image block in a picture is,splicing functions for image block features;
step S314: vector image block using average poolingProcessing, averaging pooled operating functionsIn whichRepresents the average pooling function and is the average pooling function,is represented at a positionkOn the upper partWThe image block characteristics of the mean value in an image,all image block vectors representing the kth position in the W images; and the image block features after the average pooling are spliced intoAn image feature mapWherein,k∈(1,N),Splicing functions for image block features;
step S315: for the spliced image feature mapAndprocessing the vector into a one-dimensional vector by a convolution layer, a maximum pooling layer and an average pooling layer, inputting the vector into a full-connection layer, and finally processing the vector by a nonlinear functionsoftmaxThe function, its expression is as follows:in the formula a i 、a j Weight, x, representing input vector i And x j As input variables, E 1 Is the output class number.
In addition, the generation of training samples in the model training in the embodiment is well known to those skilled in the art.
After the local image signals are processed, in order to fuse voice interaction information and carry out cooperative analysis by combining a global image signal and a voice signal of the position of the boarding vehicle in real time, information extraction is carried out on a global wide-angle image shot by a camera system through a convolutional neural network, and then the voice signal and the image signal obtained after vector coding are processed through an automatic supervision coding module at the same time. The method comprises the following specific steps:
step S320: in the collaborative analysis model, assumptionsZ a Global image signal inX 1 The vector processed by the feature extraction module isX a1 ,Speech signalX 2 The coded vector isX a2 Image signal is spliced by vector sequenceX a1 And coded speech signalX a2 Are combined into a vectorX a The module comprises four self-learning matrixes which are respectively used for representing position information, depth information, content information and relevance informationA,Q、K、VVector representation with dimensions of the vectorX a Multiplication by a weight matrix (linear transformation), the corresponding matrix of which is defined as W a 、W q 、W k 、W v And obtaining a corresponding self-learning input vector, wherein the self-supervision learning strategy based on the four vectors is as follows:position information vector in formula A = X a ×W a Depth information vector Q = X a ×W q Content information vector K = X a ×W k And degree of association information vector V = X a ×W v ,d k Is a vectorKDimension of (2), self-attention mechanism functionThe output value represents the degree of association between the information with high degree of association between the image and the voice signal and the detection object and between the information when the model decodes the vector information.
Step S321: the invention can adopt sine coding mode and other coding modes aiming at the coding modes of the image block and the voice signal. The coding mode is changed, so that the relative relevance between different signals can be ensured to be learned by the constructed model.
In the self-attention module. Slave typeAs can be seen in (1), attention mechanism function output values and vectorsVAnd AQK T Is proportional, i.e. the signal processed by the module is determined by the associated information learned from the signal itself.
Step S322: position information, depth information, content information, and phraseSound correlation degree information vector transformation matrix W a ,W q 、W k And W v And the model is continuously optimized in the training process of the model, so that the model is ensured to realize the global image signal learning. Based on the attention mechanism module, the self-attention mechanism module is improved in the following mode: in the formulaIs a splicing function with a total dimension after splicing ofd model Then, the corresponding parameter matrices are respectively:andwherein R is the output variable domain of the self-attention mechanism module,respectively represent the number of columns of the matrix,for the number of rows in the matrix,Focus() is a function of the self-attention mechanism,H i is shown asThe attention mechanism sub-module outputs.
Step S323: also, the self-attention module transforms the matrix through self-learning according to the principle of the self-attention mechanismRespectively calculating the attention degree of a single self-attention mechanism, wherein the output value of the attention degree corresponds to the emphasis degree of the attention message; all the output values of the self-attention module are spliced through a total coefficient matrixThe attention degrees of the image signal and the voice signal are output, so that the interaction and the cooperative learning of the image signal and the voice signal are realized.
Step S324: in a collaborative analysis model, the feature dimensionality is reduced through an improved attention mechanism, the network layer can automatically screen out a key feature set through continuous training, and then the combination of features is realized through a nonlinear function, so that the relevance is further enhanced.
Step S325: in a network, different activation functions can be used, and the invention adopts a nonlinear function:in the formulaW z This can be achieved by fully connecting layers, representing two vectors multiplied by corresponding elements,b 0 in order to be a term of the offset,xrepresenting input variables, non-linear piecewise exponential functionsNLEThe expression of (a) is as follows:the piecewise activation function can improve the model speed and inhibit irrelevant information on one hand, and on the other hand, the function retains the attribute of nonlinear transformation, the piecewise activation function is different from linearity, and the information obtained after screening is further strengthened by adopting exponential segmentation. The piecewise function is applied to the input variablexWhen the slope is negative, a smaller slope is adopted, and the phenomenon that network neurons do not work due to disappearance of gradients in the network is avoided.
And outputting a global detection and analysis result through the collaborative analysis model, wherein the output result comprises road information, distance information of vehicles before and after the distance, the number information of vehicles in other lanes, parking point pedestrian condition information and emergency parking information.
Step S400: selecting a corresponding executed state instruction as a remote driving control instruction based on two outputs of the S300 local information analysis model and the collaborative analysis model;
the local attention mechanism-based module mainly completes simple task analysis in road conditions, including traffic light identification and pedestrian and obstacle detection. And outputting corresponding waiting and parking state instructions.
The cooperative analysis model learns the global correlation information, controls and infers based on the detection and analysis results, and outputs driving state instructions mainly comprising road switching, acceleration, deceleration, left and right turning and parking point intelligent identification parking points. If the cooperative analysis model detects that the number of the road vehicles is less, switching to a road model with better road condition execution according to the model lane change instruction; if the distance between the vehicle and the front vehicle is detected to be close, the signal corresponds to a deceleration signal, and if the distance between the vehicle and the front vehicle is detected to be far, the signal is an acceleration signal. The information obtained by sorting the different lane information corresponds to lane change information. If the parking point passenger information is detected, a parking instruction is sent out; and if the detected passenger information comprises help seeking, deceleration and parking information, giving a parking instruction by combining the road condition.
In the description of the present invention, unless otherwise expressly specified or limited, the terms "connected" and "connected" are to be construed broadly, e.g., as meaning a fixed connection, a removable connection, or an integral connection; can be mechanically or electrically connected; may be directly connected or indirectly connected through an intermediate. Those skilled in the art understand the specific meanings of the above terms in the present invention according to specific situations.
In the description of the present invention, unless otherwise specified, the terms "upper", "lower", "left", "right", "inner", "outer", and the like, indicate orientations or positional relationships based on the orientations or positional relationships shown in the drawings, are merely for convenience of description and simplification of description, and do not indicate or imply that the device or element referred to must have a particular orientation, be constructed and operated in a particular orientation, and thus, should not be construed as limiting the present invention.
Finally, it should be noted that the above-mentioned technical solution is only one embodiment of the present invention, and it will be apparent to those skilled in the art that various modifications and variations can be easily made based on the application method and principle of the present invention disclosed, and the method is not limited to the above-mentioned specific embodiment of the present invention, so that the above-mentioned embodiment is only preferred, and not restrictive.
Claims (10)
1. A passenger boarding vehicle automatic parking remote driving control method is characterized by comprising the following steps:
step S100: acquiring signals, namely acquiring local image signals and global image signals of the surrounding environment through camera equipment arranged on the boarding vehicle roof, and acquiring real-time voice signals of passengers through voice acquisition equipment in a passenger boarding vehicle;
step S200: based on the image signal and the voice signal obtained in S100, an image signal processing module and a voice signal coding module are constructed to preprocess signals in different modes;
step S300: training a local information analysis model based on the preprocessed local image signals, and training an image and voice collaborative analysis model based on the preprocessed global image signals and voice signals;
step S400: and selecting a corresponding executed state instruction as a remote driving control instruction through a path selection and planning module based on the two outputs of the S300 local information analysis model and the collaborative analysis model.
2. The passenger boarding vehicle automatic parking remote driving control method according to claim 1, characterized in that step S200: aiming at the image signals, a value normalization method is adopted to respectively obtain the preprocessed local image signalsXAnd a global image signalX 1 。
3. The passenger boarding vehicle automatic parking remote driving control method according to claim 1, characterized in that step S200: for speech signalsZ b The collected voice signal is vector-coded by adopting a voice vector coding algorithm to obtain a voice signalX 2 。
4. The passenger boarding vehicle automatic parking remote driving control method according to claim 1, characterized in that: the local information analysis model realizes road recognition, pedestrian recognition, vehicle recognition, signal lamp recognition and dynamic obstacle recognition in the driving process.
5. The passenger boarding vehicle automatic parking remote driving control method according to claim 1, characterized in that the local information analysis model comprises the following specific steps: step S311: the input image signal is a preprocessed partial image signalWhereinC×HWhich is representative of the size of the image,Wrepresenting the number of images;
step S312: to pairXFeature extraction is carried out through a deep convolution feature module to obtain a feature mapLocal key information is selected by adopting multi-region pooling operation, and the specific calculation steps are as follows:
step S313: for characteristic diagramX C Randomly dividing N image blocks with different sizes, and calculating by maximum poolingWThe image block with the largest pixel value is contained in the same position of the image, and the maximum value solving function of the image block isWhereinIs a function of the maximum value of the signal,all image block vectors representing the kth position in the W images;
finally, the image blocks with the maximum pixel value in each corresponding position are spliced into an image feature mapWhereinIn the formulaRepresenting the characteristics of the spliced image, k is the same as (1,N),is represented at a positionkOn the upper partWThe largest image block of a picture is,splicing functions for image block features;
step S314: vector image block using average poolingProcessing, averaging pooled operating functionsWhereinRepresents the average pooling function of the samples of the sample,is represented at a positionkUpper part ofWThe image block characteristics of the mean value in an image,all image block vectors representing the kth position in the W images; and the image block features after the average pooling processing are spliced into an image feature mapWherein,k∈(1,N),Splicing functions for image block features;
6. The passenger boarding vehicle automatic parking remote driving control method according to claim 5, wherein step S315 is adopted in a full connection layersoftmaxThe function, its expression is as follows:
7. The passenger boarding vehicle automatic parking remote driving control method according to claim 1, characterized in that in an image and voice cooperative analysis model: the global image signal and the voice signal of the position of the boarding vehicle are combined for real-time analysis, information extraction is carried out on a global wide-angle image shot by a camera system through a convolutional neural network, and then the voice signal and the image signal obtained after vector-based coding are processed through an automatic supervision coding module at the same time.
8. The passenger boarding vehicle automatic parking remote driving control method according to claim 6, characterized in that: the image and voice cooperative analysis model comprises four self-learning matrixes which are respectively used for representing position information, depth information, content information and relevance information and are respectively used forA,Q、K、VA vector representation; then vector X is added a Multiplied by a weight matrix, whichThe correspondence matrix is defined as W a 、W q 、W k 、W v And obtaining a corresponding self-learning input vector, wherein the self-supervision learning strategy based on the four vectors is as follows:position information vector in formula A = X a ×W a Depth information vector Q = X a ×W q Content information vector K = X a ×W k And degree of association information vector V = X a ×W v ,d k Is a vectorKDimension of (2), self-attention mechanism functionThe output value represents the degree of association between the information with high degree of association between the image and the voice signal and the detection object and between the information when the model decodes the vector information.
9. The passenger boarding vehicle automatic parking remote driving control method according to claim 7, characterized in that the self-attention mechanism module is improved in the following manner: in the formulaIs a splicing function with a total dimension after splicing ofd model Then, the corresponding parameter matrices are respectively:andwherein R is the output variable domain of the self-attention mechanism module,respectively represent the number of columns of the matrix,for the number of rows in the matrix,Focus() is a function of the self-attention mechanism,H i is shown asThe attention mechanism sub-module outputs.
10. The passenger boarding vehicle automatic parking remote driving control method according to claim 1, characterized in that: and outputting a global detection and analysis result through the collaborative analysis model, wherein the output result comprises road information, distance information of vehicles before and after the distance, the number information of vehicles in other lanes, parking point pedestrian condition information and emergency parking information.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310084318.1A CN115782835B (en) | 2023-02-09 | 2023-02-09 | Automatic parking remote driving control method for passenger boarding vehicle |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310084318.1A CN115782835B (en) | 2023-02-09 | 2023-02-09 | Automatic parking remote driving control method for passenger boarding vehicle |
Publications (2)
Publication Number | Publication Date |
---|---|
CN115782835A true CN115782835A (en) | 2023-03-14 |
CN115782835B CN115782835B (en) | 2023-04-28 |
Family
ID=85430541
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310084318.1A Active CN115782835B (en) | 2023-02-09 | 2023-02-09 | Automatic parking remote driving control method for passenger boarding vehicle |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115782835B (en) |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190187707A1 (en) * | 2017-12-18 | 2019-06-20 | PlusAI Corp | Method and system for personalized driving lane planning in autonomous driving vehicles |
CN110099836A (en) * | 2017-01-17 | 2019-08-06 | Lg 电子株式会社 | The method of vehicle and control display therein |
CN110162040A (en) * | 2019-05-10 | 2019-08-23 | 重庆大学 | A kind of low speed automatic Pilot trolley control method and system based on deep learning |
CN110758241A (en) * | 2019-08-30 | 2020-02-07 | 华为技术有限公司 | Occupant protection method and apparatus |
CN111968338A (en) * | 2020-07-23 | 2020-11-20 | 南京邮电大学 | Driving behavior analysis, recognition and warning system based on deep learning and recognition method thereof |
CN113614749A (en) * | 2021-06-25 | 2021-11-05 | 华为技术有限公司 | Processing method, device and equipment of artificial intelligence model and readable storage medium |
CN115344049A (en) * | 2022-09-14 | 2022-11-15 | 江苏天一航空工业股份有限公司 | Automatic path planning and vehicle control method and device for passenger boarding vehicle |
CN115662166A (en) * | 2022-09-19 | 2023-01-31 | 长安大学 | Automatic driving data processing method and automatic driving traffic system |
-
2023
- 2023-02-09 CN CN202310084318.1A patent/CN115782835B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110099836A (en) * | 2017-01-17 | 2019-08-06 | Lg 电子株式会社 | The method of vehicle and control display therein |
US20190187707A1 (en) * | 2017-12-18 | 2019-06-20 | PlusAI Corp | Method and system for personalized driving lane planning in autonomous driving vehicles |
CN110162040A (en) * | 2019-05-10 | 2019-08-23 | 重庆大学 | A kind of low speed automatic Pilot trolley control method and system based on deep learning |
CN110758241A (en) * | 2019-08-30 | 2020-02-07 | 华为技术有限公司 | Occupant protection method and apparatus |
CN111968338A (en) * | 2020-07-23 | 2020-11-20 | 南京邮电大学 | Driving behavior analysis, recognition and warning system based on deep learning and recognition method thereof |
CN113614749A (en) * | 2021-06-25 | 2021-11-05 | 华为技术有限公司 | Processing method, device and equipment of artificial intelligence model and readable storage medium |
CN115344049A (en) * | 2022-09-14 | 2022-11-15 | 江苏天一航空工业股份有限公司 | Automatic path planning and vehicle control method and device for passenger boarding vehicle |
CN115662166A (en) * | 2022-09-19 | 2023-01-31 | 长安大学 | Automatic driving data processing method and automatic driving traffic system |
Also Published As
Publication number | Publication date |
---|---|
CN115782835B (en) | 2023-04-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110647839B (en) | Method and device for generating automatic driving strategy and computer readable storage medium | |
CN106599773B (en) | Deep learning image identification method and system for intelligent driving and terminal equipment | |
CN110356412B (en) | Method and apparatus for automatic rule learning for autonomous driving | |
CN112731925B (en) | Cone barrel identification and path planning and control method for formula car | |
CN114418895A (en) | Driving assistance method and device, vehicle-mounted device and storage medium | |
CN110516380B (en) | Deep reinforcement test method and system based on vehicle driving simulation data | |
CN111931683B (en) | Image recognition method, device and computer readable storage medium | |
CN110281949B (en) | Unified hierarchical decision-making method for automatic driving | |
CN112417973A (en) | Unmanned system based on car networking | |
CN112489072B (en) | Vehicle-mounted video perception information transmission load optimization method and device | |
CN116630702A (en) | Pavement adhesion coefficient prediction method based on semantic segmentation network | |
CN113379711A (en) | Image-based urban road pavement adhesion coefficient acquisition method | |
CN117237884A (en) | Interactive inspection robot based on berth positioning | |
CN112009491B (en) | Deep learning automatic driving method and system based on traffic element visual enhancement | |
CN115880658A (en) | Automobile lane departure early warning method and system under night scene | |
CN114782915A (en) | Intelligent automobile end-to-end lane line detection system and equipment based on auxiliary supervision and knowledge distillation | |
CN117115690A (en) | Unmanned aerial vehicle traffic target detection method and system based on deep learning and shallow feature enhancement | |
CN116729433A (en) | End-to-end automatic driving decision planning method and equipment combining element learning multitask optimization | |
CN116630920A (en) | Improved lane line type identification method of YOLOv5s network model | |
CN115782835A (en) | Automatic parking remote driving control method for passenger boarding vehicle | |
CN111160230B (en) | Road irregular area detection network based on deep learning | |
CN111931768A (en) | Vehicle identification method and system capable of self-adapting to sample distribution | |
CN117612140B (en) | Road scene identification method and device, storage medium and electronic equipment | |
CN115131762B (en) | Vehicle parking method, system and computer readable storage medium | |
CN110991337B (en) | Vehicle detection method based on self-adaptive two-way detection network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
EE01 | Entry into force of recordation of patent licensing contract |
Application publication date: 20230314 Assignee: Jiangsu Tianyi Airport Equipment Maintenance Service Co.,Ltd. Assignor: Jiangsu Tianyi Aviation Industry Co.,Ltd. Contract record no.: X2023980044219 Denomination of invention: A Remote Driving Control Method for Automatic Parking of Passenger Boarding Vehicles Granted publication date: 20230428 License type: Common License Record date: 20231024 |
|
EE01 | Entry into force of recordation of patent licensing contract |