Image-based urban road pavement adhesion coefficient acquisition method
Technical Field
The invention belongs to the technical field of intelligent automobiles, relates to a road surface adhesion coefficient acquisition method, and more particularly relates to an image-based urban road surface adhesion coefficient acquisition method.
Background
With the development of automobile intellectualization, the performance requirements of users on vehicle-mounted intelligent driving auxiliary systems and unmanned driving systems are higher and higher, the improvement of the performance of most intelligent driving auxiliary systems depends on the accuracy of dynamic control, the design of a high-performance dynamic control system needs to accurately acquire road surface information in real time, and an estimator based on a dynamic model can acquire a real-time and accurate road surface adhesion coefficient estimation value.
At the present stage, more and more intelligent vehicles are provided with cameras and other devices to acquire road information and surrounding vehicle information, so that new opportunities are brought to the research of a road adhesion coefficient identification method, and the intelligent vehicle has the advantages that the conditions of the front road surface can be sensed in advance, so that certain prediction capability is provided, the intelligent driving vehicle can adjust a control strategy in advance under the condition that the road surface changes suddenly, the capability of responding to dangerous working conditions is improved, and the challenge is to how to acquire the road adhesion coefficient information by using image sensing information.
Disclosure of Invention
In order to overcome the problems in the prior art, the invention provides an urban road pavement adhesion coefficient acquisition method based on an image.
The method is realized by the following technical scheme:
an urban road pavement adhesion coefficient obtaining method based on images comprises the following specific steps:
step one, establishing a road surface image information base
The precondition for obtaining the road adhesion coefficient based on the image is that a perfect road image information base can be established, and the sample image is properly processed to ensure that the characteristic information in the image is fully obtained;
firstly, acquiring pavement image data, wherein adverse factors on imaging effects need to be made up in the pavement image acquisition process, the image acquisition equipment is not limited to one or a certain type of image acquisition equipment, and the requirements on equipment performance and installation position are as follows: the video resolution of 1280 multiplied by 720 and above is provided, the video frame rate is 30 frames per second and above, the maximum effective shooting distance is more than 70 meters, and the wide dynamic technology is provided to quickly adapt to the light intensity change; the installation position of the equipment should ensure that the road surface shot in the acquired image information occupies more than half of the whole image area;
according to the conditions of urban road surfaces under different weather conditions, through comparative analysis and by combining the types of the urban road surfaces in China, the road surface types to be identified are specifically defined as 5 road surface types including an asphalt road surface, a cement road surface, a loose snow road surface, a compacted snow road surface and an ice plate road surface, a video file in the data acquisition process is decomposed into pictures at intervals of 10 frames, the pictures are sorted according to the 5 attribution types according to road surface characteristics in GB/T920 plus 2002 road surface grade and surface layer type code and pavement adhesion coefficient survey analysis in cold regions, the same type of road surface images are uniformly stored under the same folder, and the establishment of a road surface image information base is completed;
step two, establishing a pavement image data set
The method comprises the steps that an original collected image still contains a large number of non-road surface elements, and the acquisition precision of a road surface adhesion coefficient is seriously influenced, so that an image sample and a pixel-level label of a region corresponding to a road surface are needed in the road surface adhesion coefficient acquisition method based on the image, the image in a road surface image information base collected in the step one needs to be subjected to road surface range labeling, Labelme in software Anaconda is selected as a labeling tool, the labeling tool is used for manually labeling each image in a sample set one by one, a create polygon button is clicked in the labeling process, points are drawn along the boundary of the road surface region in the image, a labeling frame can completely cover the road surface region, and the labeling category is named as road; after the labeling is finished, a json file can be generated and is converted by using a self-contained json _ to _ dataset. py script program in software Anaconda to obtain a json folder, five files with names and suffixes of img.png, label.png, label _ viz.png, info.yaml and label _ names are contained under the folder, only the file with the label.png picture format is required to be converted to obtain a 8-bit gray label image, the labeling process is sequentially carried out on the pictures in the road image information base by using Labelmes in the Anaconda to obtain a gray label image set of the pictures in the road image information base, and the gray label image set of the pictures in the road image information base is a road image data set;
step three, establishing and training a road surface image area extraction network
The extraction process of the pavement image area is realized under the Anaconda environment through a semantic segmentation network, the whole semantic segmentation network is of an encoder-decoder structure, and the specific design is as follows:
3.1, firstly, zooming the image to be recognized into a picture with the size of 769 multiplied by 3, and taking the picture as the input of a semantic segmentation network;
3.2, then setting the first layer as a convolutional layer, adopting 32 filters with the size of 3 × 3, the step length is 2, filling the filter with 1, and obtaining the size of an output characteristic graph of the convolutional layer, which is 385 × 385 × 32, through batch regularization and a ReLU activation function;
3.3, inputting the convolution layer output characteristic diagram into a maximum value pooling layer, wherein the size is 3 multiplied by 3, the step length is 2, and the size of the obtained pooling layer output characteristic diagram is 193 multiplied by 32;
3.4, taking the output characteristic diagram of the pooling layer as the input of the bottleneck module structure, wherein the bottleneck module structure is realized in detail as follows: firstly, copying input characteristic channel to increase characteristic dimension, one branch directly passing through depth convolution with size of 3X 3 and step length of 2, another branch equally dividing into two sub-branches by channel splitting, one sub-branch is subjected to 3 x 3 depth convolution and 1 x 1 point-by-point convolution, the other sub-branch is directly subjected to a characteristic multiplexing mode, then the two sub-branches are connected through channel splicing, the channel arrangement sequence is disturbed through channel cleaning, after the same depth convolution with the size of 3 multiplied by 3 and the step length of 2, the data are spliced with a copy channel, finally the information exchange between groups is realized through the point-by-point convolution with the size of 1 multiplied by 1, therefore, the size of the output characteristic diagram of the whole bottleneck module structure is reduced by half, the number of channels is doubled, and the size of the output characteristic diagram passing through the bottleneck module structure once is 97 multiplied by 64;
3.5, taking the output result of the process 3.4 as input, passing through the bottleneck module structure again to obtain the output characteristic graph with the size of 49 × 49 × 128 after passing through the bottleneck module structure twice, taking the result as input, passing through the bottleneck module structure again to obtain the output characteristic graph with the size of 25 × 25 × 256 after passing through the bottleneck module structure three times, and reducing the size by 32 times compared with the original image, wherein the whole part is used as an encoder part of a semantic segmentation network;
3.6, the decoder part adopts a jump structure, 2 times of upsampling is carried out on the output characteristic diagram of the three-time bottleneck module structure which is processed by the process 3.5 by using a bilinear interpolation method to obtain a characteristic diagram with the size of 49 multiplied by 256, and the characteristic diagram and the output characteristic diagram of the two-time bottleneck module structure which is processed by the process 3.5 are added pixel by pixel, and in the process, the output characteristic channels of the two-time bottleneck module structure which is processed by the process 3.5 need to be copied to ensure that the result is still 256 output channels;
3.7, performing 2 times of upsampling on the result obtained in the process 3.6 by using a bilinear interpolation method again to obtain a characteristic diagram with the size of 97 multiplied by 256, and adding the characteristic diagram with the output characteristic diagram of the primary bottleneck module structure passing through the process 3.4 pixel by pixel;
3.8, converting the output channel number into a semantic category number by passing the result of the process 3.7 through a 1 × 1 convolutional layer, setting a Dropout layer to reduce the occurrence of an overfitting phenomenon, finally obtaining a feature map with the same size as the original image by 8 times of upsampling, giving a semantic category prediction result of each pixel point according to the maximum probability by using an Argmax function, and finally obtaining the whole semantic segmentation network;
randomly disordering the pavement image data set established in the step two, selecting 80% of sample pictures as a training set, and selecting 20% of sample pictures as a verification set; during semantic segmentation network training; the size of the read training set picture tensor is randomly scaled between 0.5 time and 2 times according to 0.25 step length, the picture tensor is randomly cut according to the size of 769 multiplied by 769 pixels and is randomly turned left and right, the purpose of data enhancement is achieved, the adaptability of a segmentation network is improved, and pixel point values are normalized from 0-255 to 0-1;
selecting a Poly learning rate rule when training a semantic segmentation network, wherein a learning rate attenuation expression is an expression (1), an initial learning rate is 0.001, training iteration steps are iters, the maximum training step max _ iter is set to be 20K steps, and power is set to be 0.9; using an Adam optimization solution algorithm, dynamically adjusting the learning rate of each parameter by using first moment estimation and second moment estimation of the gradient, setting the batch processing size to be 16 according to the performance of computer hardware, storing the model parameters once every 10-30min, and simultaneously using a verification set to perform performance evaluation on the network;
after the network training is finished, a proper semantic segmentation evaluation index is required to be selected for evaluating the performance of the model, before that, a confusion matrix is introduced, as shown in table 1, each row of the two-classification confusion matrix represents a prediction class, each column of the two-classification confusion matrix represents a real attribution class of data, and a specific numerical value represents the number of samples predicted to be a certain class;
TABLE 1 two-class confusion matrix schematic
The evaluation index of the semantic segmentation network is an average intersection-to-union ratio MIoU, which represents the ratio of the intersection and the union of each type of prediction result and the true value, and the result of the sum and the re-averaging is shown in formula (2):
when the MIoU is trained to be more than 60%, the training can be considered to be finished, the trained model and model parameters are stored, a road surface image area extraction network can be obtained, and the actually acquired original image is input into the road surface image area extraction network, so that the extraction of the road surface area in the image can be finished;
step four, establishing and training a road surface type recognition network
The extraction process of the road surface area in the real-time image information can be completed through the network in the third step, and the identification of the road surface type is completed on the basis of the extraction result of the road surface area in the third step;
after the image pavement data set is processed by the semantic segmentation network, an image set only containing a pavement area is obtained and is used as a final data set of a training and evaluation pavement type recognition network, so that the pavement type recognition network is built under the Anaconda environment, and the specific network structure is designed as follows:
4.1, firstly, scaling the image to be classified and identified into a picture with the size of 224 multiplied by 3 as the input of a convolutional neural network;
4.2, then setting the first layer as a convolutional layer, adopting 32 filters with the size of 3 × 3, the step length is 2, filling the filter with 1, and obtaining the size of an output characteristic diagram of the convolutional layer with the size of 112 × 112 × 32 through batch regularization and a ReLU activation function;
4.3, inputting the convolution layer output characteristic diagram into a maximum value pooling layer, wherein the size is 3 multiplied by 3, the step length is 2, and the size of the output characteristic diagram of the pooling layer is 56 multiplied by 32;
4.4, taking the output characteristic diagram of the pooling layer as the input of the bottleneck module structure, wherein the bottleneck module structure is implemented in detail as follows: firstly, copying input characteristic channel to increase characteristic dimension, one branch directly passing through depth convolution with size of 3X 3 and step length of 2, another branch equally dividing into two sub-branches by channel splitting, one sub-branch is subjected to 3 x 3 depth convolution and 1 x 1 point-by-point convolution, the other sub-branch is directly subjected to a characteristic multiplexing mode, then the two sub-branches are connected through channel splicing, the channel arrangement sequence is disturbed through channel cleaning, after the same depth convolution with the size of 3 x 3 and the step length of 2, the data are spliced with a copy channel, finally the information exchange between groups is realized through the point-by-point convolution with the size of 1 x 1, it can be seen that the size of the output characteristic diagram of the whole bottleneck module structure is reduced by half, the number of channels is doubled, and the size of the output characteristic diagram passing through the bottleneck module structure for one time is 28 multiplied by 64;
4.5, taking the output result of the process 4.4 as input, passing through the bottleneck module structure again to obtain an output characteristic diagram with the size of 14 multiplied by 128 after passing through the bottleneck module structure twice, taking the result as input, and passing through the bottleneck module structure again to obtain an output characteristic diagram with the size of 7 multiplied by 256 after passing through the bottleneck module structure three times;
4.6, converting the output result in the process 4.5 into a characteristic diagram with the size of 1 × 1 × 256 by using a global average pooling layer with the size of 7 × 7;
4.7, using a layer of full connection layer and a Softmax function as a network classifier, converting the output characteristic diagram in the process 4.6 into probability values belonging to various categories, and determining a network classification result according to the maximum probability value by using an Argmax function;
then, making the images only containing the road surface area obtained in the step three into a data set for training a road surface type identification network, storing different types of road surface images in a classified manner according to the folder names established in the step one, sequentially reading the image data in different folders, adding 5-bit 0/1 label information and road surface adhesion coefficient information, referring to table 2, adjusting the size of the image to 224 x 224 pixels in a bilinear interpolation manner, normalizing the pixel point values from 0-255 to 0-1, disturbing the road surface image data set, randomly extracting each type of image according to a proportion of 20% as a verification set, and taking the rest as a training set;
TABLE 2 pavement image Category labels
After the training set and the verification set of the road surface type recognition network are manufactured, training and evaluating a network model are started, the batch processing size is set to 64, a cross entropy loss function is selected, an Adam optimization solving algorithm is used, the basic learning rate is 0.0001, when the training is completed until the MIoU is more than 80%, the training can be considered to be completed, the model and the training result are stored according to the iteration times epoch, and the well-trained road surface type recognition network can be obtained;
step five, obtaining the road surface adhesion coefficient information
The road surface adhesion coefficient information acquisition process is as follows: the method comprises the steps of shooting a front road image through a camera in the driving process of a vehicle, transmitting the front road image shot by the camera to a road image area extraction network to obtain a road area, transmitting the image only containing the road area to a road type identification network for classification and identification, judging the adhesion coefficient range of the road where the road is located according to the corresponding vehicle speed after the road identification is finished, referring to a table 2, and taking the intermediate values of the upper limit and the lower limit of the road adhesion coefficient range as the current road adhesion coefficient to finish the acquisition of the road adhesion coefficient information.
Compared with the prior art, the invention has the beneficial effects that:
the invention discloses a road adhesion coefficient acquisition method based on images, which can provide road adhesion coefficient information for the development of an intelligent driving auxiliary system and an unmanned driving system; the method realizes the acquisition of the adhesion coefficient by acquiring the front road image, and can realize the advance acquisition of the road adhesion information; the method designs a method for extracting the network and identifying the network serial of the road surface type based on the road surface image area, so that the real-time and rapid acquisition of the front road surface attachment information can be realized.
Drawings
FIG. 1 is a simplified flow chart of a method for obtaining an image-based road surface adhesion coefficient of an urban road in the method;
FIG. 2 is a network structure diagram of the road surface image area extraction in the method;
FIG. 3 is a diagram of the bottleneck module in the present method;
FIG. 4 is a diagram of a road surface type recognition network in the present method;
Detailed Description
The invention provides an urban road adhesion coefficient acquisition method based on images, which aims to solve the problem of acquisition of road adhesion coefficient information required by development of intelligent vehicle driving assistance and unmanned technology.
The invention relates to an image-based urban road pavement adhesion coefficient acquisition method, which comprises the following specific steps:
step one, establishing a road surface image information base
The precondition for obtaining the road adhesion coefficient based on the image is that a perfect road image information base can be established, and the sample image is properly processed to ensure that the characteristic information in the image is fully obtained;
firstly, acquiring pavement image data, wherein adverse factors on imaging effects need to be made up in the pavement image acquisition process, the image acquisition equipment is not limited to one or a certain type of image acquisition equipment, and the requirements on equipment performance and installation position are as follows: the video resolution of 1280 multiplied by 720 and above is provided, the video frame rate is 30 frames per second and above, the maximum effective shooting distance is more than 70 meters, and the wide dynamic technology is provided to quickly adapt to the light intensity change; the installation position of the equipment should ensure that the road surface shot in the acquired image information occupies more than half of the whole image area;
according to the conditions of urban road surfaces under different weather conditions, through comparative analysis and by combining the types of the urban road surfaces in China, the road surface types to be identified are specifically defined as 5 road surface types including an asphalt road surface, a cement road surface, a loose snow road surface, a compacted snow road surface and an ice plate road surface, a video file in the data acquisition process is decomposed into pictures at intervals of 10 frames, the pictures are sorted according to the 5 attribution types according to road surface characteristics in GB/T920 plus 2002 road surface grade and surface layer type code and pavement adhesion coefficient survey analysis in cold regions, the same type of road surface images are uniformly stored under the same folder, and the establishment of a road surface image information base is completed;
step two, establishing a pavement image data set
The method comprises the steps that an original collected image still contains a large number of non-road surface elements, and the acquisition precision of a road surface adhesion coefficient is seriously influenced, so that an image sample and a pixel-level label of a region corresponding to a road surface are needed in the road surface adhesion coefficient acquisition method based on the image, the image in a road surface image information base collected in the step one needs to be subjected to road surface range labeling, Labelme in software Anaconda is selected as a labeling tool, the labeling tool is used for manually labeling each image in a sample set one by one, a create polygon button is clicked in the labeling process, points are drawn along the boundary of the road surface region in the image, a labeling frame can completely cover the road surface region, and the labeling category is named as road; after the labeling is finished, a json file can be generated and is converted by using a self-contained json _ to _ dataset. py script program in software Anaconda to obtain a json folder, five files with names and suffixes of img.png, label.png, label _ viz.png, info.yaml and label _ names are contained under the folder, only the file with the label.png picture format is required to be converted to obtain a 8-bit gray label image, the labeling process is sequentially carried out on the pictures in the road image information base by using Labelmes in the Anaconda to obtain a gray label image set of the pictures in the road image information base, and the gray label image set of the pictures in the road image information base is a road image data set;
step three, training of road surface image area extraction network
The extraction network of the pavement image area is realized in an Anaconda environment through a semantic segmentation network, the structure of the semantic segmentation network is shown in FIG. 2, the whole semantic segmentation network is an encoder-decoder structure, and the specific design is as follows:
3.1, firstly, zooming the image to be recognized into a picture with the size of 769 multiplied by 3, and taking the picture as the input of a semantic segmentation network;
3.2, then setting the first layer as a convolutional layer, adopting 32 filters with the size of 3 × 3, the step length is 2, filling the filter with 1, and obtaining the size of an output characteristic graph of the convolutional layer, which is 385 × 385 × 32, through batch regularization and a ReLU activation function;
3.3, inputting the convolution layer output characteristic diagram into a maximum value pooling layer, wherein the size is 3 multiplied by 3, the step length is 2, and the size of the obtained pooling layer output characteristic diagram is 193 multiplied by 32;
3.4, taking the output characteristic diagram of the pooling layer as the input of the bottleneck module structure, wherein the bottleneck module structure is shown in fig. 3, and the detailed implementation process of the bottleneck module structure is as follows: firstly, copying input characteristic channel to increase characteristic dimension, one branch directly passing through depth convolution with size of 3X 3 and step length of 2, another branch equally dividing into two sub-branches by channel splitting, one sub-branch is subjected to 3 x 3 depth convolution and 1 x 1 point-by-point convolution, the other sub-branch is directly subjected to a characteristic multiplexing mode, then the two sub-branches are connected through channel splicing, the channel arrangement sequence is disturbed through channel cleaning, after the same depth convolution with the size of 3 multiplied by 3 and the step length of 2, the data are spliced with a copy channel, finally the information exchange between groups is realized through the point-by-point convolution with the size of 1 multiplied by 1, therefore, the size of the output characteristic diagram of the whole bottleneck module structure is reduced by half, the number of channels is doubled, and the size of the output characteristic diagram passing through the bottleneck module structure once is 97 multiplied by 64;
3.5, taking the output result of the process 3.4 as input, passing through the bottleneck module structure again to obtain the output characteristic graph with the size of 49 × 49 × 128 after passing through the bottleneck module structure twice, taking the result as input, passing through the bottleneck module structure again to obtain the output characteristic graph with the size of 25 × 25 × 256 after passing through the bottleneck module structure three times, and reducing the size by 32 times compared with the original image, wherein the whole part is used as an encoder part of a semantic segmentation network;
3.6, the decoder part adopts a jump structure, 2 times of upsampling is carried out on the output characteristic diagram of the three-time bottleneck module structure which is processed by the process 3.5 by using a bilinear interpolation method to obtain a characteristic diagram with the size of 49 multiplied by 256, and the characteristic diagram and the output characteristic diagram of the two-time bottleneck module structure which is processed by the process 3.5 are added pixel by pixel, and in the process, the output characteristic channels of the two-time bottleneck module structure which is processed by the process 3.5 need to be copied to ensure that the result is still 256 output channels;
3.7, performing 2 times of upsampling on the result obtained in the process 3.6 by using a bilinear interpolation method again to obtain a characteristic diagram with the size of 97 multiplied by 256, and adding the characteristic diagram with the output characteristic diagram of the primary bottleneck module structure passing through the process 3.4 pixel by pixel;
3.8, converting the output channel number into a semantic category number by passing the result of the process 3.7 through a 1 × 1 convolutional layer, setting a Dropout layer to reduce the occurrence of an overfitting phenomenon, finally obtaining a feature map with the same size as the original image by 8 times of upsampling, giving a semantic category prediction result of each pixel point according to the maximum probability by using an Argmax function, and finally obtaining the whole semantic segmentation network;
randomly disordering the pavement image data set established in the step two, selecting 80% of sample pictures as a training set, and selecting 20% of sample pictures as a verification set; during semantic segmentation network training; the size of the read training set picture tensor is randomly scaled between 0.5 time and 2 times according to 0.25 step length, the picture tensor is randomly cut according to the size of 769 multiplied by 769 pixels and is randomly turned left and right, the purpose of data enhancement is achieved, the adaptability of a segmentation network is improved, and pixel point values are normalized from 0-255 to 0-1;
selecting a Poly learning rate rule when training a semantic segmentation network, wherein a learning rate attenuation expression is an expression (1), an initial learning rate is 0.001, training iteration steps are iters, the maximum training step max _ iter is set to be 20K steps, and power is set to be 0.9; using an Adam optimization solution algorithm, dynamically adjusting the learning rate of each parameter by using first moment estimation and second moment estimation of the gradient, setting the batch processing size to be 16 according to the performance of computer hardware, storing the model parameters once every 10-30min, and simultaneously using a verification set to perform performance evaluation on the network;
after the network training is finished, a proper semantic segmentation evaluation index is required to be selected for evaluating the performance of the model, before that, a confusion matrix is introduced, as shown in table 1, each row of the two-classification confusion matrix represents a prediction class, each column of the two-classification confusion matrix represents a real attribution class of data, and a specific numerical value represents the number of samples predicted to be a certain class;
TABLE 1 two-class confusion matrix schematic
The evaluation index of the semantic segmentation network is an average intersection-to-union ratio MIoU, which represents the ratio of the intersection and the union of each type of prediction result and the true value, and the result of the sum and the re-averaging is shown in formula (2):
when the MIoU is trained to be more than 60%, the training can be considered to be finished, the trained model and model parameters are stored, a road surface image area extraction network can be obtained, and the actually acquired original image is input into the road surface image area extraction network, so that the extraction of the road surface area in the image can be finished;
step four, training the road surface recognition network
The extraction process of the road surface area in the real-time image information can be completed through the network in the third step, and the identification of the road surface type is completed on the basis of the extraction result of the road surface area in the third step;
after the image pavement data set is processed by the semantic segmentation network, an image set only containing a pavement area is obtained, and the image pavement data set is used as a final data set of a training and evaluation pavement type recognition network, so that the pavement type recognition network is built under an Anaconda environment, as shown in FIG. 4, the specific network structure is designed as follows:
4.1, firstly, scaling the image to be classified and identified into a picture with the size of 224 multiplied by 3 as the input of a convolutional neural network;
4.2, then setting the first layer as a convolutional layer, adopting 32 filters with the size of 3 × 3, the step length is 2, filling the filter with 1, and obtaining the size of an output characteristic diagram of the convolutional layer with the size of 112 × 112 × 32 through batch regularization and a ReLU activation function;
4.3, inputting the convolution layer output characteristic diagram into a maximum value pooling layer, wherein the size is 3 multiplied by 3, the step length is 2, and the size of the output characteristic diagram of the pooling layer is 56 multiplied by 32;
4.4, taking the output characteristic diagram of the pooling layer as the input of the bottleneck module structure, wherein the bottleneck module structure is implemented in detail as follows: firstly, copying input characteristic channel to increase characteristic dimension, one branch directly passing through depth convolution with size of 3X 3 and step length of 2, another branch equally dividing into two sub-branches by channel splitting, one sub-branch is subjected to 3 x 3 depth convolution and 1 x 1 point-by-point convolution, the other sub-branch is directly subjected to a characteristic multiplexing mode, then the two sub-branches are connected through channel splicing, the channel arrangement sequence is disturbed through channel cleaning, after the same depth convolution with the size of 3 x 3 and the step length of 2, the data are spliced with a copy channel, finally the information exchange between groups is realized through the point-by-point convolution with the size of 1 x 1, it can be seen that the size of the output characteristic diagram of the whole bottleneck module structure is reduced by half, the number of channels is doubled, and the size of the output characteristic diagram passing through the bottleneck module structure for one time is 28 multiplied by 64;
4.5, taking the output result of the process 4.4 as input, passing through the bottleneck module structure again to obtain an output characteristic diagram with the size of 14 multiplied by 128 after passing through the bottleneck module structure twice, taking the result as input, and passing through the bottleneck module structure again to obtain an output characteristic diagram with the size of 7 multiplied by 256 after passing through the bottleneck module structure three times;
4.6, converting the output result in the process 4.5 into a characteristic diagram with the size of 1 × 1 × 256 by using a global average pooling layer with the size of 7 × 7;
4.7, using a layer of full connection layer and a Softmax function as a network classifier, converting the output characteristic diagram in the process 4.6 into probability values belonging to various categories, and determining a network classification result according to the maximum probability value by using an Argmax function;
then, making the images only containing the road surface area obtained in the step three into a data set for training a road surface type identification network, storing different types of road surface images in a classified manner according to the folder names established in the step one, sequentially reading the image data in different folders, adding 5-bit 0/1 label information and road surface adhesion coefficient information, referring to table 2, adjusting the size of the image to 224 x 224 pixels in a bilinear interpolation manner, normalizing the pixel point values from 0-255 to 0-1, disturbing the road surface image data set, randomly extracting each type of image according to a proportion of 20% as a verification set, and taking the rest as a training set;
TABLE 2 pavement image Category labels
After the training set and the verification set of the road surface type recognition network are manufactured, training and evaluating a network model are started, the batch processing size is set to 64, a cross entropy loss function is selected, an Adam optimization solving algorithm is used, the basic learning rate is 0.0001, when the training is completed until the MIoU is more than 80%, the training can be considered to be completed, the model and the training result are stored according to the iteration times epoch, and the well-trained road surface type recognition network can be obtained;
step five, obtaining the road surface adhesion coefficient information
The road surface adhesion coefficient information acquisition process is as follows: the method comprises the steps of shooting a front road image through a camera in the driving process of a vehicle, transmitting the front road image shot by the camera to a road image area extraction network to obtain a road area, transmitting the image only containing the road area to a road type identification network for classification and identification, judging the adhesion coefficient range of the road where the road is located according to the corresponding vehicle speed after the road identification is finished, referring to a table 2, and taking the intermediate values of the upper limit and the lower limit of the road adhesion coefficient range as the current road adhesion coefficient to finish the acquisition of the road adhesion coefficient information.