CN112785610B - Lane line semantic segmentation method integrating low-level features - Google Patents
Lane line semantic segmentation method integrating low-level features Download PDFInfo
- Publication number
- CN112785610B CN112785610B CN202110049820.XA CN202110049820A CN112785610B CN 112785610 B CN112785610 B CN 112785610B CN 202110049820 A CN202110049820 A CN 202110049820A CN 112785610 B CN112785610 B CN 112785610B
- Authority
- CN
- China
- Prior art keywords
- lane line
- image
- low
- semantic segmentation
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/12—Edge-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/588—Recognition of the road, e.g. of lane markings; Recognition of the vehicle driving pattern in relation to the road
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20092—Interactive image processing based on input by user
- G06T2207/20104—Interactive definition of region of interest [ROI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Abstract
The invention provides a lane line semantic segmentation method fusing low-level features, which comprises the following steps: collecting video in the running process of the unmanned vehicle, and screening images containing lane lines from the video to form a lane line data set; preprocessing the data set to obtain a training set required by training; inputting lane line images in the training set into a network model for training, respectively extracting second-layer image features, third-layer image features and last-layer image features by the network model, and carrying out multiple low-layer feature fusion operations; and carrying out up-sampling operation of a preset multiple on the semantic information image fused with the low-level features to obtain a lane line semantic segmentation image. The method can realize semantic segmentation of the lane line image, identify the lane line in the driving scene, has the advantages of higher accuracy, stronger anti-interference capability, good robustness and the like, and can be applied to unmanned vehicles.
Description
Technical Field
The invention relates to the field of computer vision, in particular to a lane line image semantic segmentation method, and more particularly relates to a lane line semantic segmentation method with low-level features fused.
Background
The lane line semantic segmentation technology is one application of computer vision in the field of unmanned vehicles, and the external environment around the vehicle is perceived through a vehicle-mounted camera, so that the unmanned vehicles are assisted to know the surrounding world. The main core of the lane line semantic segmentation technology is that classification of the category of each pixel point on the image is realized for lane line image information acquired in the running process of the vehicle.
The most main problem of the current lane line detection technology is that accuracy and real-time cannot be simultaneously considered. Traditional lane line detection methods rely on predefined lane line models to detect lane lines, which are susceptible to environmental factor changes. In some complex environments, the traditional road detection method is not high in robustness and is easily influenced by road environments such as illumination intensity, road width change, lane line blurring, lane line edge degradation and the like, and although some methods can accurately detect lane lines, the execution speed is too slow to meet the use requirement; some approaches have improved real-time performance, but at the cost of accuracy. In the automatic driving process, accurate identification of lane line information from images plays a key role in automatic driving. Therefore, a high-efficiency and real-time semantic segmentation algorithm is designed, and the rapid development of the unmanned automobile can be promoted.
Disclosure of Invention
In order to solve the problems of the traditional lane line detection method in the background technology, the invention provides the lane line semantic segmentation method fused with low-level features, which can efficiently detect lane line information in an image in real time, has good algorithm stability and can be applied to an unmanned system.
The invention provides a lane line semantic segmentation method fusing low-level features, which comprises the following steps:
1) Collecting video data in the running process of the unmanned vehicle, and screening out images containing lane lines from the video data to form a data set;
2) Image preprocessing is carried out on images in the data set, and a training set required by training is obtained;
3) Inputting the images in the training set into a network model for training, extracting the second layer image features, the third layer image features and the last layer image features in the network model, and carrying out multiple low layer feature fusion operations to obtain semantic information images fused with the low layer features;
4) And carrying out up-sampling operation of a preset multiple on the semantic information image fused with the low-level features to obtain a final lane line semantic segmentation image.
Further, in the step 1), the lane line data set collection process specifically includes the following steps:
1.1 Horizontally mounting the vehicle-mounted camera on a vehicle, driving in various urban road scenes, and shooting video at a preset acquisition frame rate in the driving process to obtain video data comprising lane lines;
1.2 Extracting the last frame image of each second to form a road scene data set;
1.3 Deleting images which do not contain lane lines in the road scene data set, wherein the rest images form the lane line data set, and the images in the data set contain a plurality of urban road scenes.
Further, the urban road scene includes high speed, tunnel, urban area, suburb and campus.
Further, in the step 2), the preprocessing of the image in the dataset comprises the following specific steps:
2.1 Cutting the images in the data set respectively, and extracting the interested areas of the images in the data set;
2.2 Data enhancement processing is carried out on the data obtained in the step 2.1), more images containing lane lines are obtained, and a lane line training set required by training is formed.
Further, the data enhancement processing in step 2.2) means that for a photo in the training set, one of brightness, contrast, scaling, rotation is randomly selected, and the attribute value of the image is randomly changed.
Further, in the step 3), the multiple low-level feature fusion operations are performed, and the specific process is as follows:
3.2 Performing up-sampling operation of a preset multiple on the image features of the last layer, performing convolution operation of 1 multiplied by 1 on the image features of the third layer, and fusing the two layers of image features by using fusion operation to obtain a semantic information image of the first fused low-layer features;
3.3 The semantic information image of the low-level features is fused for the first time, the up-sampling operation of a preset multiple is carried out, meanwhile, the convolution operation of 1 multiplied by 1 is carried out on the features of the second-level images, and fusion operation is used for fusing the two features to obtain the semantic information image of the low-level features fused for the second time, namely the semantic information image of the low-level features fused.
Further, the network model in the step 3) is a deep v3+ network model.
Further, in the step 3), the step of inputting the images in the training set into the network model for training specifically includes: the network model adopts a learning strategy of poly to determine the initial learning rateAnd ending the learning rate, wherein the learning rate needs to be multiplied in the training process of the network modelitem is the number of iterations.
Further, in the step 3), the network model needs to calculate the cross loss function during each training, and then adopts a random gradient descent method to optimize the training;
wherein, the cross-over loss function is:
wherein L represents a cross-loss function value; y is a real sample label;to predict the output value.
Further, after the final lane line semantic segmentation image is obtained in the step 4), the pixel precision and the average intersection ratio are adopted as evaluation indexes of the lane line semantic segmentation result,
wherein, the pixel Precision (PA) is:
average cross-over ratio (MIoU) is:
wherein k+1 represents a category of semantic segmentation, and comprises k target classes and 1 background class; p is p ii Representing the number of pixels originally belonging to class i and also predicted as class i, p ij Representing the number of pixels originally belonging to class i but predicted to class j, p ji Representing the number of pixels originally belonging to the j-th class but predicted as the i-th class.
Compared with the prior art, the invention has the following beneficial effects:
(1) Compared with the traditional lane line detection method, the method can be used for fusing the low-layer characteristic information of the image for multiple times to obtain more lane line boundary information, fusing the low-layer lane line boundary information with the high-layer lane line semantic information, improving the dividing precision of the lane lines, and has the advantages of good instantaneity, high accuracy, good robustness, capability of detecting the lane lines in various scenes and good algorithm universality.
(2) The method disclosed by the invention fuses the boundary information of the lane lines in the image for multiple times, so that the lane line detection process is not influenced by light, the lane lines on the image can be accurately identified and segmented under the condition that the lane lines are fuzzy, and the method has stronger anti-interference capability.
(3) According to the characteristics of various lane lines and easy interference from external environment, the deep labV & lt3+ & gt is selected as the network model, the network model has strong characteristic extraction capability, can adapt to multi-scale shape characteristics, and simultaneously combines an image enhancement method to improve the generalization capability of the network model.
Drawings
Fig. 1 is a schematic flow chart of an embodiment of the present invention.
Fig. 2 is a network configuration diagram of an embodiment of the present invention.
Fig. 3 is a road scene image.
Fig. 4 is a semantic segmentation result image according to an embodiment of the present invention.
Detailed Description
The invention will be further described with reference to the accompanying drawings and specific examples.
Referring to fig. 1-4, the lane line semantic segmentation method with fusion of low-level features provided in this embodiment includes the following steps:
step S1: video data in the running process of the unmanned vehicle are collected, and images containing lane lines are screened out to form a data set.
The method specifically comprises the steps of extracting the last frame of image of each second, forming a road scene data set, deleting all pictures which do not contain lane line information, and forming a plurality of urban road scene data sets which contain high speed, tunnels, urban areas, suburbs, campuses and the like.
The lane line data set acquisition process comprises the following specific processes:
1.1 Horizontally mounting the vehicle-mounted camera on a vehicle, driving in various urban road scenes, and shooting video in the driving process to obtain video data containing lane lines;
1.2 The acquisition frame rate of the video is 30 frames/second, and under the actual road condition, the difference of 30 frames of road scenes per second is not great, so that the last frame of image per second is extracted to form a road scene data set;
1.3 The image which does not contain the lane lines is deleted in the road scene data set, and the rest of the image forms the lane line data set and contains a plurality of urban road scenes such as high speed, tunnels, urban areas, suburbs, campuses and the like.
Step S2: and carrying out image preprocessing on the images in the data set to obtain a training set required by training.
Preprocessing of images in a data set, wherein the specific process is as follows:
2.1 The resolution of the image in the lane line data set is 1920×1080, wherein the lane line of interest is occupied in the lower half of the image, the part of no interest is occupied in the upper half of the image, the image is required to be cut, and the region of interest with the size of 1920×720 is extracted;
2.2 And 2.1) carrying out data enhancement processing on the obtained data in the step 2.1), and adjusting the obtained data from multiple aspects of brightness, contrast, scaling, rotation and the like to obtain more images containing lane lines, thereby forming a lane line training set required by training. In specific implementation, the data enhancement operation specifically includes: for a photo in the training set, randomly selecting one of brightness, contrast, scaling and rotation, and randomly changing the attribute value of the image, wherein the weight value of the brightness is randomly valued between-1 and 1, the weight of the contrast is randomly valued between 0.1 and 0.3, the scaled weight is valued between 0 and 1, and the rotated weight is valued between-1 and 1.
Step S3: and (2) taking the training set obtained in the step (S2) as the input of a network model, extracting the second layer image features, the third layer image features and the last layer image features in the network model, and carrying out multiple low layer feature fusion operations to obtain the semantic information image fused with the low layer features.
In this embodiment, the network model adopts a deep v3+ network model, please refer to fig. 1, and the specific process is as follows:
3.1 Extracting a second layer image feature, a third layer image feature and a last layer image feature from the deep labV < 3+ > network model respectively;
3.2 Carrying out up-sampling operation of a preset multiple (2 times in the embodiment) on the image features of the last layer, gradually recovering the resolution of the image, simultaneously, in order to reduce the parameter quantity of the model and accelerate the lane line detection speed, carrying out convolution operation of 1 multiplied by 1 on the image features of the third layer, and fusing the two layers of image features by using fusion operation to obtain a semantic information image of the first fused low-layer feature;
3.3 The semantic information image of the first fusion low-level feature is subjected to up-sampling operation of a preset multiple (2 times in the embodiment), meanwhile, the second-level image feature is subjected to convolution operation of 1×1, and the semantic information image of the second fusion low-level feature is obtained by fusion of the two images through fusion operation.
In this embodiment, the training method of the network model specifically includes: the network model employs a learning strategy of "poly". In the training process, in order to avoid that the model cannot be converged due to too large learning rate or the convergence speed is too slow due to too small learning rate, the initial learning rate is set to be 0.007, and the end learning rate is set to be 1e -6 Then, in the training process of the model, the learning rate needs to be multiplied byWherein the power in the formula is 0.9, the max_iter is 30000, the iter is the iteration number, each epoch is iterated for 30000, and the model trains 43 epochs in total.
The network model needs to calculate the cross loss function during each training, and then adopts a random gradient descent method to optimize the training, and the cross loss function is adopted in the embodiment as follows:
wherein L represents a cross-loss function value; y is a real sample label;to predict the output value. When predicting output valueThe closer to true tag y, the smaller the value of the loss function L.
Step S4: and (3) carrying out up-sampling operation of a preset multiple (4 times in the embodiment) on the semantic information image fused with the low-level features, and recovering the resolution of the image to the original image size to obtain a final lane line semantic segmentation result. And then using the pixel precision and the average intersection ratio as evaluation indexes of the lane line semantic segmentation results to evaluate the precision of the semantic segmentation results.
Pixel Precision (PA):
average cross-over ratio (MIoU):
wherein k+1 represents a category of semantic segmentation, including k target classes and 1 background class; p is p ii Representing the number of pixels originally belonging to class i and also predicted as class i, p ij Representing the number of pixels originally belonging to the ith class predicted as the jth class, p ji Representing the number of pixels originally belonging to the j-th class but predicted as the i-th class.
In this embodiment, there are 5812 experimental pictures, as shown in fig. 3, two of which are 4650 pictures for training, various types of road scenes are uniformly distributed in the data set, and the rest 1162 lane line images are taken as verification set pictures, and the sizes of the images are 1920×1080.
During training, firstly, the image input size of the network model is set to 1920×720, the learning rate adopts the poly learning strategy in the step 3, and the final network model parameters are obtained after training 43 epochs.
And testing the trained network model by using the lane line verification set to obtain a prediction result. The result of the final network model on the lane line training set is shown in fig. 4, and the lane line can be clearly detected. Therefore, the method can realize semantic segmentation of the lane line image, identify the lane line in the driving scene, has the advantages of higher accuracy, stronger anti-interference capability, good robustness and the like, and can be applied to unmanned vehicles.
The foregoing describes specific embodiments of the present invention. It is to be understood that the invention is not limited to the particular embodiments described above, and that various changes or modifications may be made by those skilled in the art within the scope of the appended claims without affecting the spirit of the invention. The embodiments of the present application and features in the embodiments may be combined with each other arbitrarily without conflict.
Claims (9)
1. The lane line semantic segmentation method integrating the low-level features is characterized by comprising the following steps of:
1) Collecting video data in the running process of the unmanned vehicle, and screening out images containing lane lines from the video data to form a data set;
2) Image preprocessing is carried out on images in the data set, and a training set required by training is obtained;
3) Inputting the images in the training set into a network model for training, extracting the second layer image features, the third layer image features and the last layer image features in the network model, and carrying out multiple low layer feature fusion operations to obtain semantic information images fused with the low layer features;
4) Carrying out up-sampling operation of preset multiples on the semantic information image fused with the low-level features to obtain a final lane line semantic segmentation image;
in the step 3), the multiple low-level feature fusion operations are performed, and the specific process is as follows:
carrying out up-sampling operation of a preset multiple on the image features of the last layer, carrying out convolution operation of 1 multiplied by 1 on the image features of the third layer, and fusing the image features of the two layers by using fusion operation to obtain a semantic information image of the first fused low-layer features;
and carrying out up-sampling operation of a preset multiple on the semantic information image of the first fusion low-level feature, carrying out convolution operation of 1 multiplied by 1 on the second-level image feature, and fusing the semantic information image and the second-level image feature by using fusion operation to obtain the semantic information image of the second fusion low-level feature, namely the semantic information image of the fusion low-level feature.
2. The lane line semantic segmentation method fused with low-level features according to claim 1, wherein the method comprises the following steps: in the step 1), the lane line data set acquisition process specifically includes the following steps:
1.1 Horizontally mounting the vehicle-mounted camera on a vehicle, driving in various urban road scenes, and shooting video at a preset acquisition frame rate in the driving process to obtain video data comprising lane lines;
1.2 Extracting the last frame image of each second to form a road scene data set;
1.3 Deleting images which do not contain lane lines in the road scene data set, wherein the rest images form the lane line data set, and the images in the data set contain a plurality of urban road scenes.
3. The lane line semantic segmentation method fused with low-level features according to claim 2, wherein the method comprises the following steps: the urban road scene includes high speed, tunnels, urban areas, suburbs and campuses.
4. The lane line semantic segmentation method fused with low-level features according to claim 1, wherein the method comprises the following steps: in the step 2), preprocessing of the image in the data set comprises the following specific processes:
2.1 Cutting the images in the data set respectively, and extracting the interested areas of the images in the data set;
2.2 Data enhancement processing is carried out on the data obtained in the step 2.1), more images containing lane lines are obtained, and a lane line training set required by training is formed.
5. The lane line semantic segmentation method fused with low-level features according to claim 4, wherein the method comprises the following steps: the data enhancement processing in step 2.2) refers to randomly selecting one of a plurality of attributes for a photo in a training set, and randomly changing the value of the selected attribute of the image, wherein the attributes include brightness, contrast, scaling and rotation.
6. The lane line semantic segmentation method fused with low-level features according to claim 1, wherein the method comprises the following steps: the network model in the step 3) is a deep labv3+ network model.
7. The lane line semantic segmentation method fused with low-level features according to claim 1, wherein the method comprises the following steps: the step 3) of inputting the images in the training set into the network model for training specifically includes: the network model adopts a learning strategy of poly to determine an initial learning rate and an end learning rate, and the learning rate needs to be multiplied in the training process of the network modelitem is the number of iterations.
8. The lane line semantic segmentation method fused with low-level features according to claim 1, wherein the method comprises the following steps: in the step 3), the network model needs to calculate a cross loss function during each training, and then a random gradient descent method is adopted to optimize the training;
wherein, the cross-over loss function is:
9. The lane line semantic segmentation method fused with low-level features according to any one of claims 1 to 8, wherein: after the final lane line semantic segmentation image is obtained in the step 4), the pixel precision and the average intersection ratio are adopted as evaluation indexes of lane line semantic segmentation results,
wherein, the pixel Precision (PA) is:
average cross-over ratio (MIoU) is:
wherein k+1 represents a category of semantic segmentation, and comprises k target classes and 1 background class; p is p ii Representing the number of pixels originally belonging to class i and also predicted as class i, p ij Representing the number of pixels originally belonging to class i but predicted to class j, p ji Representing the number of pixels originally belonging to the j-th class but predicted as the i-th class.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110049820.XA CN112785610B (en) | 2021-01-14 | 2021-01-14 | Lane line semantic segmentation method integrating low-level features |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110049820.XA CN112785610B (en) | 2021-01-14 | 2021-01-14 | Lane line semantic segmentation method integrating low-level features |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112785610A CN112785610A (en) | 2021-05-11 |
CN112785610B true CN112785610B (en) | 2023-05-23 |
Family
ID=75756750
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110049820.XA Active CN112785610B (en) | 2021-01-14 | 2021-01-14 | Lane line semantic segmentation method integrating low-level features |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112785610B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113609980A (en) * | 2021-08-04 | 2021-11-05 | 东风悦享科技有限公司 | Lane line sensing method and device for automatic driving vehicle |
CN116229379B (en) * | 2023-05-06 | 2024-02-02 | 浙江大华技术股份有限公司 | Road attribute identification method and device, electronic equipment and storage medium |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109902748A (en) * | 2019-03-04 | 2019-06-18 | 中国计量大学 | A kind of image, semantic dividing method based on the full convolutional neural networks of fusion of multi-layer information |
WO2020172875A1 (en) * | 2019-02-28 | 2020-09-03 | 深圳市大疆创新科技有限公司 | Method for extracting road structure information, unmanned aerial vehicle, and automatic driving system |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111860349A (en) * | 2020-07-23 | 2020-10-30 | 上海交通大学 | Intelligent vehicle lane line semantic segmentation method and system |
CN112215073A (en) * | 2020-09-10 | 2021-01-12 | 华蓝设计(集团)有限公司 | Traffic marking line rapid identification and tracking method under high-speed motion scene |
-
2021
- 2021-01-14 CN CN202110049820.XA patent/CN112785610B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020172875A1 (en) * | 2019-02-28 | 2020-09-03 | 深圳市大疆创新科技有限公司 | Method for extracting road structure information, unmanned aerial vehicle, and automatic driving system |
CN109902748A (en) * | 2019-03-04 | 2019-06-18 | 中国计量大学 | A kind of image, semantic dividing method based on the full convolutional neural networks of fusion of multi-layer information |
Also Published As
Publication number | Publication date |
---|---|
CN112785610A (en) | 2021-05-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109190752B (en) | Image semantic segmentation method based on global features and local features of deep learning | |
CN111368687A (en) | Sidewalk vehicle illegal parking detection method based on target detection and semantic segmentation | |
CN110929577A (en) | Improved target identification method based on YOLOv3 lightweight framework | |
CN111767927A (en) | Lightweight license plate recognition method and system based on full convolution network | |
CN111507226B (en) | Road image recognition model modeling method, image recognition method and electronic equipment | |
CN112785610B (en) | Lane line semantic segmentation method integrating low-level features | |
CN114693924A (en) | Road scene semantic segmentation method based on multi-model fusion | |
CN115205264A (en) | High-resolution remote sensing ship detection method based on improved YOLOv4 | |
CN112990065A (en) | Optimized YOLOv5 model-based vehicle classification detection method | |
CN116484971A (en) | Automatic driving perception self-learning method and device for vehicle and electronic equipment | |
CN113205107A (en) | Vehicle type recognition method based on improved high-efficiency network | |
CN113033454A (en) | Method for detecting building change in urban video camera | |
CN114049572A (en) | Detection method for identifying small target | |
CN115861619A (en) | Airborne LiDAR (light detection and ranging) urban point cloud semantic segmentation method and system of recursive residual double-attention kernel point convolution network | |
CN114913498A (en) | Parallel multi-scale feature aggregation lane line detection method based on key point estimation | |
CN110659601A (en) | Depth full convolution network remote sensing image dense vehicle detection method based on central point | |
CN115984543A (en) | Target detection algorithm based on infrared and visible light images | |
CN114463205A (en) | Vehicle target segmentation method based on double-branch Unet noise suppression | |
CN114419583A (en) | Yolov4-tiny target detection algorithm with large-scale features | |
CN111612803B (en) | Vehicle image semantic segmentation method based on image definition | |
CN111160282B (en) | Traffic light detection method based on binary Yolov3 network | |
CN111832463A (en) | Deep learning-based traffic sign detection method | |
CN116597270A (en) | Road damage target detection method based on attention mechanism integrated learning network | |
CN116580232A (en) | Automatic image labeling method and system and electronic equipment | |
CN116129327A (en) | Infrared vehicle detection method based on improved YOLOv7 algorithm |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |