CN116758506A - Three-dimensional vehicle detection method based on point cloud and image fusion - Google Patents
Three-dimensional vehicle detection method based on point cloud and image fusion Download PDFInfo
- Publication number
- CN116758506A CN116758506A CN202310786009.9A CN202310786009A CN116758506A CN 116758506 A CN116758506 A CN 116758506A CN 202310786009 A CN202310786009 A CN 202310786009A CN 116758506 A CN116758506 A CN 116758506A
- Authority
- CN
- China
- Prior art keywords
- point cloud
- fusion
- image
- roi
- detection method
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 47
- 230000004927 fusion Effects 0.000 title claims abstract description 43
- 230000001629 suppression Effects 0.000 claims abstract description 13
- 238000000605 extraction Methods 0.000 claims abstract description 8
- 238000007781 pre-processing Methods 0.000 claims abstract description 6
- 238000005070 sampling Methods 0.000 claims description 11
- 238000011176 pooling Methods 0.000 claims description 8
- 238000001914 filtration Methods 0.000 claims description 3
- 230000011218 segmentation Effects 0.000 claims description 3
- 238000013528 artificial neural network Methods 0.000 claims description 2
- 238000000034 method Methods 0.000 abstract description 11
- 230000004888 barrier function Effects 0.000 abstract 1
- 238000010586 diagram Methods 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 4
- 239000013598 vector Substances 0.000 description 4
- 230000007547 defect Effects 0.000 description 3
- 230000006872 improvement Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 2
- 238000005286 illumination Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 206010039203 Road traffic accident Diseases 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Databases & Information Systems (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a three-dimensional vehicle detection method based on point cloud and image fusion, in particular to a three-dimensional vehicle detection method for extracting color texture information in an image and size and position information in point cloud data by using a laser radar-camera sensor fusion architecture. The method is realized by the following steps: preprocessing point cloud data and image data; designing an improved key point module for feature extraction; the improved interesting fusion module is used for completing feature fusion of the point cloud and the image data; the use of improved non-maximum suppression reduces redundancy of the detection frame. The method mainly solves the problems that unmanned vehicles in traffic scenes are difficult to accurately measure and estimate the distance of vehicle barriers, realizes information complementation of different modes, provides information such as position, contour, speed and the like for navigation, obstacle avoidance and path planning, and makes up for a short plate for estimating the target distance of an unmanned automobile team.
Description
Technical Field
The invention relates to the field of computer vision, in particular to a three-dimensional vehicle detection method based on point cloud and image fusion.
Background
In recent years, unmanned technology is rapidly developed, becomes a focus and focus of attention, and has important significance for development of fields such as transportation, intelligent manufacturing, service robots and the like and improvement of human life quality. The increase of the number of motor vehicles brings unprecedented pressure to the traffic environment, and scientific research institutions and enterprises in all countries of the world input a large amount of manpower and material resources to develop related researches and exploration for the improvement of driving safety and the reduction of the occurrence frequency of traffic accidents.
Along with the superiority of the deep learning technology in the field of target detection, a vehicle detection method of an unmanned vehicle is rapidly developed, but the method still has great defects in the actual traffic environment, firstly, the traffic environment is intricate and complex, the shielding degree among vehicles is different, and the difficulty in vehicle detection is high; secondly, illumination is uneven, and false detection and omission among vehicles are easy to cause. Most existing unmanned automobile vision systems adopt monocular cameras, specific depth information of a vehicle to be detected cannot be obtained, a sensor is required to continuously sense the position of the vehicle in the running process of the vehicle, and multiple image processing operations are carried out, so that inaccurate positioning and overlong time are caused, and the running efficiency of the unmanned vehicle is seriously affected. The main research direction of the current environment perception is to use the fusion of point cloud data and image information to make up characteristic information, separately process the point cloud data and the image information, and then use the sub-network fusion to output results for the target detection of the three-dimensional vehicle.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a three-dimensional vehicle detection method based on point cloud and image fusion, which reduces false detection rate and omission rate of a three-dimensional vehicle.
The three-dimensional vehicle detection method based on the point cloud and the image fusion is realized by the following steps:
step one, preprocessing point cloud data and image data;
step two, designing a key point module for feature extraction;
thirdly, finishing feature fusion of the point cloud and the image data by using an improved interesting fusion module;
and step four, reducing redundancy of the detection frame by using improved non-maximum suppression.
The preprocessing step comprises the steps of removing edge clutter points and noise points in laser radar data through a statistical filtering algorithm, and reducing calculated amount for subsequent detection; the laser radar and the monocular camera are registered in time and space, so that coordinate systems of the two sensors are corresponding and unified, and a foundation for multi-mode target detection is laid.
The key point module is used for respectively extracting the point cloud data and the image data to generate key points of the vehicle to be detected through feature extraction, and estimating respective interested areas according to the key points.
The interesting fusion module is used for extracting local features on the point cloud and the image region, carrying out fusion operation on the 3D ROI and the 2D ROI corresponding to the point cloud data and the image data, completing fusion of the point cloud data and the 2D ROI by adopting a fusion strategy of attention fusion, and predicting the fused local features by using the fused local features in a three-dimensional detection frame.
The improved non-maximum value suppression module is used for setting a fixed threshold value to a threshold value of an improved self-adaptive size, and self-adaptively adjusting the threshold value of non-maximum value suppression according to the density of vehicle distribution, so that the accuracy of three-dimensional vehicle detection is improved as much as possible.
The invention has the following effective effects: 1. the filtering of the point cloud data and the clipping of the image data reduce the operation amount of the data. 2. The feature fusion of the point cloud data and the image data can reduce the false detection rate and the omission rate of vehicles in an unmanned scene, and make up for the defect of lack of depth information in a monocular camera. 3. Through information fusion of a plurality of modes, data of a plurality of angles are provided for environmental information in an unmanned scene, so that robustness of the data information in a complex environment is enhanced, and a more accurate precondition is provided for three-dimensional vehicle detection.
Drawings
In order to more clearly illustrate the embodiments of the invention or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, it being obvious that the drawings in the following description are only some embodiments of the invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a schematic flow chart of steps one to four in the present invention.
FIG. 2 is a diagram of a calibration architecture of a data preprocessing sensor in accordance with the present invention.
Fig. 3 is a diagram showing a time synchronization structure of a sensor according to the present invention.
Fig. 4 is a block diagram of a region of interest attention fusion module in accordance with the present invention.
FIG. 5 is a flow chart of improved non-maximum suppression in the present invention.
Detailed Description
The present invention will be described in detail with reference to specific examples.
The three-dimensional vehicle detection method based on the point cloud and the image fusion is realized by the following steps, as shown in fig. 1;
step one, preprocessing point cloud data and image data; step two, designing a key point module for feature extraction; thirdly, finishing feature fusion of the point cloud and the image data by using an improved interesting fusion module; and step four, reducing redundancy of the detection frame by using improved non-maximum suppression.
In the first step, the specific steps include:
according to different traffic scenes and different vehicle detection difficulties, collecting a data set, and judging whether outliers exist according to the density of the point cloud according to the sparse distribution characteristics of the outliers in the point cloud data. Firstly, searching a neighborhood point of each point according to the position of the point, then calculating to obtain the distance from each point to the neighborhood point, wherein the distance value meets the characteristic of Gaussian distribution, and the mean value and the variance can be obtained by modeling the distance parameter according to the characteristic of the Gaussian distribution, and the mean value and the variance have the following calculation formulas:
wherein d is ij For the distance from m points to k clinical points, an average confidence interval, the distance threshold, is calculated, and if a point is outside the center interval, the point is determined to be an outlier, and if it is within the interval, it is not an outlier. The threshold calculation formula is as follows:
d=μ+λσ (3)
where λ is a coefficient of the threshold. By clipping the fixed point cloud and the size of the image, feature matching is achieved with multiple views having the same aspect ratio.
And converting measured values of different sensor coordinate systems into the same coordinate system, establishing a coordinate conversion relation among an accurate laser radar coordinate system, a three-dimensional world coordinate system, a camera coordinate system, an image coordinate system and a pixel coordinate system, and converting measuring points under the laser radar coordinate system into the pixel coordinate system corresponding to the camera through the coordinate system to realize the spatial synchronization of the two, wherein a multi-sensor spatial calibration structure diagram is shown in figure 2.
The sampling frequencies of the camera sensor and the laser radar are different, so that the two sensors are synchronized in time after the space calibration is finished, the sampling period of the laser radar is 50ms, namely the sampling frequency is 20Hz, the camera sensor samples 30 pictures per second, namely the sampling frequency is 33.3ms, the time synchronization among the sensors is realized by adopting an interval sampling mode, the least common multiple of 100ms of the two sensors is selected as the sampling period of the fusion system, and the sensor time registration schematic diagram is shown in fig. 3.
In the second step, the specific steps include:
the key points in the point cloud are extracted by the SA layer in the PointNet++, the input point cloud is subjected to downsampling operation through the sampling layer, the downsampling operation adopts the furthest point sampling to obtain the center of the local area, and the point cloud data are respectively input to the F-FPS module and the D-FPS module. For the positioning of objects in the point cloud, especially far-small and occluded targets, the texture and color of the objects have certain importance, so that an image segmentation network is adopted to guide the selection of key points, and the segmentation characteristics of the point cloud are realized by using a characteristic propagation (FP) layer in PointNet++. The output of the SA layer is transmitted into the FP layer, the interpolation method based on the weighted average of the K nearest inverse distances is used for carrying out up-sampling operation on the points, and the calculation formula is as follows:
w in i (x)=1/d(x,x i ) p Representing an i-th neighbor point x corresponding to a certain point and k nearest neighbors i Inverse square of the euclidean distance between them. The interpolation point feature is connected with the jump connection point feature in the corresponding SA layer. The connected features pass through a PointNet of one unit, the properties of the connected features are similar to 1X 1 convolution in convolution, the feature vectors of points are updated by using a shared full connection layer and a ReLu layer, and finally the FN layer outputs semantic features corresponding to the same points of the original data. The image contains rich color and texture information, so that additional data information can be provided for the point cloud, and the accuracy of three-dimensional detection is improved. After the image is input, the image enters a deep v3+ neural network to obtain the segmentation characteristics and classification scores of the pixels, so that the selection of key points can ignore the background.
In the third step, the specific steps include:
and obtaining a central point of the target according to the position and the characteristics of the key point, and obtaining the spatial offset between the central point and the corresponding true value by adopting a partial network of Votenote. Firstly, inputting the point cloud with XYZ axes in the previous step into a network, carrying out up-sampling operation on the point cloud, learning deep features, and then outputting a subset of M points with XYZ coordinates and rich C-dimensional feature vectors. These point subsets with (3+C) dimensions are referred to as seed points, and votes are generated from each seed point independently by a shared voting module implemented by a fully connected layer, reLu, and batch normalized multi-layer perceptron (MLP). The multilayer perceptron characterizes the seed i To output the offset delta x of the European space i And a characteristic offset Δf i So that it votes v i =[y i ;g i ]From y i =x i +Δx i And g i =f i +Δf i The composition is formed. When the seed point s i When the three-dimensional offset is positioned on the surface of an object, the 3D offset is predicted as shown in the following formula:
where M is the number of seeds on the surface of the object,is the position x of the seed i Offset to the center of the three-dimensional bounding box of the object in which it is located.
And carrying out pooling operation on surrounding points of the previously obtained center points by adopting a 3D ROI pooling layer, learning local features of cluster points near each center point, and carrying out coding operation on the ROI by using a 3D boundary box. Using the obtained center points to center (x o ,y o ,z o ) Parameterization is achieved with the enlarged dimensions as the length and width of the region of interest and the height of the enlarged dimensions as the height value of the region of interest. The dimension of the region of interest may be defined as (x o ,y o ,z o ,δ+h i ,δ+w i ,δ+l i ) Where δ is an expanding parameter. The points in each 3D ROI are moved to relative positions according to the center point for better learning of local features, and then clustered on points within the 3D ROI using a sub-network with stacked multi-layer perceptron layers to extract pooled features of the local ROIs. After the 3D ROI is obtained, it is projected onto the image to generate a corresponding 2D ROI, and the local texture features of the 2D ROI are learned by the 2D ROI pooling layer following in a similar manner to the processing of the 3D ROI. And finally, carrying out fusion operation on the 3D ROI and the 2DROI corresponding to the three-dimensional point cloud and the two-dimensional image, and completing fusion of the three-dimensional point cloud and the 2D ROI through a fusion strategy of attention fusion, wherein the local features are finally fused together for 3D boundary box prediction.
Using an attention fusion algorithm to obtain a two-part ROIFully fusing, weighting the ROI of the two-dimensional image and the ROI of the point cloud data through channels, wherein the ROI in the image is F i ∈ N×N×C ROI in the point cloud is F l ∈ N×N×C Using channel splicing operation to splice F i And F l Realizing fusion and representing the fused data as F f ∈ N×N×2C . Then F is carried out f Input to global average pooling operation, and output F is obtained after pooling operation fa ∈ 2C 。F fa Output is F through the full connection layer fc ∈ 2C F is to F fc Remodelling to F 1 ∈ 2×C Then for F 1 The channel uses the Softmax function to obtain the following equation:
wherein A and B are F 1 Is F, a and b are the first row vector and the second row vector of l And F i C is the number of channels per ROI. And finally, the ROI fuses a calculation formula through an attention mechanism as follows:
wherein F is w Is F l And F i The attention mechanism fusion structure is shown in figure 4.
In the fourth step, the specific steps include:
according to the object distribution density, the threshold value of non-maximum value suppression is adaptively adjusted, when the object to be detected is denser, the correct prediction frame is reserved as far as possible, and when the object to be detected is sparser, the redundant prediction frame is removed by adopting the threshold value transferred to by the smaller non-maximum value. The density of the predicted candidate frames is defined as the maximum IoU of all the rest of the real frames, and the density of the target to be detected is expressed as follows:
wherein i is a generated prediction candidate frame, and j is a real frame. The threshold value thereof can be defined as follows:
N M =Max(N t ,d M ) (9)
m is the prediction candidate frame with the highest confidence score in the set. Finally, combining with Soft-NMS to obtain confidence penalty formula as follows:
wherein the method comprises the steps ofThe improved non-maximum suppression architecture is shown in fig. 5 by the confidence penalty formula, which allows for an adaptive process for threshold selection for the NMS.
The invention and the current main stream algorithm are put into the same traffic scene under the same data set for experiment, the data in the comparison algorithm are all from the KITTI official network, three different difficulties of simplicity, generality and difficulty are respectively selected for comparison analysis, and the detection methods are divided into an image detection method, a point cloud detection method and a detection method of point cloud and image fusion, and the test comparison data are shown in the table 1. The three-dimensional vehicle detection method of the pure image is faster, but the detection precision still has a larger room for improvement, and the three-dimensional vehicle detection method of the pure point cloud is far higher than the image detection method in precision, but slightly lower than the vehicle detection algorithm of the point cloud and image fusion.
Table 1 three-dimensional vehicle detection algorithm time and accuracy comparison
And respectively selecting an algorithm from different detection modes to carry out experimental comparison with the method, taking a traffic scene with more complex data set as a comparison experimental environment, wherein the vehicle detection precision of the method is 83.68 percent, the average time for processing single frame data is 0.11s under the traffic scenes such as vehicle shielding, poor illumination, far-small vehicles and the like, and the method has good detection result, and the false detection rate and the omission rate are effectively reduced.
Claims (5)
1. The three-dimensional vehicle detection method based on the point cloud and the image fusion is characterized by comprising the following steps of:
step one, preprocessing point cloud data and image data;
step two, designing a key point module for feature extraction;
thirdly, finishing feature fusion of the point cloud and the image data by using an improved interesting fusion module;
and step four, reducing redundancy of the detection frame by using improved non-maximum suppression.
2. The three-dimensional vehicle detection method of point cloud and image fusion according to claim 1, characterized in that: in the first step, the specific steps include:
the laser radar and the monocular camera acquire data and respectively output point cloud data and image data, the joint calibration in space is realized according to the joint calibration principle of the sensors, and the time synchronization among the sensors is realized by selecting an interval sampling mode; and reducing the interference of outliers in the point cloud data through a statistical filtering algorithm.
3. The three-dimensional vehicle detection method of point cloud and image fusion according to claim 1, characterized in that: in the second step, the specific steps include:
the improved key point module performs feature extraction in the point cloud data and the image data respectively to generate key points of a vehicle to be detected, the key point module in the point cloud performs feature extraction by an SA layer in PointNet++, and simultaneously performs downsampling and depth feature extraction on the point cloud, and downsampled 3D points and corresponding features are obtained; the image data is input to the deeplabv3+ neural network to obtain the segmentation characteristics and classification scores of the pixels, so that the selection of the key points can ignore the background.
4. The three-dimensional vehicle detection method of point cloud and image fusion according to claim 1, wherein: in the third step, the specific steps include:
and obtaining the central point of the target vehicle according to the positions and the characteristics of the key points, and acquiring the spatial offset between the central point and the corresponding true value by adopting a partial network of Votenote. And carrying out pooling operation on surrounding points of the previously obtained center points by adopting a 3D ROI pooling layer, learning local features of cluster points near each center point, and carrying out coding operation on the ROI by using a 3D boundary box. After the 3D ROI is obtained, it is projected onto the image to generate a corresponding 2D ROI, whose local texture features are learned by the 2D ROI pooling layer followed. Finally, the fusion of the 3D ROI and the 2D ROI is done using an improved attention fusion strategy, these local features being fused together for prediction of the 3D bounding box.
5. The three-dimensional vehicle detection method of point cloud and image fusion according to claim 1, wherein: in the fourth step, the specific steps include:
the improved target detection under the non-maximum suppression self-adaptive traffic scene is designed, the non-maximum suppression threshold value is self-adaptively adjusted according to the distribution density of the vehicles to be detected, and the detection accuracy of the vehicles is improved. According to the density of the object distribution, the threshold value of non-maximum suppression is self-adaptively adjusted, when the object to be detected is denser, the correct prediction frame is reserved as far as possible, and when the object to be detected is sparser, the redundant prediction frame is removed by adopting the smaller threshold value of non-maximum suppression.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310786009.9A CN116758506A (en) | 2023-06-29 | 2023-06-29 | Three-dimensional vehicle detection method based on point cloud and image fusion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310786009.9A CN116758506A (en) | 2023-06-29 | 2023-06-29 | Three-dimensional vehicle detection method based on point cloud and image fusion |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116758506A true CN116758506A (en) | 2023-09-15 |
Family
ID=87949495
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310786009.9A Pending CN116758506A (en) | 2023-06-29 | 2023-06-29 | Three-dimensional vehicle detection method based on point cloud and image fusion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116758506A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116968758A (en) * | 2023-09-19 | 2023-10-31 | 江西五十铃汽车有限公司 | Vehicle control method and device based on three-dimensional scene representation |
-
2023
- 2023-06-29 CN CN202310786009.9A patent/CN116758506A/en active Pending
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116968758A (en) * | 2023-09-19 | 2023-10-31 | 江西五十铃汽车有限公司 | Vehicle control method and device based on three-dimensional scene representation |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110188696B (en) | Multi-source sensing method and system for unmanned surface equipment | |
CN110675418B (en) | Target track optimization method based on DS evidence theory | |
CN114565900A (en) | Target detection method based on improved YOLOv5 and binocular stereo vision | |
CN112149550B (en) | Automatic driving vehicle 3D target detection method based on multi-sensor fusion | |
CN110688905B (en) | Three-dimensional object detection and tracking method based on key frame | |
JP7091686B2 (en) | 3D object recognition device, image pickup device and vehicle | |
CN109961461B (en) | Multi-moving-object tracking method based on three-dimensional layered graph model | |
CN114495064A (en) | Monocular depth estimation-based vehicle surrounding obstacle early warning method | |
CN112183330B (en) | Target detection method based on point cloud | |
CN117058646B (en) | Complex road target detection method based on multi-mode fusion aerial view | |
CN115187964A (en) | Automatic driving decision-making method based on multi-sensor data fusion and SoC chip | |
CN113281718B (en) | 3D multi-target tracking system and method based on laser radar scene flow estimation | |
CN114972968A (en) | Tray identification and pose estimation method based on multiple neural networks | |
CN112927264A (en) | Unmanned aerial vehicle tracking shooting system and RGBD tracking method thereof | |
CN116758506A (en) | Three-dimensional vehicle detection method based on point cloud and image fusion | |
CN114089329A (en) | Target detection method based on fusion of long and short focus cameras and millimeter wave radar | |
CN116486287A (en) | Target detection method and system based on environment self-adaptive robot vision system | |
CN117519168A (en) | Robot navigation obstacle avoidance method based on visual image and laser radar | |
CN116978009A (en) | Dynamic object filtering method based on 4D millimeter wave radar | |
CN116310673A (en) | Three-dimensional target detection method based on fusion of point cloud and image features | |
CN114608522B (en) | Obstacle recognition and distance measurement method based on vision | |
CN117576665B (en) | Automatic driving-oriented single-camera three-dimensional target detection method and system | |
CN115909268A (en) | Dynamic obstacle detection method and device | |
CN114092778A (en) | Radar camera data fusion system and method based on characterization learning | |
CN112950786A (en) | Vehicle three-dimensional reconstruction method based on neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication |