CN115457132A - Method for positioning grid map based on binocular stereo camera - Google Patents

Method for positioning grid map based on binocular stereo camera Download PDF

Info

Publication number
CN115457132A
CN115457132A CN202211127974.7A CN202211127974A CN115457132A CN 115457132 A CN115457132 A CN 115457132A CN 202211127974 A CN202211127974 A CN 202211127974A CN 115457132 A CN115457132 A CN 115457132A
Authority
CN
China
Prior art keywords
grid
occupied
distance
free
current
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211127974.7A
Other languages
Chinese (zh)
Inventor
章雨昂
仲维
刘晋源
王维民
樊鑫
刘日升
罗钟铉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dalian University of Technology
Original Assignee
Dalian University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dalian University of Technology filed Critical Dalian University of Technology
Priority to CN202211127974.7A priority Critical patent/CN115457132A/en
Publication of CN115457132A publication Critical patent/CN115457132A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/05Geographic models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Remote Sensing (AREA)
  • Computer Graphics (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a method for positioning a grid map based on a binocular stereo camera, belonging to the field of image processing and computer vision. The method comprises the steps of obtaining images by using a visible light binocular stereo camera and an infrared binocular stereo camera together, obtaining a disparity map by using the binocular stereo camera to obtain real-time three-dimensional point cloud information, and constructing a high-quality grid mapping algorithm to realize the mapping from 3d point cloud to 2d grid map, wherein the grid height and grid threshold screening algorithm is realized according to prior information; adding a blocking counting algorithm to process the blocking problem in the barrier by combining with the space geometric knowledge; removing possible noise interference in the occupied grids through the posterior probability of the grids; using a target detection algorithm to assist in clustering the target objects and estimating the actual distance of the target objects; optimizing a dynamic target object by combining particle filtering and multi-frame information; and fusing the grid map results obtained by the visible light camera and the infrared camera and outputting the fused grid map results.

Description

Method for positioning grid map based on binocular stereo camera
Technical Field
The invention belongs to the field of image processing and computer vision, and relates to a method for positioning a grid map based on a binocular stereo camera.
Background
Stereoscopic vision has been one of the most popular research directions in the field of computer vision since the middle of the 20 th century. The stereoscopic vision obtains a parallax image by simulating human eyes to observe objects and by a stereoscopic matching technology, thereby obtaining three-dimensional information of a real scene, namely a depth map, and obtaining a three-dimensional point cloud model. At present, the binocular stereo vision technology is rapidly developed in the fields of automatic driving, robot navigation and path planning, vision measurement and the like. One of the most important problems in the field of robot navigation and path planning is the construction of a navigation map. In the practical application process of navigation and path planning, common navigation maps are generally divided into three categories, namely, topological maps, semantic maps and scale maps. The topological map generally has no actual distance information, only represents the connectivity and the path relation of two positions, and is most common in large-scale path navigation and planning; the scale map generally has actual distance information, including a point cloud map, a grid map, a feature map and the like, and is common in navigation map construction, positioning and small-scale path planning; the semantic map is added with labels on the basis of the scale map, and any place on the map can be represented by the labels and is commonly used for man-machine interaction. Among them, a scale map is of great interest because it can display real environmental information around the robot in real time. The most widely applied grid map in the scale map is the grid map, because the real scene around the carrier can be visually presented on a 2-dimensional plane, and the state of each grid represents whether an obstacle exists in the real scene, the grid map is more suitable for map construction and obstacle positioning.
Most conventional grid maps are generated by a laser radar (LIDAR) construction. The sensor of the laser radar can emit a laser signal towards the surrounding environment, when the sensor encounters an obstacle, the laser can be reflected back to the sensor, so that the distance from the sensor to the obstacle can be calculated through the round-trip time difference, and the laser radar aims to provide a complete 360-degree panoramic view, and the visualization of the surrounding environment of the vehicle is realized in a three-dimensional point cloud mode by using laser pulses. However, the disadvantages of the method are obvious, and for civil vehicles, compared with other devices, the laser radar is an expensive choice, so that most of the laser radar can only be used on high-grade vehicles, mass production and popularization cannot be achieved, and due to the fact that a radar sensor generates noise, the image resolution is generally low, and a real scene cannot be presented; for military use, although the laser radar can feed back surrounding environment information in real time, a large number of signals are transmitted to the outside while emitting laser, and the signals are likely to be utilized by enemies to expose the real position of the laser radar, so that great hidden danger is caused.
So far, an algorithm for navigation and positioning based on binocular stereo vision gradually enters the sight of people. The binocular stereo vision technology simulates stereo perception of human eyes to a three-dimensional space through a group of binocular stereo cameras, the binocular cameras are used for shooting a left image and a right image at the same time, then characteristic points in the left image and the right image are extracted for stereo matching, the parallax in the images is calculated by using a triangular distance measurement principle to obtain a parallax image, a depth image is further calculated through the parallax image, and therefore three-dimensional point information of each object in the images is obtained, a point cloud image is generated, and real three-dimensional information in a scene is restored. The advantages of navigation and positioning based on binocular stereo vision are: compared with a laser radar sensor and a structured light sensor, the cost of the binocular camera is much lower; the problem of target identification is reduced, whether the distance is calculated after the obstacle is identified is not needed, and the depth map is directly calculated on the disparity map; compared with the problem that the laser radar sensor is noisy, the binocular stereo vision estimates the depth by calculating the parallax, and the error is smaller. The binocular stereo vision has the defects that the calculation amount is large, the processing speed of a CPU is required to be fast enough, and therefore a special chip or an integrated calculation method on a circuit board is required to realize the calculation of the characteristic points and the stereo matching. And special FPGA has been researched at home and abroad to complete the calculation of binocular camera image stereo matching.
At present, most grid map construction methods construct two-dimensional maps, lack height information and cannot meet the requirements of equipment with three-dimensional mobility, such as unmanned aerial vehicles and the like. At present, a method for constructing a grid map by using a binocular stereo camera is urgently needed, so that the grid map can contain information of a three-dimensional real scene and can be used as civil vehicle-mounted navigation map equipment or military application. Compared with a laser radar sensor, the general binocular stereo camera has a poor effect in dark days or extremely severe weather.
Disclosure of Invention
In order to solve the problems, the invention provides a method for positioning a grid map by using a binocular stereo vision camera. The invention combines the results of a visible light binocular camera and an infrared binocular camera, so that the visible light binocular camera and the infrared binocular camera can show good imaging effect in various unknown scenes, and aims to abandon the expensive method for generating the grid map by using a laser radar sensor, obtain a disparity map by using a binocular stereo camera, estimate a depth map through the disparity map, and construct the grid map after converting the disparity map into three-dimensional point cloud. And constructing a high-performance operation platform by using a binocular stereo camera and a GPU, and constructing a high-performance solving algorithm to obtain a high-quality grid map containing three-dimensional information. The method can complete the acquisition of input data only by using a binocular stereo camera, and then realizes a grid height and grid threshold value screening algorithm by using spatial prior information and statistics; adding a blocking counting algorithm to process the blocking problem in the barrier by combining with the space geometric knowledge; removing noise interference possibly existing in the occupancy grid through the grid posterior probability; using a target detection algorithm to assist in clustering the target objects and estimating the actual distance of the target objects; and optimizing the dynamic target object by combining particle filtering and multi-frame information. The invention realizes an algorithm for constructing the two-dimensional grid map containing three-dimensional information by using the binocular stereo camera, and adds an efficient post-processing mechanism on the basis of the traditional mapping method to ensure that the grid information is more accurate.
The specific content comprises the following steps: the method comprises the steps of using a visible light binocular stereo camera and an infrared binocular stereo camera to jointly obtain images, respectively calculating a disparity map through the visible light images and the infrared images, fusing the disparity maps obtained by the visible light camera and the infrared camera to generate a fused disparity map, obtaining real-time three-dimensional point cloud information according to the fused disparity map, and constructing a high-quality grid mapping algorithm to realize mapping from 3d point cloud to 2d grid map, wherein the grid height and grid threshold screening algorithm is realized according to prior information; adding a blocking counting algorithm to process the blocking problem in the barrier by combining with the space geometric knowledge; removing possible noise interference in the occupied grid through the posterior probability of the grid; using a target detection algorithm to assist in clustering the target objects and estimating the actual distance of the target objects; and optimizing the dynamic target object by combining particle filtering and multi-frame information.
The technical scheme of the invention is as follows:
the method for positioning the grid map based on the binocular stereo camera comprises the following steps:
1) Acquiring three-dimensional point cloud information:
1-1) respectively acquiring RGB images of a real scene and corresponding parallax images thereof through a binocular visible light camera and an infrared camera;
1-2) fusing the visible light parallax image and the infrared parallax image, wherein the fusing method comprises the following steps: comparing the confidence coefficient of each parallax in the visible light parallax image and the infrared parallax image, keeping the parallax value with large confidence coefficient in the visible light parallax image and the infrared parallax image, and finally obtaining a fusion parallax image;
1-3) generating a three-dimensional point cloud according to the fusion disparity map, and converting the three-dimensional point cloud into a world coordinate system.
2) Initializing a grid map:
2-1) determining the size of the grids, wherein the side length of each grid represents actual distance information;
2-2) determining the spatial position of the grid corresponding to the real scene: the range in the z direction (closest and farthest distances) of the perceived obstacle, and the range in the y direction (highest height h) of the obstacle are set max And a minimum height h min ) And the calculated range (field angle) of the obstacle in the x direction;
2-3) initializing grid state: idling;
3) Mapping the three-dimensional point cloud to a two-dimensional grid:
3-1) carrying out noise point filtering on the generated three-dimensional point cloud through bilateral filtering, and mapping the filtered three-dimensional point cloud to a two-dimensional grid map plane;
3-2) counting the number of three-dimensional points mapped in the grid and the height information of the three-dimensional points;
4) Grid height and threshold screening:
4-1) obtaining the height of the three-dimensional points mapped into each grid, sorting the height of the three-dimensional points in each grid, removing the height of the three-dimensional points of the first 10% and the second 10% after sorting, and carrying out weighted average on the heights of the three-dimensional points after screening to obtain the final grid height h i
4-2) obtaining the number of three-dimensional points num mapped into each grid i Calculating a mapping threshold T for each grid i The formula is as follows:
Figure BDA0003848944790000041
wherein img width Is the width of the image, h i Height of the current grid, Z k The grid depth of the k-th row is theta, the field angle size of the initialization setting is theta, and alpha is a preset smoothing coefficient; formula (1) represents the number of pixels represented by a plane formed by the grid of the k-th row and the height of the grid as the threshold value of the number of grids;
4-3) comparing the number of three-dimensional points num in each grid i And a mapping threshold T for each grid i If num i >T i If the grid is in the occupied state, otherwise, the grid is in the idle state;
4-4) if the grid is in the occupied state, the height h of the grid is set i Comparison with the y-direction Range of the initialization grid settings (h) max And h min ) If h is min <h i <h max Then h is given i Setting the height value of the current grid, otherwise setting the current grid state to be idle;
5) Screening sheltered obstacles: processing occlusion by a occlusion counter;
5-1) judging whether the grid belongs to a blocking state:
there are three cases of a blocking grid:
a. out of a set distance range or too close to the binocular camera;
b. the grid and the origin are spaced apart by one or more obstacles;
c. the height of a background obstacle in the same obstacle is lower than that of a foreground obstacle;
5-2) find the initial occupied grid, its height is noted gridh i
5-3) using polar coordinates, the occupied-state grid height gridh of the same column after the first occupied-state grid i-1 And gridh i Comparison when gridh i >gridh i-1 When the blocking value is greater than the blocking threshold β, the grid is set to the blocking state, i.e., the barrier is blocked.
6) Noise point elimination: removal of noise interference possibly present in an occupied grid by a grid posterior probability
6-1) calculating the z-direction distance error of the current three-dimensional point under a world coordinate system, wherein the formula is as follows:
Figure BDA0003848944790000051
wherein b is a base line of the binocular stereo camera, f is the focal length of the camera, z is a z-direction coordinate of the three-dimensional point in a world coordinate system, and delta d Is a third scale factor in the disparity calculation;
6-2) calculating the x-direction distance error of the current three-dimensional point in a world coordinate system, wherein the formula is as follows:
Figure BDA0003848944790000052
d is the distance from the origin of the camera to the three-dimensional point in the world coordinate system, x is the x-direction coordinate of the three-dimensional point in the world coordinate system, and δ z is the z-direction distance error of the three-dimensional point in the world coordinate system;
6-3) converting the distance error of the current three-dimensional point in the z and x directions under a world coordinate system into the relative row and column error of the two-dimensional grid map, wherein the formula is as follows:
Figure BDA0003848944790000053
Figure BDA0003848944790000054
wherein delta z And delta x Distance errors of the three-dimensional points in the z direction and the x direction under a world coordinate system are obtained, and dz and dx are the resolutions of grids relative to the world coordinate system;
6-4) calculating the probability density of the occupation of the grid within the error range of the row and the column by taking the current grid as the center, wherein the formula is as follows:
wherein P is occupied Representing the occupancy probability density of the current grid:
Figure BDA0003848944790000061
P free free probability density representing the grid:
P free =1-P occupied
δ row and delta col gridNum, the relative row-column error of the grid occupied Representing the number of cells occupying the state gridNum total Is delta row And delta col RangeTotal number of inner grids. The occupancy probability density of the current grid is δ with the current grid as the center row And delta col The ratio of the number of all occupied grids in the range to the number of all grids in the range; similarly, the free probability density of the current grid is the complement of the occupied probability density.
6-5) calculating the distance probability density of the current grid to the nearest occupied grid and the free grid, wherein the formula is as follows: wherein P is distance (anchored) represents the distance probability density of the current grid to the nearest occupied grid:
Figure BDA0003848944790000062
P distance (free) represents the distance probability density of the current grid to the nearest free grid:
Figure BDA0003848944790000063
wherein
Figure BDA0003848944790000064
And
Figure BDA0003848944790000065
representing the row-column difference of the current grid to the most recently occupied grid,
Figure BDA0003848944790000066
and
Figure BDA0003848944790000067
representing the row-column difference from the current grid to the nearest free grid;
6-6) calculating the occupation weight coefficient and the idle weight coefficient of the current grid, wherein the formula is as follows:
wherein ω is occupied Occupancy weight coefficient for current grid:
ω occupied =P occupied *P distance (occupied)
ω free for the idle weight coefficient of the current grid:
ω free =P free *P distance (free)
wherein P is occupied And P free Respectively, the occupation and idle probability densities, P, of the current grid distance (occupied) and P distance (free) distance probability densities of the current grid to the nearest occupied and free grids, respectively;
6-7) calculating the posterior occupancy probability of the current grid, wherein the formula is as follows:
Figure BDA0003848944790000071
wherein gridNum occupied Is delta row And delta col Number of occupied grids in the range gridNum free Is delta row And delta col The number of free grids in the range.
6-8) according to P posterior If P is a noise point, judging whether the occupied grid is a noise point posterior Less than the threshold, the occupied grid is considered likely to be noisy, and the occupied grid is changed to a free grid.
7) Clustering the target objects: target object clustering is assisted through target detection algorithm and actual distance of target objects is estimated
7-1) obtaining a target detection result by using yoloV5 algorithm, and obtaining the number Num of target objects in a three-dimensional scene object Dis, distance estimate per object i And the Center of each target object in the world coordinate system i (x,y,z);
7-2) clustering the targets using a modified K-Means clustering algorithm: center of each target i (x, y, z) as the initialized clustering center point, and the number Num of the target objects object As a K value, the K-Means clustering algorithm K unknown defect is made up, the operation speed is increased, and the formula is as follows:
Figure BDA0003848944790000072
Figure BDA0003848944790000073
wherein theta is n (x, y, z) is a three-dimensional point data set, center k For the cluster centers, k represents the number of cluster centers, α nk For the binary variable 0,1 represents that the three-dimensional point belongs to class k, 0 represents not,
Figure BDA0003848944790000074
Figure BDA0003848944790000075
when the value of j is represented, the value of theta is minimized n (x,y,z)-Cente j || 2 And in fact, distributing the three-dimensional point data set to the nearest clustering center, continuously performing iterative computation until the algorithm converges to minimize loss, and finishing clustering.
8) Multi-frame fusion: optimization of dynamic objects by particle filtering in combination with multi-frame information
8-1) obtaining the screened grid map model, and adding particles with different speeds into the grids in the occupied state. The particles are marked as S, S = { S = } i |s i =(x i ,y i ,vx i ,vy i ,p i ),i=1,...N S In which x i Lines representing the grid in which the particles are located, y i Columns representing the grid in which the particles are located, vx i And vy i Respectively representing the row and column velocities of the particles, p i Representing the period of the particle.
8-2) predicting the motion state of each particle according to the information of the next frame; constructing a prediction model, wherein the formula is as follows:
the angle and distance between two adjacent frames are:
Figure BDA0003848944790000081
Figure BDA0003848944790000082
wherein
Figure BDA0003848944790000083
Is the angular velocity, v is the velocity, Δ t is the time interval between two frames; the relative movement between grids is:
Figure BDA0003848944790000084
Figure BDA0003848944790000085
DX and Dy are resolution ratios of the grid lines and the columns respectively; the grid position in the current frame may be denoted as x n ,y n
Figure BDA0003848944790000086
The prediction model is:
Figure BDA0003848944790000087
the delta y, the delta x, the delta vy and the delta vx are random disturbance and are derived from a Gaussian distribution with zero mean and a state transition covariance matrix of a Kalman filter.
8-3) reassign the particles in the grid by weighting and resampling:
a. calculating an occupancy probability density and a vacancy probability density for each grid
Wherein p is density (m (x, y) | occupied) represents the occupancy probability density of the grid:
Figure BDA0003848944790000088
p density (m (x, y) | free) represents the free probability density of the grid:
P density (m(x,y)|free)=1-P density (m(x,y)|occupied)
δ x and delta y Represents the error of the transverse and longitudinal distances projected by the three-dimensional points to the grid, grid (actual) represents the occupied grid, and the meaning of the formula is that the current grid delta x ,δ y Taking the ratio of the number of all occupied state grids in the range to the number of all grids in the range as the occupation probability density of the current grid;
b. calculating the occupation distance probability and the idle distance probability of each grid:
the occupied distances are represented in rows and columns, respectively, as:
Figure BDA0003848944790000091
Figure BDA0003848944790000092
row and Col represent the closest Row and column occupying the grid to the current grid, respectively
The free distance is represented in rows and columns, respectively, as:
Figure BDA0003848944790000093
Figure BDA0003848944790000094
the distance probability is expressed using a multidimensional gaussian function:
Figure BDA0003848944790000095
c. calculating an occupancy weight and an idle weight for each grid:
ω occupied (x,y)=p density (m(x,y)|occupied)*p distance (m(x,y|occupied)
ω free (x,y)=p density (m(x,y)|free)*p distance (m(x,y)|free)
d. the occupancy posterior probability for each grid is calculated:
Figure BDA0003848944790000096
wherein N is i To occupy the actual number of particles of the grid, N max Is the maximum number of particles allowed in a grid;
e. calculating the number of resamples N per grid Ri =P i *N max And calculating the ratio of the number of resamples to the actual number
Figure BDA0003848944790000097
If f i If > 1, add f to the grid i Particles in whole number multiples of f i If < 1, deleting the particle;
8-4) judging the grid state again according to the number of particles and the particle speed:
the occupancy probability of a grid is the ratio of the number of particles of the current grid to the maximum number of particles of the grid:
Figure BDA0003848944790000101
the current grid velocity is the vector sum of the particle velocities in the current grid:
Figure BDA0003848944790000102
if P of the grid occupied If the grid speed is less than the speed of all particles in the grid, the grid is in an occupied stateThe standard deviation of the degrees indicates that the barrier in the grid is in a static state;
9) And counting the grids in the occupied state in the two-dimensional grids to generate a two-dimensional grid map.
The beneficial effects of the invention are:
compared with a common grid map algorithm, the method combines spatial geometric knowledge to add a blocking counting algorithm to process the problem of shielding in the barrier, so that the output result is more accurate; according to the invention, the grid height and grid threshold screening algorithm is realized according to the prior information, and the generalization capability and robustness of the algorithm are improved; noise interference possibly existing in the occupied grid is removed through the posterior probability of the grid, and the accuracy of the grid map is improved; the target detection algorithm is used for assisting in clustering the target objects and estimating the actual distance of the target objects, so that the problem that the target objects of a common grid map are ambiguous is solved; the result of the grid map is optimized by combining particle filtering and multi-frame information, so that the result optimization between continuous frames is realized; the parallax image results obtained by the visible light camera and the infrared camera are fused to generate a fused parallax image, the grid map is constructed according to the fused parallax image, and the good grid map result can be displayed in a complex scene.
Drawings
Fig. 1 is a left image captured by a binocular camera in the embodiment.
Fig. 2 is a disparity map in the embodiment.
Fig. 3 is an overall algorithm flowchart.
Fig. 4 is a flow chart of altitude and threshold and occlusion policies.
Fig. 5 is a flowchart of the target detection clustering strategy.
Fig. 6 is a flow chart of implementing multi-frame dynamic optimization by particle filtering.
Fig. 7 shows the grid map result without any optimization strategy.
Fig. 8 is the grid map results of fig. 7 after adding height and threshold screening and after blocking strategy.
Fig. 9 is a grid map result after the target detection clustering target object strategy and the particle filtering strategy are added to fig. 8.
Detailed Description
The following further describes a specific embodiment of the present invention with reference to the drawings and technical solutions.
The basic flow of the method for creating the two-dimensional grid map based on the binocular stereo camera is shown in fig. 3, and the method comprises the following specific steps:
1) Acquiring three-dimensional point cloud information:
1-1) respectively acquiring RGB (red, green and blue) images of a real scene and corresponding parallax images thereof through a binocular visible light camera and an infrared camera; the left image and the disparity map are shown in fig. 1 and fig. 2, respectively.
1-2) fusing the visible light parallax image and the infrared parallax image, wherein the fusing method is to compare the confidence coefficient of each parallax in the visible light parallax image and the infrared parallax image, retain the parallax value with high confidence coefficient in the visible light parallax image and the infrared parallax image, and finally obtain a fused parallax image;
1-3) generating a three-dimensional point cloud according to the fusion disparity map, and converting the three-dimensional point cloud into a world coordinate system.
2) Initializing a grid map:
2-1) determining the size of the grids, wherein the side length of each grid represents actual distance information;
2-2) determining the spatial position of the grid corresponding to the real scene: setting a range (closest and farthest distances) in a z direction of a perceived obstacle, as well as a range (highest height and lowest height) of the obstacle in a y direction and a calculation range (angle of view) of the obstacle in an x direction;
2-3) initializing grid states:
initializing grid states including idle, occupied and blocked;
3) The 3d point cloud maps to a 2d grid:
3-1) carrying out noise point filtering on the generated three-dimensional point cloud through bilateral filtering, and mapping the filtered three-dimensional point cloud to a 2-dimensional grid map plane;
3-2) counting the number of three-dimensional points mapped in the grid and the height information of the three-dimensional points;
4) Grid height and threshold screening (flow shown in fig. 4):
4-1) obtaining three-dimensional point heights mapped into each grid, for each gridSorting the three-dimensional point heights in the grids, removing the three-dimensional point heights of the front 10% and the rear 10% after sorting, and performing weighted average on the three-dimensional point heights after screening to obtain the final grid height h i
4-2) obtaining the number of three-dimensional points num mapped into each grid i Calculating a mapping threshold T for each grid i The formula is as follows:
Figure BDA0003848944790000121
wherein img width Is the width of the image, h i Height of the current grid, Z k The grid depth of the k-th row, theta, the field angle size of the initialization setting, and alpha, a preset smoothing coefficient, wherein the formula is expressed as the number of pixels represented by a plane formed by the grid of the k-th row and the height of the grid as the threshold value of the number of the grids;
4-3) comparing the number of three-dimensional points num in each grid i And a mapping threshold T for each grid i If num i >T i If the grid is in the occupied state, otherwise, the grid is in the idle state;
4-4) if the grid is in the occupied state, the height h of the grid is set i Comparison with the y-direction Range of the initialization grid settings (h) max And h min ) If h is min <h i <h max Then h will be i Setting the height value of the current grid, otherwise setting the current grid state to be idle;
5) Screening sheltered obstacles: handling occlusion by blocking counters
5-1) judging whether the grid belongs to a blocking state:
there are three cases of a blocking grid:
a. out of a set distance range or too close to the binocular camera;
b. the grid and the origin are spaced apart by one or more obstacles;
c. the height of a background obstacle in the same obstacle is lower than that of a foreground obstacle;
5-2) find the initial occupied grid, its height is noted gridh i
5-3) using polar coordinates, the occupied-state grid height gridh of the same column after the first occupied-state grid i-1 And gridh i Comparison, when gridh i >gridh i-1 When the blocking value is larger than the blocking threshold value beta, setting the grid to be in a blocking state;
6) Noise point elimination: removal of noise interference possibly present in an occupied grid by a grid posterior probability
6-1) calculating the z-direction distance error of the current three-dimensional point under a world coordinate system, wherein the formula is as follows:
Figure BDA0003848944790000122
wherein b is a base line of the binocular stereo camera, f is the focal length of the camera, z is a z-direction coordinate of the three-dimensional point in a world coordinate system, and delta d A third scale factor in the disparity calculation;
6-2) calculating the x-direction distance error of the current three-dimensional point in a world coordinate system, wherein the formula is as follows:
Figure BDA0003848944790000131
wherein d is the distance from the origin of the camera to the three-dimensional point in the world coordinate system, x is the x-direction coordinate of the three-dimensional point in the world coordinate system, and delta z The z-direction distance error of the three-dimensional point under a world coordinate system is obtained;
6-3) converting the distance error of the current three-dimensional point in the z and x directions under a world coordinate system into the relative row error of the two-dimensional grid map, wherein the formula is as follows:
Figure BDA0003848944790000132
Figure BDA0003848944790000133
wherein delta z And delta x For the z-direction (x-direction) distance error of a three-dimensional point in the world coordinate system, dz and dx are the resolutions of the grid with respect to the world coordinate system. 100mm in this invention;
6-4) calculating the probability density of the occupation of the grid within the error range of the row and the column by taking the current grid as the center, wherein the formula is as follows:
wherein P is occupied Representing the occupancy probability density of the current grid:
Figure BDA0003848944790000134
P free free probability density representing the grid:
P free =1-P occupied
δ row and delta col gridNum, the relative row-column error of the grid occupied Representing the number of cells occupying the state gridNum total Is delta row And delta col The total number of grids within the range. The occupancy probability density of the current grid is δ with the current grid as the center row And delta col The ratio of the number of all occupied grids in the range to the number of all grids in the range; similarly, the free probability density of the current grid is the complement of the occupied probability density.
6-5) calculating the distance probability density of the current grid to the nearest occupied (free) grid, and the formula is as follows: wherein P is distance (occupied) represents the distance probability density of the current grid to the nearest occupied grid:
Figure BDA0003848944790000141
P distance (free) represents the distance probability density of the current grid to the nearest free grid:
Figure BDA0003848944790000142
wherein
Figure BDA0003848944790000143
And
Figure BDA0003848944790000144
representing the row-column difference of the current grid to the most recently occupied grid,
Figure BDA0003848944790000145
and
Figure BDA0003848944790000146
representing the row-column difference from the current grid to the nearest free grid;
6-6) calculating an occupation weight coefficient and an idle weight coefficient of the current grid, wherein the formula is as follows:
wherein ω is occupied Occupancy weight coefficient for current grid:
ω occupied =P occupied *P distance (occupied)
ω free for the idle weight coefficient of the current grid:
ω free =P free *P distance (free)
wherein P is occupied And P free Respectively, the occupation (free) probability density, P, of the current grid distance (occupied) and P distance (free) distance probability densities of the respective current grid to the nearest occupied (free) grid;
6-7) calculating the posterior occupancy probability of the current grid, wherein the formula is as follows:
Figure BDA0003848944790000147
wherein gridNum occupied Is delta row And delta col Number of occupied grids in the range, gridNum free Is delta row And delta col The number of free grids in the range.
6-8) according to P posterior If P is a noise point, judging whether the occupied grid is a noise point posterior If < 0.3, the occupied grid is considered possibly noisy, and the occupied grid is changed to a free grid.
7) Clustering the target objects: clustering the target objects and estimating their actual distance assisted by a target detection algorithm (as shown in FIG. 5)
7-1) obtaining a target detection result by using yoloV5 algorithm, and obtaining the number Num of target objects in a three-dimensional scene object Dis, distance estimate per object i And a Center of each object in a world coordinate system i (x,y,z);
7-2) clustering objects using a modified K-Means clustering algorithm, the Center of each object having been obtained by an object detection algorithm i (x, y, z) and a k value (Num) object ) The Center of each object is measured i (x, y, z) as the initialized clustering center point, and the number Num of the target objects objec t is used as a K value, so that the defect that the K-Means clustering algorithm K is unknown is overcome, the operation speed is increased, and the formula is as follows:
Figure BDA0003848944790000151
Figure BDA0003848944790000152
wherein theta is n (x, y, z) is a three-dimensional point data set, center k Is a clustering center, k represents the number of clustering centers, α nk For the binary variable 0,1 represents that the three-dimensional point belongs to class k, 0 represents not,
Figure BDA0003848944790000157
Figure BDA0003848944790000158
when the value of j is represented, the value of theta is minimized n (x,y,z)-Cente j || 2 The three-dimensional point data set is distributed to the nearest clustering center, iterative calculation is continuously carried out until the algorithm converges to minimize loss, and clustering is finished;
8) Multi-frame fusion: optimization of dynamic objects by particle filtering in combination with multiframe information (as shown in FIG. 6)
8-1) obtaining the screened grid map model, and adding particles with different speeds into the grids in the occupied state. The particles are denoted as S, S = { S = { (S) i |s i =(x i ,y i ,vx i ,vy i ,p i ),i=1,...N S In which x i Lines representing the grid in which the particles are located, y i Columns representing the grid in which the particles are located, vx i And vy i Respectively representing the row and column velocities of the particles, p i Representing the period of the particles.
8-2) predicting the motion state of each particle according to the information of the next frame, and constructing a prediction model, wherein the formula is as follows:
the angle and distance between two adjacent frames are:
Figure BDA0003848944790000153
Figure BDA0003848944790000154
wherein
Figure BDA0003848944790000155
Is the angular velocity, v is the velocity, Δ t is the time interval between two frames; the relative movement between the grids is:
Figure BDA0003848944790000156
Figure BDA0003848944790000161
DX and Dy are resolution of grid array and row respectively, and the resolution is 0.1m in the invention; the grid position in the current frame may be denoted as x n ,y n
Figure BDA0003848944790000162
The prediction model is:
Figure BDA0003848944790000163
the delta y, the delta x, the delta vy and the delta vx are random disturbance and are derived from a Gaussian distribution with zero mean and a state transition covariance matrix of a Kalman filter.
8-3) reassign the particles in the grid by weighting and resampling:
a. calculating an occupancy probability density and a vacancy probability density for each grid
Wherein p is density (m (x, y) | occupied) represents the occupancy probability density of the grid:
Figure BDA0003848944790000164
p density (m (x, y) | free) represents the free probability density of the grid:
P density (m(x,y)|free)=1-P density (m(x,y)|occupied)
δ x and delta y Representing the lateral and longitudinal distance error of the projection of the three-dimensional point to the grid, grid (registered) representing the occupied grid, the meaning of the formula is that the current grid delta x ,δ y The ratio of the number of all occupied state grids in the range to the number of all the grids in the range is used as the occupation probability density of the current grid;
b. calculating the occupation distance probability and the idle distance probability of each grid:
the occupied distances are represented in rows and columns, respectively, as:
Figure BDA0003848944790000165
Figure BDA0003848944790000166
row and Col represent the closest Row and column occupying the grid to the current grid, respectively
The free distance is represented in rows and columns, respectively, as:
Figure BDA0003848944790000171
Figure BDA0003848944790000172
the distance probability is expressed by using a multidimensional Gaussian function:
Figure BDA0003848944790000173
c. calculating an occupancy weight and an idle weight for each grid:
ω occupied (x,y)=p density (m(x,y)|occupied)*p distance (m(x,y|occupied)
ω free (x,y)=p density (m(x,y)|free)*p distance (m(x,y)|free)
d. the occupancy posterior probability for each grid is calculated:
Figure BDA0003848944790000174
wherein N is i To occupy the actual number of particles of the grid, N max Is the maximum number of particles allowed in a grid;
a. calculating the number of resamples N per grid Ri =P i *N max And calculating the ratio of the resample amount to the actual amount
Figure BDA0003848944790000175
If f i If > 1, add f to the grid i Particles in whole number multiples of f i If the number is less than 1, deleting the particles;
8-4) judging the grid state again according to the number of particles and the particle speed:
the occupancy probability of a grid is the ratio of the number of particles of the current grid to the maximum number of particles of the grid:
Figure BDA0003848944790000176
the current grid velocity is the vector sum of the particle velocities in the current grid:
Figure BDA0003848944790000177
if P of the grid occupied If the velocity of the grid is less than the standard deviation of the velocities of all particles in the grid, the obstacle in the grid is in a static state;
9) Counting the grids in an occupied state in the two-dimensional grids, and generating a two-dimensional grid map: setting the resolution of the grid map, and visually outputting the grid map result; FIG. 7 shows the result of the grid map without any optimization strategy, and it can be seen that there are many noisy points and blocked obstacles in the graph; as shown in fig. 8, the grid map results after adding height and threshold screening and after blocking strategy, noise is reduced a little, and the number of blocked obstacles is reduced; as shown in fig. 9, the grid map result after the target detection clustering target strategy and the particle filtering strategy are added can be seen that noise is significantly less and the removal of the blocking obstacle is significant.

Claims (6)

1. The method for positioning the grid map based on the binocular stereo camera is characterized by comprising the following steps of:
1) Acquiring three-dimensional point cloud information:
1-1) respectively acquiring RGB (red, green and blue) images of a real scene and corresponding parallax images thereof through a binocular visible light camera and an infrared camera;
1-2) fusing the visible light parallax image and the infrared parallax image, wherein the fusing method comprises the following steps: comparing the confidence coefficient of each parallax in the visible light parallax image and the infrared parallax image, keeping the parallax value with large confidence coefficient in the visible light parallax image and the infrared parallax image, and finally obtaining a fusion parallax image;
1-3) generating a three-dimensional point cloud according to the fusion disparity map, and converting the three-dimensional point cloud into a world coordinate system;
2) Initializing a grid map:
2-1) determining the size of the grids, wherein the side length of each grid represents actual distance information;
2-2) determining the spatial position of the grid corresponding to the real scene: setting the range of the perceived obstacle in the z direction to include the nearest distance and the farthest distance, and the range of the obstacle in the y direction to include the highest height h max And a minimum height h min And the field angle of the obstacle in the calculation range in the x direction;
2-3) initializing grid states: idling;
3) Mapping the three-dimensional point cloud to a two-dimensional grid:
3-1) carrying out noise point filtering on the generated three-dimensional point cloud through bilateral filtering, and mapping the filtered three-dimensional point cloud onto a two-dimensional grid map plane;
3-2) counting the number of three-dimensional points mapped in the grid and the height information of the three-dimensional points;
4) Grid height and threshold screening:
4-1) obtaining the height of the three-dimensional points mapped into each grid, sorting the heights of the three-dimensional points in each grid, removing the heights of the three-dimensional points of the first 10% and the second 10% after sorting, and carrying out weighted average on the heights of the three-dimensional points after screening to obtain the final grid height h i
4-2) obtaining the number of three-dimensional points num mapped into each grid i Calculating a mapping threshold T for each grid i The formula is as follows:
Figure FDA0003848944780000011
wherein img width Is the width of the image, h i Height of the current grid, Z k The grid depth of the k-th row is theta, the field angle size of the initialization setting is theta, and alpha is a preset smoothing coefficient; formula (1) shows that the number of pixels represented by a plane formed by the grid of the k-th row and the height of the grid is used as the threshold value of the number of grids;
4-3) comparing the number of three-dimensional points num in each grid i And a mapping threshold T for each grid i If num i >T i If the grid is in the occupied state, otherwise, the grid is in the idle state;
4-4) if the grid is in the occupied state, the height h of the grid is set i If h is compared to the y-direction range of the initialization grid setting min <h i <h max Then h will be i Setting the height value of the current grid, otherwise setting the state of the current grid to be idle;
5) Screening blocking obstacles: processing occlusion by a occlusion counter;
6) Noise point elimination: removing noise interference possibly existing in the occupancy grid through the grid posterior probability;
7) Clustering the target objects: clustering target objects and estimating the actual distance of the target objects by the aid of a target detection algorithm;
8) Multi-frame fusion: optimizing a dynamic target object by combining particle filtering and multi-frame information;
9) And counting the grids in the occupied state in the two-dimensional grids to generate a two-dimensional grid map.
2. The method for positioning a grid map based on binocular stereo cameras of claim 1, wherein the grid state including idle, occupied and blocked is initialized in the step 2-3).
3. The method for positioning the grid map based on the binocular stereo camera according to claim 1, wherein the screening of the blocking obstacles in the step 5) comprises the steps of:
5-1) judging whether the grid belongs to a blocking state:
there are three cases of a blocking grid:
a. out of a set distance range or too close to the binocular camera;
b. the grid and the origin are spaced apart by one or more obstacles;
c. the height of the background obstacle in the same obstacle is lower than that of the foreground obstacle;
5-2) find the initial occupied grid, its height is noted gridh i
5-3) using polar coordinates, the occupied-state grid height gridh of the same column after the first occupied-state grid i-1 And gridh i Comparison when gridh i >gridh i-1 When the blocking value is greater than the blocking threshold β, the grid is set to the blocking state, i.e., the barrier is blocked.
4. The method for positioning a grid map based on binocular stereo cameras of claim 1, wherein the noise interference possibly existing in the occupancy grid is removed by the grid posterior probability in the step 6), comprising the steps of:
6-1) calculating the z-direction distance error of the current three-dimensional point in a world coordinate system, wherein the formula is as follows:
Figure FDA0003848944780000031
wherein b is a base line of the binocular stereo camera, f is the focal length of the camera, z is a z-direction coordinate of the three-dimensional point in a world coordinate system, and delta d A third scale factor in the disparity calculation;
6-2) calculating the x-direction distance error of the current three-dimensional point in a world coordinate system, wherein the formula is as follows:
Figure FDA0003848944780000032
wherein d is the distance from the origin of the camera to the three-dimensional point in the world coordinate system, x is the x-direction coordinate of the three-dimensional point in the world coordinate system, and delta z The distance error of the three-dimensional point in the z direction under a world coordinate system is obtained;
6-3) converting the distance error of the current three-dimensional point in the z and x directions under a world coordinate system into the relative row error of the two-dimensional grid map, wherein the formula is as follows:
Figure FDA0003848944780000033
Figure FDA0003848944780000034
wherein delta z And delta x Distance errors of the three-dimensional points in the z direction and the x direction under a world coordinate system are shown, and dz and dx are the resolutions of grids relative to the world coordinate system; 100mm in this invention;
6-4) calculating the probability density of the grid occupation within the error range of the row and the column by taking the current grid as the center, wherein the formula is as follows:
wherein P is occupied Representing the occupancy probability density of the current grid:
Figure FDA0003848944780000035
P free free probability density representing the grid:
P free =1-P occupied
δ row and delta col gridNum, the relative row-column error of the grid occupied Representing the number of grids occupying the state gridNum total Is composed of row And delta col A total number of grids within the range; the occupation probability density of the current grid is centered on the current grid row And delta col The ratio of the number of all occupied grids in the range to the number of all grids in the range; similarly, the idle probability density of the current grid is the complement of the occupancy probability density;
6-5) calculating the distance probability density from the current grid to the nearest occupied grid and the idle grid, wherein the formula is as follows:
wherein P is distance (occupied) represents the distance probability density of the current grid to the nearest occupied grid:
Figure FDA0003848944780000041
P distance (free) represents the distance probability density of the current grid to the nearest free grid:
Figure FDA0003848944780000042
wherein
Figure FDA0003848944780000043
And
Figure FDA0003848944780000044
representing the row-column difference of the current grid to the most recently occupied grid,
Figure FDA0003848944780000045
and
Figure FDA0003848944780000046
representing the row-column difference from the current grid to the nearest free grid;
6-6) calculating the occupation weight coefficient and the idle weight coefficient of the current grid, wherein the formula is as follows:
wherein ω is occupied Occupancy weight coefficient for the current grid:
ω occupied =P occupied *P distance (occupied)
ω free for the idle weight coefficient of the current grid:
ω free =P free *P distance (free)
wherein P is occupied And P free Respectively, the occupation and idle probability densities, P, of the current grid distance (occupied) and P distance (free) distance probability densities of the current grid to the nearest occupied and free grids, respectively;
6-7) calculating the posterior occupancy probability of the current grid, and the formula is as follows:
Figure FDA0003848944780000047
wherein gridNum occupied Is delta row And delta col Number of occupied grids in the range, gridNum free Is delta row And delta col The number of free grids within the range;
6-8) according to P posterior If P is a noise point, whether the occupied grid is a noise point is judged posterior Less than the threshold, the occupied grid is considered likely to be noisy, and the occupied grid is changed to a free grid.
5. The method for positioning the grid map based on the binocular stereo camera according to claim 1, wherein the step 7) of clustering the objects and estimating the actual distance thereof by the aid of an object detection algorithm comprises the steps of:
7-1) obtaining a target detection result by using a yoloV5 algorithm to obtain the number Num of target objects in a three-dimensional scene object Dis, distance estimate per object i And the Center of each target object in the world coordinate system i (x,y,z);
7-2) use of improved K-Means polymerizationClustering the target object by using a class algorithm: center of each target i (x, y, z) as the initialized clustering center point, and the number Num of the target objects object As the k value, the formula is as follows:
Figure FDA0003848944780000051
Figure FDA0003848944780000052
wherein theta is n (x, y, z) is a three-dimensional point data set, center k Is a clustering center, k represents the number of clustering centers, α nk For the binary variable 0,1 represents that the three-dimensional point belongs to class k, 0 represents not,
Figure FDA0003848944780000053
Figure FDA0003848944780000054
when the value of j is taken, the value of theta is minimized n (x,y,z)-Cente j || 2 And actually, distributing the three-dimensional point data set to the nearest clustering center, continuously iterating and calculating until the algorithm converges to minimize loss, and finishing clustering.
6. The method for positioning the grid map based on the binocular stereo camera according to claim 1, wherein the dynamic object is optimized by combining multi-frame information through particle filtering in step 8), comprising the steps of:
8-1) obtaining the screened grid map model, and adding particles with different speeds into the grid in the occupied state; the particles are denoted as S, S = { S = { (S) i |s i =(x i ,y i ,vx i ,vy i ,p i ),i=1,...N S In which x is i Lines representing the grid in which the particles are located, y i The columns of the grid in which the particles are located,vx i and vy i Representing the row and column velocities, p, of the particles, respectively i Represents the period of the particle;
8-2) predicting the motion state of each particle according to the information of the next frame; constructing a prediction model, wherein the formula is as follows:
the angle and distance between two adjacent frames are:
Figure FDA0003848944780000055
Figure FDA0003848944780000061
wherein
Figure FDA0003848944780000062
Is the angular velocity, v is the velocity, Δ t is the time interval between two frames; the relative movement between the grids is:
Figure FDA0003848944780000063
Figure FDA0003848944780000064
DX and Dy are resolution ratios of the grid lines and the columns respectively; the grid position in the current frame is denoted x n ,y n
Figure FDA0003848944780000065
The prediction model is:
Figure FDA0003848944780000066
delta y, delta x, delta vy and delta vx are random disturbance and are derived from the Gaussian distribution of zero mean and a state transition covariance matrix of a Kalman filter;
8-3) reassign the particles in the grid by weighting and resampling:
a. calculating an occupancy probability density and a vacancy probability density for each grid
Wherein p is density (m (x, y) | occupied) represents the occupancy probability density of the grid:
Figure FDA0003848944780000067
p density (m (x, y) | free) represents the free probability density of the grid:
P density (m(x,y)|free)=1-P density (m(x,y)|occupied)
δ x and delta y Representing the lateral and longitudinal distance error of the projection of the three-dimensional point to the grid, grid (registered) representing the occupied grid, the meaning of the formula is that the current grid delta x ,δ y The ratio of the number of all occupied state grids in the range to the number of all the grids in the range is used as the occupation probability density of the current grid;
b. calculating the occupation distance probability and the free distance probability of each grid:
the occupied distances are represented in rows and columns, respectively, as:
Figure FDA0003848944780000071
Figure FDA0003848944780000072
row and Col represent the closest Row and column occupying the grid to the current grid, respectively
The free distance is represented in rows and columns, respectively, as:
Figure FDA0003848944780000073
Figure FDA0003848944780000074
the distance probability is expressed by using a multidimensional Gaussian function:
Figure FDA0003848944780000075
c. calculating an occupancy weight and an idle weight for each grid:
ω occupied (x,y)=p density (m(x,y)|occupied)*p distance (m(x,y|occupied)
ω free (x,y)=p density (m(x,y)|free)*p distance (m(x,y)| free)
d. calculate the occupancy posterior probability for each grid:
Figure FDA0003848944780000076
wherein N is i To occupy the actual number of particles of the grid, N max Is the maximum number of particles allowed in a grid;
e. calculating the number of resamples N per grid Ri =P i *N max And calculating the ratio of the number of resamples to the actual number
Figure FDA0003848944780000077
If f is i If > 1, add f to the grid i Particles in integral multiples of f i If the number is less than 1, deleting the particles;
8-4) judging the grid state again through the particle number and the particle speed:
the occupancy probability of a grid is the ratio of the number of particles of the current grid to the maximum number of particles of the grid:
Figure FDA0003848944780000078
the current grid velocity is the vector sum of the particle velocities in the current grid:
Figure FDA0003848944780000079
if P of the grid occupied If the velocity of the grid is less than the standard deviation of the velocities of all the particles in the grid, the obstacle in the grid is in a static state.
CN202211127974.7A 2022-09-16 2022-09-16 Method for positioning grid map based on binocular stereo camera Pending CN115457132A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211127974.7A CN115457132A (en) 2022-09-16 2022-09-16 Method for positioning grid map based on binocular stereo camera

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211127974.7A CN115457132A (en) 2022-09-16 2022-09-16 Method for positioning grid map based on binocular stereo camera

Publications (1)

Publication Number Publication Date
CN115457132A true CN115457132A (en) 2022-12-09

Family

ID=84304646

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211127974.7A Pending CN115457132A (en) 2022-09-16 2022-09-16 Method for positioning grid map based on binocular stereo camera

Country Status (1)

Country Link
CN (1) CN115457132A (en)

Similar Documents

Publication Publication Date Title
CN110942449B (en) Vehicle detection method based on laser and vision fusion
Häne et al. Obstacle detection for self-driving cars using only monocular cameras and wheel odometry
US20180322646A1 (en) Gaussian mixture models for temporal depth fusion
CN111563415B (en) Binocular vision-based three-dimensional target detection system and method
CN110853075B (en) Visual tracking positioning method based on dense point cloud and synthetic view
WO2020215194A1 (en) Method and system for detecting moving target object, and movable platform
CN106681353A (en) Unmanned aerial vehicle (UAV) obstacle avoidance method and system based on binocular vision and optical flow fusion
Perrollaz et al. Probabilistic representation of the uncertainty of stereo-vision and application to obstacle detection
CN114708585A (en) Three-dimensional target detection method based on attention mechanism and integrating millimeter wave radar with vision
CN115032651A (en) Target detection method based on fusion of laser radar and machine vision
CN111814602B (en) Intelligent vehicle environment dynamic target detection method based on vision
CN111880191B (en) Map generation method based on multi-agent laser radar and visual information fusion
CN114325634A (en) Method for extracting passable area in high-robustness field environment based on laser radar
CN113050074B (en) Camera and laser radar calibration system and calibration method in unmanned environment perception
CN112097732A (en) Binocular camera-based three-dimensional distance measurement method, system, equipment and readable storage medium
CN114359744A (en) Depth estimation method based on fusion of laser radar and event camera
CN117274749B (en) Fused 3D target detection method based on 4D millimeter wave radar and image
Laflamme et al. Driving datasets literature review
CN112257668A (en) Main and auxiliary road judging method and device, electronic equipment and storage medium
CN117501311A (en) Systems and methods for generating and/or using three-dimensional information with one or more cameras
CN113487631A (en) Adjustable large-angle detection sensing and control method based on LEGO-LOAM
Giosan et al. Superpixel-based obstacle segmentation from dense stereo urban traffic scenarios using intensity, depth and optical flow information
CN117115690A (en) Unmanned aerial vehicle traffic target detection method and system based on deep learning and shallow feature enhancement
CN114648639B (en) Target vehicle detection method, system and device
Pfeiffer et al. Ground truth evaluation of the Stixel representation using laser scanners

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination