CN111028285A - Depth estimation method based on binocular vision and laser radar fusion - Google Patents

Depth estimation method based on binocular vision and laser radar fusion Download PDF

Info

Publication number
CN111028285A
CN111028285A CN201911221616.0A CN201911221616A CN111028285A CN 111028285 A CN111028285 A CN 111028285A CN 201911221616 A CN201911221616 A CN 201911221616A CN 111028285 A CN111028285 A CN 111028285A
Authority
CN
China
Prior art keywords
binocular
disparity map
laser radar
confidence
depth estimation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911221616.0A
Other languages
Chinese (zh)
Inventor
陈昆
沈会良
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201911221616.0A priority Critical patent/CN111028285A/en
Publication of CN111028285A publication Critical patent/CN111028285A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/593Depth or shape recovery from multiple images from stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • G06T7/85Stereo camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • G06T2207/10044Radar image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a depth estimation method based on binocular vision and laser radar fusion, which is characterized in that data collected by a laser radar and a binocular camera are registered through joint calibration; obtaining a laser radar disparity map according to the result of the combined calibration; obtaining a binocular disparity map through a binocular stereo matching algorithm, performing confidence analysis on the binocular disparity map, removing points with low confidence in the binocular disparity map, and obtaining the binocular disparity map after confidence processing; extracting and fusing the characteristics of the obtained laser radar disparity map and the binocular disparity map after the confidence coefficient processing; further feature extraction is carried out through a cascading hourglass structure, and parallax regression is carried out; the output of the front and rear cascaded hourglass structures is utilized by adopting relay supervision; and outputting an accurate and dense disparity map after fusion. The method designs a more effective network structure, and the characteristics of the laser radar disparity map and the binocular disparity map are better extracted and fused to obtain a more accurate disparity map.

Description

Depth estimation method based on binocular vision and laser radar fusion
Technical Field
The invention belongs to the field of robot and computer vision application, and particularly relates to a depth estimation method based on binocular vision and laser radar fusion.
Background
In many robotic and computer vision applications, perceiving the three-dimensional geometry of a scene or object through depth estimation is undoubtedly key to many tasks, such as autopilot, mobile robotics, positioning, obstacle avoidance, path planning, three-dimensional reconstruction, etc.
To estimate reliable depth information of a scene, two techniques may be used: depth estimation is performed using a lidar scanner or using stereo matching algorithms on binocular images. For complex outdoor scenes, lidar scanners are the most practical 3D perception solution, and the three-dimensional perception of lidar scanners can provide very accurate depth information, with errors in centimeters. However, since the lidar point cloud is sparse and occupies less than 6% of the image points, the reconstruction of a three-dimensional image by using the lidar is limited in practical application, and cannot cover all objects in a scene, and although some efforts to interpolate depth information of the sparse three-dimensional depth points exist, the performance of the lidar is also limited. On the other hand, depth estimation by binocular vision can yield dense depth information. However, due to the small baseline of the binocular camera, the limited sensing range and the inherent limitation of the stereo matching algorithm (caused by occlusion, illumination and other factors), the accuracy of the depth information obtained by binocular stereo vision is often not high.
Therefore, in order to obtain better depth estimation information, the depth information acquired by the laser radar and the depth information acquired by the binocular vision need to be fused, so that accurate and dense depth information is obtained. Attempts have also been made in the prior art based on this, however the results are not satisfactory.
The invention patent with the patent application number of CN 201810448954.7 provides a fusion method based on a low-beam laser radar and a binocular camera. And generating an error coefficient according to the data corresponding to the same object in the image data and the radar data, and generating the calibrated image data according to the error coefficient. The method is greatly influenced by environmental factors, image data acquired by a binocular camera in a complex environment can be greatly interfered, and a generated disparity map has a plurality of regions which are estimated incorrectly, so that an accurate fusion result is not obtained.
The patent with the patent application number of CN 201810575904.5 provides a network based on the existing binocular stereo matching algorithm, and the network is used for training, tuning and finally outputting an optimized disparity map by adding laser radar disparity data as a supervision, namely a system error compensation module. The fusion method belongs to front-end fusion, namely, a laser radar disparity map is added in the process of calculating the binocular disparity map to be used as a supervision item to restrict the solving process of the binocular disparity map. The method has high requirements on the binocular stereo matching algorithm, and because the supervision information is added into the binocular stereo matching algorithm, the structure of the whole network needs to be adjusted every time different binocular stereo matching algorithms are changed, which is tedious.
The patent application number of CN 201710851841.7 provides a fusion correction method of stereoscopic vision and low beam laser radar in unmanned driving. According to the method, the disparity map after semantic segmentation is obtained through semantic segmentation, and then the binocular disparity map is compensated through laser radar data to obtain the compensated full-pixel disparity map. And finally, inputting the two disparity maps as a neural network to obtain a final disparity map. However, the result of semantic segmentation is not accurate in a complex environment, and the original depth information cannot be recovered more accurately.
Disclosure of Invention
Aiming at the defects in the prior art, the invention provides a depth estimation method based on binocular vision and laser radar fusion. According to the method, a more effective network structure is designed, the characteristics of the laser radar disparity map and the binocular disparity map are better extracted and fused, and a more accurate disparity map is obtained.
A depth estimation method based on binocular vision and laser radar fusion comprises the following steps:
(1) performing binocular calibration on the camera, and performing combined calibration on the camera and the laser radar;
(2) acquiring a binocular disparity map through a binocular stereo matching algorithm according to a binocular calibration result;
(3) acquiring a laser point cloud disparity map, and performing confidence processing on the binocular disparity map, wherein the confidence processing adopts a convolutional neural network method, and the method is used for acquiring the confidence of an input disparity map end to end;
(4) respectively extracting the characteristics of the input laser point cloud disparity map and the binocular disparity map after the confidence coefficient processing, and splicing and fusing;
(5) performing further feature extraction through a cascaded hourglass structure, performing parallax regression and adding relay supervision, and fully utilizing the output result of each hourglass structure;
(6) and outputting the fused disparity map.
In the above technical solution, further, the step (1) specifically includes: the method comprises the steps of synchronously acquiring images and laser point cloud data through a binocular camera and a laser radar, carrying out camera monocular calibration and binocular calibration according to the acquired images to determine camera internal parameters and external parameters, and acquiring the external parameters between the camera and the laser radar by combining the laser point cloud data.
Further, the step (2) is specifically: and correcting the images according to the images acquired by the binocular cameras and the calibration parameters of the binocular cameras so as to align the lines of the images, and then obtaining the corresponding binocular disparity maps by using a binocular stereo matching algorithm.
Further, the step (3) is specifically: projecting the laser point cloud onto the image through a combined calibration result to obtain a corresponding laser point cloud disparity map; and removing points with the confidence coefficient less than 95% in the original binocular disparity map through a convolutional neural network, and obtaining the binocular disparity map after confidence coefficient processing.
Further, the step (4) is specifically as follows: taking the binocular disparity map and the laser point cloud disparity map after the confidence degree processing as input, performing characteristic preliminary extraction through a multilayer convolutional neural network to obtain two groups of sixteen-channel characteristic maps, and then splicing and fusing the obtained characteristic maps to obtain a thirty-two channel characteristic map; the convolution kernel adopts hole convolution, so that fewer parameters are used under the condition of the same receptive field, and the time consumption of the network is reduced; and a BN layer is added to prevent the problems of overfitting, gradient disappearance and the like.
Further, the step (5) is specifically: according to the obtained thirty-two channel characteristic diagram, further characteristic extraction is carried out through a two-layer cascade hourglass structure, finally the overall loss of the network is calculated according to the ground truth, and due to the fact that the two-layer cascade hourglass structure is formed in all, the loss calculation is carried out on the output of the two-layer hourglass structure and the ground truth respectively, and then the final loss is obtained through weighted summation, so that the shallow network can be trained fully, and the performance of the whole network is improved;
Figure BDA0002301015040000031
wherein D isF1,DF2Disparity map, D, representing the output of the first and second hourglass structuresGRepresents a disparity map corresponding to a ground channel, and p represents DGAll pixels with a median value other than null, λ1And λ2Respectively representing the corresponding weight coefficient when two layers of outputs find the loss.
Further, the step (6) is specifically: the results output using the second layer hourglass structure were used as the disparity map after final fusion at the time of testing.
The main idea of the method of the invention is as follows:
registering data acquired by a laser radar and a binocular camera through joint calibration; obtaining a laser radar disparity map according to the result of the combined calibration; obtaining a binocular disparity map through a binocular stereo matching algorithm, performing confidence analysis on the binocular disparity map, removing points with low confidence in the binocular disparity map, and obtaining the binocular disparity map after confidence processing; extracting and fusing the characteristics of the obtained laser radar disparity map and the binocular disparity map after the confidence coefficient processing; extracting deeper features through a cascaded hourglass structure and performing parallax regression; the output of the front and rear cascaded hourglass structures is utilized by adopting relay supervision; and outputting an accurate and dense disparity map after fusion.
Compared with the prior art, the invention has the following advantages:
1. the invention provides a depth estimation framework with high precision and high flexibility;
2. the invention provides a fusion method for complementing advantages of binocular vision and a laser radar, and the method removes points with low confidence in a binocular disparity map through confidence analysis, so that the influence of mismatching points on final fusion is eliminated.
3. The invention adopts a rear-end fusion method innovatively, and is different from the original front-end fusion method in that laser point cloud is used as a supervision item and added into the solving process of the binocular stereo matching algorithm.
4. According to the invention, the hourglass structure is adopted for extracting the characteristics of the parallax regression module for the first time, the strategy of firstly down-sampling and then up-sampling is adopted for the hourglass structure, and the skip level connection is added to assist up-sampling, so that the deep characteristics can be extracted, the calculated amount of network parameters is considered, the relay supervision is added into the cascaded hourglass structure, the loss output by each hourglass structure is calculated respectively, the output information of different layers of networks is fully utilized, the integration of a plurality of models is equivalent, and the robustness of the network is improved.
5. The invention trains the network under a large number of different scenes, so that the adaptability to the environment is stronger, and a better fusion result can be obtained under a complex environment.
Drawings
FIG. 1 is an overall flow diagram of the process of the present invention;
fig. 2 is an original image and a disparity image obtained in embodiment 1 of the present invention.
Detailed Description
The invention is further described with reference to the following figures and specific examples.
Fig. 1 shows a process of performing binocular vision and lidar fusion according to the present invention, which includes the following steps:
1. the method comprises the steps of synchronously acquiring calibration plate images and laser point cloud data through a binocular camera and a laser radar, carrying out camera monocular calibration and binocular calibration according to the acquired images to determine camera internal parameters and external parameters of a left camera and a right camera, and acquiring the external parameters before the cameras and the laser radar by combining the laser radar point cloud data.
1.1 in the step, the optical axes of the lenses of the binocular camera are kept parallel, the focal lengths are consistent, the positions of the camera and the laser radar are fixed, and the relative position between the camera and the laser radar is guaranteed to be unchanged; simultaneously collecting two paths of image signals and laser radar point cloud data under the control of a signal synchronization unit;
1.2 according to the monocular and binocular calibration principle of the camera, acquiring the internal reference and external reference information of the camera and the plane equation a of the calibration plate in each frame of camera coordinate systemm,iMeanwhile, a plane equation a of the calibration plate in each frame of laser radar coordinate system is obtained according to the laser point cloudl,i. Theta denotes the normal vector of the plane, X denotes the coordinates of a point on the plane, and d denotes the distance from the origin of the coordinate system to the plane.
am,im,iX+dm,i=0
al,il,iX+dl,i=0
1.3 after plane equations of the calibration plate under different coordinate systems are obtained, restraining an RT matrix by minimizing the following equations, and solving to obtain final external reference information. i represents the serial number of each group of images and laser point cloud data, and l represents the number of points on the mid-plane of each group.
Figure BDA0002301015040000051
2. And processing the image and the laser point cloud data by using the binocular calibration and the combined calibration to obtain a binocular disparity map and a laser point cloud disparity map.
2.1 in this step, first, the result of binocular calibration is used to obtain the left and right aligned images, and the corresponding disparity map is obtained according to the PSMNet end-to-end network. The PSmNet extracts feature information of left and right views through a pyramid pooling structure, then constructs a matching cost cube according to the extracted feature information, and finally performs classification regression on parallax through three cascaded hourglass structures.
2.2 after obtaining the preliminary disparity map, carrying out confidence analysis on the disparity map through the confidence analysis, and removing points with confidence lower than Thresh. The following formula is shown in detail:
Figure BDA0002301015040000052
wherein M isiRepresenting the gray value, d, of the binocular disparity map at pixel point iiAnd (4) representing the gray value of the confidence map output after the confidence analysis at the pixel point i, and Thresh representing the confidence threshold set by us.
And 2.3 projecting the laser point cloud into the image according to the camera laser radar external parameter obtained by calibration, converting the depth information of the laser point cloud into parallax information according to the relation between parallax and depth, and obtaining a laser point cloud parallax map.
Figure BDA0002301015040000053
Where B and f represent the length of the binocular camera baseline and the focal length of the camera, respectively, Z represents depth, and D represents parallax.
3. Taking the binocular disparity map and the laser point cloud disparity map obtained in the step 2 as input, performing primary feature extraction through a multilayer convolutional neural network to obtain two groups of sixteen-channel feature maps, and then splicing and fusing the obtained multi-channel feature maps to obtain a thirty-two-channel feature map; the convolution kernel adopts the cavity convolution, so that fewer parameters are used under the condition of the same receptive field, and the time consumption of the network is reduced; a BN layer is added in order to prevent problems of overfitting, disappearance of gradients etc.
4. And extracting deeper features through a two-layer cascaded hourglass structure according to the extracted features, and finally calculating the overall loss of the network according to the ground route. Because the two cascaded hourglass structures are arranged in total, the output of the two cascaded hourglass structures is respectively subjected to loss calculation with the ground route, and then the final loss is obtained through weighted summation, so that the shallow network can be trained sufficiently, and the performance of the whole network is improved.
Figure BDA0002301015040000061
Wherein D isF1,DF2Disparity map, D, representing the output of the first and second hourglass structuresGRepresents a disparity map corresponding to a ground channel, and p represents DGAll pixels with a median value other than null, λ1And λ2Respectively representing the corresponding weight coefficient when two layers of outputs find the loss.
5. The results output using the second layer hourglass structure were used as the disparity map after final fusion at the time of testing.
The method adds confidence coefficient analysis to reduce errors caused by mismatching in the binocular disparity map, and the specific working process of the confidence coefficient analysis comprises the following steps:
step 1) after obtaining a preliminary disparity map, carrying out confidence coefficient analysis on the disparity map through confidence coefficient analysis;
step 2) outputting a confidence coefficient analysis result graph;
and 3) considering that the fusion process is a complementary process, the accuracy of the binocular disparity map and the laser point cloud disparity map is ensured as much as possible. The points with the confidence coefficient lower than a certain higher threshold Thresh are removed, and the situation that pixels with overlarge errors cannot occur in the fusion process is guaranteed.
By means of a rear-end fusion method (namely, directly fusing the binocular disparity map and the laser radar disparity map), the whole system can flexibly select different binocular stereo matching algorithms without modifying the structure of a network, and the flexibility is higher.
The hourglass structure is adopted for feature extraction, deep features of the disparity map can be better extracted by the hourglass structure, so that information is richer, and the acquisition of a subsequent disparity map is facilitated; the hourglass structure adopts a strategy of first down-sampling and then up-sampling, and simultaneously skip-level connection is added to assist up-sampling, so that the deep feature of extraction is ensured, and the calculation amount of network parameters is considered (the network calculation amount can be greatly reduced by first down-sampling and then up-sampling); relay supervision is added into the cascaded hourglass structure, loss output by each hourglass structure is calculated respectively, output information of networks in different layers is fully utilized, equivalently, a plurality of models are integrated, and the robustness of the network is improved.
Example 1
The embodiment mainly measures the quality of depth estimation in a road scene, and mainly comprises the restoration of a detailed part and an overall contour. In fig. 2, (a), (b), and (c) respectively show the original pattern, the disparity map obtained by the PSMNet method, and the disparity map obtained by the fusion method of the present invention. It can be seen from the results that the contour of the person and the vehicle is restored more accurately by the method of the present invention compared with the original binocular disparity map. Meanwhile, the depth estimation of the original binocular disparity map on the far rod has obvious errors, a gap exists between the depth estimation and the far rod, and the depth information of the whole rod is well estimated by the method.
Example 2
This example evaluates the method of the invention with a Kitti2015 dataset. The Kitti2015 data set consists of 200 sets of data, including left and right views and corresponding ground truth, and corresponding lidar data can be obtained from Kitti's raw data. The data set evaluates the quality of the depth estimation by comparing the disparity map with the ground channel and calculating the error rate, wherein the lower the error rate, the better the quality of the depth estimation is. The error rate is defined as the proportion of the number of the pixels with the disparity value and the ground resolution difference of more than 3% or more than 5% in all the pixels. Specific results are shown in table 1:
TABLE 1 expression of the respective methods on Kitti2015
Figure BDA0002301015040000071
As can be seen from the table, the error rate of the PSMNet method we used on Kitti2015 dataset is 3.98%. And then, the result graphs obtained by PSmNet are fused, and the error rate of the parallax graph obtained by fusing by the method is 1.46%. It can be seen that after the binocular disparity map and the laser radar disparity map are fused by the method, the error rate of the original binocular disparity map can be reduced by about 60%.

Claims (7)

1. A depth estimation method based on binocular vision and laser radar fusion is characterized by comprising the following steps:
(1) performing binocular calibration on the camera, and performing combined calibration on the camera and the laser radar;
(2) acquiring a binocular disparity map through a binocular stereo matching algorithm according to a binocular calibration result;
(3) acquiring a laser point cloud disparity map, and performing confidence processing on the binocular disparity map, wherein the confidence processing adopts a convolutional neural network method;
(4) respectively extracting the characteristics of the input laser point cloud disparity map and the binocular disparity map after the confidence coefficient processing, and splicing and fusing;
(5) performing further feature extraction through a cascaded hourglass structure, performing parallax regression and adding relay supervision;
(6) and outputting the fused disparity map.
2. The binocular vision and lidar fusion based depth estimation method of claim 1, wherein the step (1) is specifically: the method comprises the steps of synchronously acquiring images and laser point cloud data through a binocular camera and a laser radar, carrying out camera monocular calibration and binocular calibration according to the acquired images to determine camera internal parameters and external parameters, and acquiring the external parameters between the camera and the laser radar by combining the laser point cloud data.
3. The binocular vision and lidar fusion based depth estimation method of claim 1, wherein the step (2) is specifically: and correcting the images according to the images acquired by the binocular cameras and the calibration parameters of the binocular cameras so as to align the lines of the images, and then obtaining the corresponding binocular disparity maps by using a binocular stereo matching algorithm.
4. The binocular vision and lidar fusion based depth estimation method of claim 1, wherein the step (3) is specifically: projecting the laser point cloud onto the image through a combined calibration result to obtain a corresponding laser point cloud disparity map; and removing points with the confidence coefficient less than 95% in the original binocular disparity map through a convolutional neural network, and obtaining the binocular disparity map after confidence coefficient processing.
5. The binocular vision and lidar fusion based depth estimation method of claim 1, wherein the step (4) is specifically: taking the binocular disparity map and the laser point cloud disparity map after the confidence degree processing as input, performing characteristic preliminary extraction through a multilayer convolutional neural network to obtain two groups of sixteen-channel characteristic maps, and then splicing and fusing the obtained characteristic maps to obtain a thirty-two channel characteristic map; the convolution kernel adopts hole convolution and is added with a BN layer.
6. The binocular vision and lidar fusion based depth estimation method of claim 1, wherein the step (5) is specifically: according to the obtained thirty-two channel characteristic diagram, further characteristic extraction is carried out through a two-layer cascade hourglass structure, finally the overall loss of the network is calculated according to the ground route, the loss calculation is carried out on the output of the two-layer hourglass structure and the ground route respectively, and then the final loss is obtained through weighted summation;
Figure FDA0002301015030000021
wherein D isF1,DF2Disparity map, D, representing the output of the first and second hourglass structuresGRepresents a disparity map corresponding to a ground channel, and p represents DGAll pixels with a median value other than null, λ1And λ2Respectively representing the corresponding weight coefficient when two layers of outputs find the loss.
7. The binocular vision and lidar fusion based depth estimation method of claim 1, wherein the step (6) is specifically: the results output using the second layer hourglass structure were used as the disparity map after final fusion at the time of testing.
CN201911221616.0A 2019-12-03 2019-12-03 Depth estimation method based on binocular vision and laser radar fusion Pending CN111028285A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911221616.0A CN111028285A (en) 2019-12-03 2019-12-03 Depth estimation method based on binocular vision and laser radar fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911221616.0A CN111028285A (en) 2019-12-03 2019-12-03 Depth estimation method based on binocular vision and laser radar fusion

Publications (1)

Publication Number Publication Date
CN111028285A true CN111028285A (en) 2020-04-17

Family

ID=70207849

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911221616.0A Pending CN111028285A (en) 2019-12-03 2019-12-03 Depth estimation method based on binocular vision and laser radar fusion

Country Status (1)

Country Link
CN (1) CN111028285A (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112233163A (en) * 2020-12-14 2021-01-15 中山大学 Depth estimation method and device for laser radar stereo camera fusion and medium thereof
CN112258590A (en) * 2020-12-08 2021-01-22 杭州迦智科技有限公司 Laser-based depth camera external parameter calibration method, device and storage medium thereof
CN112684424A (en) * 2020-12-30 2021-04-20 同济大学 Automatic calibration method for millimeter wave radar and camera
CN113096175A (en) * 2021-03-24 2021-07-09 苏州中科广视文化科技有限公司 Depth map confidence estimation method based on convolutional neural network
CN113096176A (en) * 2021-03-26 2021-07-09 西安交通大学 Semantic segmentation assisted binocular vision unsupervised depth estimation method
CN113110451A (en) * 2021-04-14 2021-07-13 浙江工业大学 Mobile robot obstacle avoidance method with depth camera and single line laser radar fused
CN113160298A (en) * 2021-03-31 2021-07-23 奥比中光科技集团股份有限公司 Depth truth value acquisition method, device and system and depth camera
CN113281779A (en) * 2021-05-20 2021-08-20 中山大学 3D object rapid detection method, device, equipment and medium
CN113379831A (en) * 2021-06-22 2021-09-10 北京航空航天大学青岛研究院 Augmented reality method based on binocular camera and humanoid robot
CN113625288A (en) * 2021-06-15 2021-11-09 中国科学院自动化研究所 Camera and laser radar pose calibration method and device based on point cloud registration
CN113706599A (en) * 2021-10-29 2021-11-26 纽劢科技(上海)有限公司 Binocular depth estimation method based on pseudo label fusion
CN114359891A (en) * 2021-12-08 2022-04-15 华南理工大学 Three-dimensional vehicle detection method, system, device and medium
CN115861401A (en) * 2023-02-27 2023-03-28 之江实验室 Binocular and point cloud fusion depth recovery method, device and medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107886477A (en) * 2017-09-20 2018-04-06 武汉环宇智行科技有限公司 Unmanned neutral body vision merges antidote with low line beam laser radar
US10003787B1 (en) * 2016-12-21 2018-06-19 Canon Kabushiki Kaisha Method, system and apparatus for refining a depth map
CN108537871A (en) * 2017-03-03 2018-09-14 索尼公司 Information processing equipment and information processing method
CN109978933A (en) * 2019-01-03 2019-07-05 北京中科慧眼科技有限公司 The confidence level detection method of parallax information data, device and automated driving system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10003787B1 (en) * 2016-12-21 2018-06-19 Canon Kabushiki Kaisha Method, system and apparatus for refining a depth map
CN108537871A (en) * 2017-03-03 2018-09-14 索尼公司 Information processing equipment and information processing method
CN107886477A (en) * 2017-09-20 2018-04-06 武汉环宇智行科技有限公司 Unmanned neutral body vision merges antidote with low line beam laser radar
CN109978933A (en) * 2019-01-03 2019-07-05 北京中科慧眼科技有限公司 The confidence level detection method of parallax information data, device and automated driving system

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
JIA-REN CHANG; YONG-SHENG CHEN: "Pyramid Stereo Matching Network", 《2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
KIHONG PARK; SEUNGRYONG KIM; KWANGHOON SOHN: "High-Precision Depth Estimation with the 3D LiDAR and Stereo Fusion", 《2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA)》 *
POGGI M, MATTOCCIA S: "Learning from scratch a confidence measure", 《BMVC》 *
陈少杰、朱振才等: "基于3D特征点的激光雷达与立体视觉配准方法", 《激光与光电子学进展》 *

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112258590A (en) * 2020-12-08 2021-01-22 杭州迦智科技有限公司 Laser-based depth camera external parameter calibration method, device and storage medium thereof
CN112233163A (en) * 2020-12-14 2021-01-15 中山大学 Depth estimation method and device for laser radar stereo camera fusion and medium thereof
CN112684424B (en) * 2020-12-30 2022-08-16 同济大学 Automatic calibration method for millimeter wave radar and camera
CN112684424A (en) * 2020-12-30 2021-04-20 同济大学 Automatic calibration method for millimeter wave radar and camera
CN113096175A (en) * 2021-03-24 2021-07-09 苏州中科广视文化科技有限公司 Depth map confidence estimation method based on convolutional neural network
CN113096175B (en) * 2021-03-24 2023-10-24 苏州中科广视文化科技有限公司 Depth map confidence estimation method based on convolutional neural network
CN113096176A (en) * 2021-03-26 2021-07-09 西安交通大学 Semantic segmentation assisted binocular vision unsupervised depth estimation method
CN113096176B (en) * 2021-03-26 2024-04-05 西安交通大学 Semantic segmentation-assisted binocular vision unsupervised depth estimation method
CN113160298A (en) * 2021-03-31 2021-07-23 奥比中光科技集团股份有限公司 Depth truth value acquisition method, device and system and depth camera
CN113160298B (en) * 2021-03-31 2024-03-08 奥比中光科技集团股份有限公司 Depth truth value acquisition method, device and system and depth camera
CN113110451A (en) * 2021-04-14 2021-07-13 浙江工业大学 Mobile robot obstacle avoidance method with depth camera and single line laser radar fused
CN113281779A (en) * 2021-05-20 2021-08-20 中山大学 3D object rapid detection method, device, equipment and medium
CN113625288A (en) * 2021-06-15 2021-11-09 中国科学院自动化研究所 Camera and laser radar pose calibration method and device based on point cloud registration
CN113379831A (en) * 2021-06-22 2021-09-10 北京航空航天大学青岛研究院 Augmented reality method based on binocular camera and humanoid robot
CN113379831B (en) * 2021-06-22 2022-09-09 北京航空航天大学青岛研究院 Augmented reality method based on binocular camera and humanoid robot
CN113706599A (en) * 2021-10-29 2021-11-26 纽劢科技(上海)有限公司 Binocular depth estimation method based on pseudo label fusion
CN114359891A (en) * 2021-12-08 2022-04-15 华南理工大学 Three-dimensional vehicle detection method, system, device and medium
CN114359891B (en) * 2021-12-08 2024-05-28 华南理工大学 Three-dimensional vehicle detection method, system, device and medium
CN115861401A (en) * 2023-02-27 2023-03-28 之江实验室 Binocular and point cloud fusion depth recovery method, device and medium

Similar Documents

Publication Publication Date Title
CN111028285A (en) Depth estimation method based on binocular vision and laser radar fusion
CN110942449B (en) Vehicle detection method based on laser and vision fusion
CN109685842B (en) Sparse depth densification method based on multi-scale network
CN112634341B (en) Method for constructing depth estimation model of multi-vision task cooperation
JP5926228B2 (en) Depth detection method and system for autonomous vehicles
CN112419494B (en) Obstacle detection and marking method and device for automatic driving and storage medium
CN110689562A (en) Trajectory loop detection optimization method based on generation of countermeasure network
CN111210481A (en) Depth estimation acceleration method of multiband stereo camera
CN107886477A (en) Unmanned neutral body vision merges antidote with low line beam laser radar
CN113936139A (en) Scene aerial view reconstruction method and system combining visual depth information and semantic segmentation
CN111260773A (en) Three-dimensional reconstruction method, detection method and detection system for small obstacles
CN111340922A (en) Positioning and mapping method and electronic equipment
CN104539928A (en) Three-dimensional printing image synthesizing method for optical grating
CN116258817B (en) Automatic driving digital twin scene construction method and system based on multi-view three-dimensional reconstruction
AU2021103300A4 (en) Unsupervised Monocular Depth Estimation Method Based On Multi- Scale Unification
CN111209840B (en) 3D target detection method based on multi-sensor data fusion
CN111260715A (en) Depth map processing method, small obstacle detection method and system
CN114494462A (en) Binocular camera ranging method based on Yolov5 and improved tracking algorithm
CN111105451B (en) Driving scene binocular depth estimation method for overcoming occlusion effect
KR20220014678A (en) Method and apparatus for estimating depth of images
CN113313740B (en) Disparity map and surface normal vector joint learning method based on plane continuity
CN112270701B (en) Parallax prediction method, system and storage medium based on packet distance network
CN103544732A (en) Three-dimensional reconstruction method for lunar vehicle
CN117434294A (en) Multi-aperture pure-vision optical flow velocity measurement method for unmanned aerial vehicle
CN116824433A (en) Visual-inertial navigation-radar fusion self-positioning method based on self-supervision neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200417