CN112750155B - Panoramic depth estimation method based on convolutional neural network - Google Patents

Panoramic depth estimation method based on convolutional neural network Download PDF

Info

Publication number
CN112750155B
CN112750155B CN202110053166.XA CN202110053166A CN112750155B CN 112750155 B CN112750155 B CN 112750155B CN 202110053166 A CN202110053166 A CN 202110053166A CN 112750155 B CN112750155 B CN 112750155B
Authority
CN
China
Prior art keywords
panoramic
depth
image
neural network
convolutional neural
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110053166.XA
Other languages
Chinese (zh)
Other versions
CN112750155A (en
Inventor
何炳蔚
邓清康
胡誉生
张立伟
陈彦杰
林立雄
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fuzhou University
Original Assignee
Fuzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fuzhou University filed Critical Fuzhou University
Priority to CN202110053166.XA priority Critical patent/CN112750155B/en
Publication of CN112750155A publication Critical patent/CN112750155A/en
Application granted granted Critical
Publication of CN112750155B publication Critical patent/CN112750155B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a panoramic depth estimation method based on a convolutional neural network, which comprises the following steps of: step S1, collecting RGB images, depth images and point cloud data of the outdoor environment, and splicing the RGB images and the depth images into a panoramic image according to the cylindrical projection principle; step S2, constructing a convolutional neural network model, and training a panoramic image based on the convolutional neural network model to obtain a trained convolutional neural network model; and step S3, inputting the panoramic image to be detected into the trained convolutional neural network model to obtain a dense panoramic depth prediction image. The method can adjust and optimize the local details of the panoramic image, thereby estimating the dense and accurate panoramic depth image.

Description

Panoramic depth estimation method based on convolutional neural network
Technical Field
The invention belongs to the field of image recognition and artificial intelligence, and particularly relates to a panoramic depth estimation method based on a convolutional neural network.
Background
Depth estimation is one of the basic tasks in computer vision. With the development of computer technology, deep learning has made a series of breakthrough progresses in the field of computer vision. Accurate 3D perception is always desirable because it plays a crucial role in numerous tasks of robotics and computer vision, such as automatic driving, positioning and mapping, path planning and 3D reconstruction. Various techniques have been proposed to obtain depth estimates, but none of them have some drawbacks. For example, an RGB-D camera is only suitable for short-range depth acquisition; the 3D LIDAR provides only sparse point cloud depth information; stereo cameras cannot produce reliable depth estimates in areas with uniform appearance or large variations in illumination. It is also common that the field angle of the depth estimate for any camera is limited by the size of the camera field angle.
Disclosure of Invention
In view of the above, an object of the present invention is to provide a method for estimating a panoramic depth based on a convolutional neural network, which is capable of estimating a dense and accurate panoramic depth image by fusing a panoramic image and LIDAR depth information in a cascade manner and adjusting and optimizing local details of the panoramic image according to a proposed PDCBN (panoramic depth regularization) network layer.
In order to achieve the purpose, the invention adopts the following technical scheme:
a panoramic depth estimation method based on a convolutional neural network comprises the following steps:
step S1, collecting RGB images, depth images and point cloud data of the outdoor environment, and splicing the RGB images and the depth images into a panoramic image according to the cylindrical projection principle;
step S2, constructing a convolutional neural network model, and training a panoramic image based on the convolutional neural network model to obtain a trained convolutional neural network model;
and step S3, inputting the panoramic image to be detected into the trained convolutional neural network model to obtain a dense panoramic depth prediction image.
Further, in step S1, an open city simulator cara is used to collect RGB images, depth images, and point cloud data of the outdoor environment.
Further, the step S1 is specifically:
step 1-1: loading a plurality of RGB cameras, a plurality of depth cameras and a 64-line LIDAR on a data acquisition vehicle of an open city simulator Carla, wherein the depth cameras correspond to the RGB cameras to form a 360-degree panoramic view, and controlling the data acquisition vehicle in the Carla to acquire RGB images, depth images and point cloud data in an outdoor environment;
step 1-2: based on the principle of cylindrical projection, performing cylindrical projection on each RGB image and depth image, and stitching the RGB images and the depth image into a panoramic image according to the overlapped area after the cylindrical projection;
and 1-3, cutting the stitched image into a panoramic image with a preset proportion.
Further, the step S1-2 specifically includes: setting a single image as a quadrilateral ABCD which represents a plane to be processed, and changing the single image into a curved surface EFGE1F1G1 after cylindrical projection;
assuming that the original image width is w, the height is h, and the camera view field angle is α, the camera focal length f is expressed as:
f=w/(2*tan(α/2)) (1)
the position of a certain pixel point on the image is (x, y), and the coordinates of the pixel point after cylindrical projection are (x)1,y1):
Figure BDA0002899889290000031
Figure BDA0002899889290000032
And after the image is projected in the cylindrical surface, stitching the image into a panoramic image according to the overlapped view field angle theta of the single camera and the left and right adjacent cameras.
Further, the step S2 is specifically:
step 2-1, constructing a convolutional neural network model to be trained, and taking a panoramic RGB image and a sparse panoramic depth image projected from a LIDAR as input;
and 2-2, calculating a loss function loss value by using a back propagation algorithm, reducing errors through iterative calculation, and performing parameter learning to enable a predicted value to approach a true value so as to obtain a trained convolutional neural network model.
Further, each layer of the convolutional neural network model sequentially performs operations of convolution, panorama condition regularization, activation and pooling, and a ReLU activation function is used:
Figure BDA0002899889290000033
let the input fused feature size be C × H × W, and the minimum batch λ ═ Fi]The panoramic depth condition regularization network layer is defined as:
Figure BDA0002899889290000034
wherein ε is a numerical stability small constant, αi,c,h,wi,c,h,wAre learnable parameters.
Setting LIDAR depth information as a function
Figure BDA0002899889290000041
Since the new parameters generated depend on LIDAR depth information, the network layer is named PDCBN;
Figure BDA0002899889290000042
different pixel values are arranged at different positions, and the mapping should be carried out according to pixels in sequence in the mapping process; setting scAnd ycFunction:
Figure BDA0002899889290000043
Figure BDA0002899889290000044
at function scAnd ycGiven a point in the panoramic LIDAR depth image
Figure BDA0002899889290000045
If the point exists in the LIDAR projection of the point into the panoramic image, the point is considered valid, enhanced or suppressed by the PDCBN network layer
Figure BDA0002899889290000046
Depth value of (d); otherwise, the point is an invalid point, and normal BN network layer processing is used.
Further, the convolutional neural network model uses a depth true value dtruAnd the model to pixel predicted value dpreTraining the model according to the absolute error of the LIDAR, and averaging the loss on the effective pixel N according to the true value of the sparse depth of the LIDAR, wherein a loss function is defined as:
Figure BDA0002899889290000047
compared with the prior art, the invention has the following beneficial effects:
the invention can adjust and optimize the local details of the panoramic image according to the proposed PDCBN (panoramic depth condition regularization) network layer by cascading and fusing the panoramic image and LIDAR depth information, thereby estimating the dense and accurate panoramic depth image.
Drawings
FIG. 1 is a schematic overall flow diagram of an embodiment of the present invention;
FIG. 2 is a schematic view of the principle of cylindrical projection in the present invention;
FIG. 3 is a schematic view of a cylindrical projection of the present invention: (a) is a top view of the cylindrical projection, (b) is a side view of the cylindrical projection;
fig. 4 is an example of stitching effect of a panoramic image after cylindrical projection in the present invention: (a) is a panoramic RGB image, (b) is a true panoramic depth image;
FIG. 5 is a schematic diagram of the convolutional neural network structure of the present invention;
fig. 6 is an example of a qualitative result of panoramic depth estimation according to the present invention, which is a panoramic RGB image, a sparse panoramic depth image projected from a LIDAR, an estimated panoramic depth image, a real panoramic depth image, and a panoramic depth error image in sequence from top to bottom.
Detailed Description
The invention is further explained by the following embodiments in conjunction with the drawings.
Referring to fig. 1, the present invention provides a panoramic depth estimation method based on a convolutional neural network, which includes the following steps:
step S1: collecting RGB images, depth images and point cloud data of an outdoor environment through an open city simulator Carla, splicing the RGB images and the depth images into panoramic images according to a cylindrical projection principle, and dividing a generated data set into a training set and a testing set (90/10%);
step S2: constructing a convolutional neural network model to be trained, taking a panoramic RGB image and a sparse panoramic depth image projected from LIDAR as input, calculating a loss function loss value by using a back propagation algorithm, reducing errors through iterative computation, and performing parameter learning to enable a predicted value to approach a true value, thereby obtaining an optimal weight model of the convolutional neural network;
step S3: and loading the weight model trained in the step S2, and inputting the divided panoramic test set into a convolutional neural network model for panoramic depth estimation to obtain a dense panoramic depth prediction image.
In this embodiment, the specific processing procedure of step S1 is as follows:
1) the method is characterized in that 5 RGB cameras, 5 depth cameras and a 64-line LIDAR are mounted on a data acquisition vehicle of an open city simulator Carla, the depth cameras correspond to the RGB cameras, the field angle is 90 degrees, and each camera is installed in a rotating mode of 72 degrees to form a 360-degree panoramic view. Then, operating a data acquisition vehicle in Carla, and acquiring RGB images, depth images and point cloud data in an outdoor environment;
2) fig. 2 is a schematic diagram illustrating the principle of cylindrical projection of an image. The image plane is a quadrilateral ABCD and represents the plane to be processed, and after cylindrical projection, becomes a curved surface EFGE1F1G 1. Suppose thatWhen a certain pixel point on the image is located at (x, y), the pixel point after the cylindrical projection is located at (x)1,y1);
Fig. 3(a) is a top view of the principle of cylindrical projection, which can deduce the projection change process of the x value of the pixel:
Figure BDA0002899889290000061
Figure BDA0002899889290000062
Figure BDA0002899889290000063
wherein w represents the width of the image, i.e. the length of the line segment AB, theta represents the included angle between Ox and the line segment Ow, f represents the focal length of the camera, alpha represents the angle of the field of view of the camera, the size is 90 DEG, and w1Representing the width, x, of the image after cylindrical projective transformation1Indicating the post x-cylinder projection position.
The transformation can be obtained by substituting the formulas (1) and (2) into the formula (3), and the position x after the projection transformation of the point x is obtained1
Figure BDA0002899889290000071
Fig. 3(b) is a side view of the principle of cylindrical projection, which can deduce the projection change process of the y value of the pixel point, and can deduce from the similarity of the triangle:
Figure BDA0002899889290000072
where h represents the height of the image, i.e., the length of the line segment BC, f represents the focal length of the camera, θ represents the angle between Ox and the line segment Ow, h represents the distance between Ox and the line segment Ow1Representing the height, y, of the image after cylindrical projective transformation1Represents the y columnThe position of the surface after projection. And there are:
Figure BDA0002899889290000073
therefore, the position y after projective transformation of the point y is obtained by substituting the formula (6) into the formula (5)1
Figure BDA0002899889290000074
The image cylindrical projection is performed in a specific calculation process, and after the image cylindrical projection is performed, a panoramic image is stitched according to the condition that the overlapped view field angle of a single camera and the left and right adjacent cameras is 9 degrees. A panoramic image stitching example effect is shown in fig. 4.
3) The stitched image is cropped to a 256 x 3840 panoramic image, forming an integrated data set for the outdoor environment. Of these, 90% are the training set and 10% are the test set.
Fig. 5 shows the convolutional neural network structure of the present invention, and step 2 is to construct the convolutional neural network model to be trained according to the proposed convolutional neural network structure.
In this embodiment, the convolutional neural network for panoramic depth estimation is an end-to-end depth learning framework that takes a panoramic color image and a sparse panoramic depth image projected from a LIDAR as inputs and outputs a dense panoramic depth image.
The entire network is mainly composed of two paths: and the coarse-scale network path and the PDCBN network optimization path. And a coarse-scale network path firstly fuses the input panoramic color image and the sparse LIDAR depth image to realize the prediction of the panoramic depth of the scene on the global level. Then, the PDCBN network optimization path adaptively adjusts and optimizes local details according to the LIDAR depth value.
Thus, the neural network model can compile global predictions and incorporate finer details, thereby generating dense and accurate panoramic depth images. The specific processing process of the network comprises the following steps:
1) the network structure cascade fuses the input panoramic RGB image and the sparse panoramic depth image projected from the LIDAR.
2) Each layer network performs convolution, panorama conditional regularization (PDCBN), activation, pooling operations in turn, using the ReLU activation function:
Figure BDA0002899889290000081
let the input fused feature size be C × H × W, and the minimum batch λ ═ Fi]. At this time, the proposed panorama depth conditional regularization network layer is defined as:
Figure BDA0002899889290000082
where ε is a numerical stability small constant, αi,c,h,wi,c,h,wAre learnable parameters. When model training is applied to the PDCBN, new parameters α, β are generated by mapping the LIDAR depth information. Setting LIDAR depth information as a function
Figure BDA0002899889290000091
Since the new parameters generated depend on the LIDAR depth information, the network layer is named PDCBN.
Figure BDA0002899889290000092
The mapping method has different pixel values at different positions, so that the mapping should be performed in sequence according to pixels in the mapping process; and the LIDAR depth information is discontinuous, a policy needs to be set to handle the depth information contained in the sparse depth image. By setting scAnd ycThe function solves the above problem:
Figure BDA0002899889290000093
Figure BDA0002899889290000094
at function scAnd ycGiven a point in the panoramic LIDAR depth image
Figure BDA0002899889290000095
If the point exists in the LIDAR projection of the point into the panoramic image, the point is considered valid, enhanced or suppressed by the PDCBN network layer
Figure BDA0002899889290000096
Depth value of (d); otherwise, the point is an invalid point, and normal BN network layer processing is used. This completes the pixel-wise sequential mapping of the panoramic LIDAR depth, thereby adaptively adjusting and optimizing the local depth values.
3) And setting a loss function in the constructed convolutional neural network model to be trained. Using true values of depth dtruAnd the model to pixel predicted value dpreThe absolute error of the model. The loss over the effective pixel N is averaged according to the true value of the sparse depth of the LIDAR. The loss function is defined as:
Figure BDA0002899889290000097
step 3 is to input the test data into the weight model trained in step 2, the resolution of the input panoramic image is 256 × 3840, the output dense panoramic depth image has good quality of depth reconstruction in the scene with recovered details and deeper, and the completed panoramic depth image is shown in fig. 6.
Preferably, in this embodiment, the setting of an evaluation index for the panoramic depth image obtained by prediction specifically includes: root Mean Square Error (RMSE), Mean Absolute Error (MAE), root mean square error of inverse depth (iRMSE) and mean absolute error of inverse depth (iMAE), where RMSE is the most important indicator.
Figure BDA0002899889290000101
Figure BDA0002899889290000102
Figure BDA0002899889290000103
Figure BDA0002899889290000104
And (4) carrying out error evaluation on the tested panoramic depth image according to the four evaluation indexes, wherein the experimental evaluation error result is shown in table 1.
Table 1: error assessment
Figure BDA0002899889290000105
Preferably, in this embodiment, the operating system for training the neural network model is ubuntu16.04, the graphics card model is NVIDIA Tesla M40, and pytorch1.0, deep learning framework and python3.5 programming languages are used. And (3) performing 20 times of loop training in the actual training process, selecting a RMSProp iterative optimizer, setting a parameter alpha to be 0.9, and setting the learning rate to be 0.001.
The above description is only a preferred embodiment of the present invention, and all equivalent changes and modifications made in accordance with the claims of the present invention should be covered by the present invention.

Claims (6)

1. A panoramic depth estimation method based on a convolutional neural network is characterized by comprising the following steps:
step S1, collecting RGB images, depth images and LIDAR projected sparse panoramic depth images of the outdoor environment, and splicing the RGB images and the depth images into panoramic images according to the cylindrical projection principle;
step S2, constructing a convolutional neural network model, and training a panoramic image based on the convolutional neural network model to obtain a trained convolutional neural network model;
step S3, inputting the panoramic image to be tested into the trained convolutional neural network model to obtain a dense panoramic depth prediction image;
each layer of the convolutional neural network model sequentially executes operations of convolution, panorama condition regularization, activation and pooling, and a ReLU activation function is used:
Figure FDA0003568012400000011
let the input panoramic image size be C × H × W, and the minimum batch λ be [ F [ ]i]The panoramic depth condition regularization network layer is defined as:
Figure FDA0003568012400000012
wherein ε is a numerical stability small constant, αi,c,h,wi,c,h,wIs a learnable parameter;
setting LIDAR depth information as a function
Figure FDA0003568012400000013
Since the new parameters generated depend on LIDAR depth information, the network layer is named PDCBN;
Figure FDA0003568012400000014
different pixel values are arranged at different positions, and the mapping should be carried out according to pixels in sequence in the mapping process; setting scAnd ycFunction:
Figure FDA0003568012400000021
Figure FDA0003568012400000022
at function scAnd ycGiven a point in a sparse panoramic depth image of a LIDAR projection
Figure FDA0003568012400000023
If the point is present from the LIDAR projection into the panoramic image, the point is deemed valid, enhanced or suppressed by the PDCBN network layer
Figure FDA0003568012400000024
Depth value of (d); otherwise, the point is an invalid point, and normal BN network layer processing is used.
2. The convolutional neural network-based panoramic depth estimation method of claim 1, wherein the step S1 adopts an open city simulator cara to acquire RGB images, depth images, and sparse panoramic depth images projected by LIDAR of outdoor environment.
3. The panoramic depth estimation method based on the convolutional neural network as claimed in claim 2, wherein the step S1 specifically comprises:
step 1-1: loading a plurality of RGB cameras, a plurality of depth cameras and a 64-line LIDAR on a data acquisition vehicle of an open city simulator Carla, wherein the depth cameras correspond to the RGB cameras to form a 360-degree panoramic view, controlling the data acquisition vehicle in the Carla, and acquiring RGB images, depth images and sparse panoramic depth images projected by the LIDAR in an outdoor environment;
step 1-2: based on the principle of cylindrical projection, carrying out cylindrical projection on each RGB image and depth image, and splicing into panoramic images according to the overlapped areas after cylindrical projection;
and 1-3, cutting the spliced image into a panoramic image with a preset proportion.
4. The panoramic depth estimation method based on the convolutional neural network as claimed in claim 3, wherein the step 1-2 is specifically: setting a single image as a quadrilateral ABCD which represents a plane to be processed, and changing the single image into a curved surface EFGE1F1G1 after cylindrical projection;
assuming that the original image width is w, the height is h, and the camera view field angle is α, the camera focal length f is expressed as:
f=w/(2*tan(α/2)) (1)
the position of a certain pixel point on the image is (x, y), and the coordinate of the pixel point after the cylindrical projection is (x)1,y1):
Figure FDA0003568012400000031
Figure FDA0003568012400000032
After the image is projected on the cylindrical surface, the panoramic image is spliced according to the coincidence view field angle theta of the single camera and the left and right adjacent cameras.
5. The panoramic depth estimation method based on the convolutional neural network of claim 1, wherein the step S2 specifically comprises:
step 2-1, constructing a convolutional neural network model to be trained, and taking a panoramic image and a sparse panoramic depth image projected from a LIDAR as input;
and 2-2, calculating a loss function loss value by using a back propagation algorithm, reducing errors through iterative calculation, and performing parameter learning to enable a predicted value to approach a true value so as to obtain a trained convolutional neural network model.
6. The convolutional neural network-based panoramic depth estimation method of claim 5, wherein the convolutional neural network model uses a true depth value dtruAnd the model to pixel predicted value dpreAbsolute error ofTraining the model and averaging the loss on the effective pixel N according to the LIDAR sparse depth true value
The loss function is defined as:
Figure FDA0003568012400000041
CN202110053166.XA 2021-01-15 2021-01-15 Panoramic depth estimation method based on convolutional neural network Active CN112750155B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110053166.XA CN112750155B (en) 2021-01-15 2021-01-15 Panoramic depth estimation method based on convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110053166.XA CN112750155B (en) 2021-01-15 2021-01-15 Panoramic depth estimation method based on convolutional neural network

Publications (2)

Publication Number Publication Date
CN112750155A CN112750155A (en) 2021-05-04
CN112750155B true CN112750155B (en) 2022-07-01

Family

ID=75652024

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110053166.XA Active CN112750155B (en) 2021-01-15 2021-01-15 Panoramic depth estimation method based on convolutional neural network

Country Status (1)

Country Link
CN (1) CN112750155B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113592932A (en) * 2021-06-28 2021-11-02 北京百度网讯科技有限公司 Training method and device for deep completion network, electronic equipment and storage medium
CN116866723B (en) * 2023-09-04 2023-12-26 广东力创信息技术有限公司 Pipeline safety real-time monitoring and early warning system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105678721A (en) * 2014-11-20 2016-06-15 深圳英飞拓科技股份有限公司 Method and device for smoothing seams of panoramic stitched image
CN106981080A (en) * 2017-02-24 2017-07-25 东华大学 Night unmanned vehicle scene depth method of estimation based on infrared image and radar data
CN107767456A (en) * 2017-09-22 2018-03-06 福州大学 A kind of object dimensional method for reconstructing based on RGB D cameras
CN109215067A (en) * 2017-07-03 2019-01-15 百度(美国)有限责任公司 High-resolution 3-D point cloud is generated based on CNN and CRF model
CN110910437A (en) * 2019-11-07 2020-03-24 大连理工大学 Depth prediction method for complex indoor scene

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10474160B2 (en) * 2017-07-03 2019-11-12 Baidu Usa Llc High resolution 3D point clouds generation from downsampled low resolution LIDAR 3D point clouds and camera images
WO2020140047A1 (en) * 2018-12-28 2020-07-02 Nvidia Corporation Distance to obstacle detection in autonomous machine applications

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105678721A (en) * 2014-11-20 2016-06-15 深圳英飞拓科技股份有限公司 Method and device for smoothing seams of panoramic stitched image
CN106981080A (en) * 2017-02-24 2017-07-25 东华大学 Night unmanned vehicle scene depth method of estimation based on infrared image and radar data
CN109215067A (en) * 2017-07-03 2019-01-15 百度(美国)有限责任公司 High-resolution 3-D point cloud is generated based on CNN and CRF model
CN107767456A (en) * 2017-09-22 2018-03-06 福州大学 A kind of object dimensional method for reconstructing based on RGB D cameras
CN110910437A (en) * 2019-11-07 2020-03-24 大连理工大学 Depth prediction method for complex indoor scene

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Convolutional Neural Network-Based Regression for Depth Prediction in Digital Holography;Tomoyoshi Shimobaba et al.;《2018 IEEE 27th International Symposium on Industrial Electronics (ISIE)》;20180615;全文 *
基于深度学习的多视图三维重建;韩婧;《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》;20200615;全文 *
面向铁路线路视频的三维场景理解;吴洁;《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》;20160715;全文 *

Also Published As

Publication number Publication date
CN112750155A (en) 2021-05-04

Similar Documents

Publication Publication Date Title
WO2020207166A1 (en) Object detection method and apparatus, electronic device, and storage medium
US11514644B2 (en) Automated roof surface measurement from combined aerial LiDAR data and imagery
CN111325794A (en) Visual simultaneous localization and map construction method based on depth convolution self-encoder
CN104330074B (en) Intelligent surveying and mapping platform and realizing method thereof
CN108734120A (en) Mark method, apparatus, equipment and the computer readable storage medium of image
CN103985133B (en) Search method and system for optimal splicing lines among images based on graph-cut energy optimization
CN113674416B (en) Three-dimensional map construction method and device, electronic equipment and storage medium
CN113330486A (en) Depth estimation
CN112750155B (en) Panoramic depth estimation method based on convolutional neural network
CN112505065A (en) Method for detecting surface defects of large part by indoor unmanned aerial vehicle
CN109255833A (en) Based on semantic priori and the wide baseline densification method for reconstructing three-dimensional scene of gradual optimization
CN113393577B (en) Oblique photography terrain reconstruction method
CN115953535A (en) Three-dimensional reconstruction method and device, computing equipment and storage medium
CN110260866A (en) A kind of robot localization and barrier-avoiding method of view-based access control model sensor
Hu et al. Deep-learning assisted high-resolution binocular stereo depth reconstruction
CN116797742A (en) Three-dimensional reconstruction method and system for indoor scene
CN112862736A (en) Real-time three-dimensional reconstruction and optimization method based on points
CN116563493A (en) Model training method based on three-dimensional reconstruction, three-dimensional reconstruction method and device
Hu et al. Geometric feature enhanced line segment extraction from large-scale point clouds with hierarchical topological optimization
CN116385660A (en) Indoor single view scene semantic reconstruction method and system
Mihajlovic et al. Deepsurfels: Learning online appearance fusion
Chen et al. I2D-Loc: Camera localization via image to lidar depth flow
Zeng et al. Surface reconstruction by propagating 3d stereo data in multiple 2d images
CN117456136A (en) Digital twin scene intelligent generation method based on multi-mode visual recognition
Cao et al. Quantifying visual environment by semantic segmentation using deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant