WO2023155353A1 - Depth image acquisition method and apparatus, and depth system, terminal and storage medium - Google Patents

Depth image acquisition method and apparatus, and depth system, terminal and storage medium Download PDF

Info

Publication number
WO2023155353A1
WO2023155353A1 PCT/CN2022/100593 CN2022100593W WO2023155353A1 WO 2023155353 A1 WO2023155353 A1 WO 2023155353A1 CN 2022100593 W CN2022100593 W CN 2022100593W WO 2023155353 A1 WO2023155353 A1 WO 2023155353A1
Authority
WO
WIPO (PCT)
Prior art keywords
depth image
initial
target
image
features
Prior art date
Application number
PCT/CN2022/100593
Other languages
French (fr)
Chinese (zh)
Inventor
杨晓立
余宇山
赵鑫
Original Assignee
奥比中光科技集团股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 奥比中光科技集团股份有限公司 filed Critical 奥比中光科技集团股份有限公司
Publication of WO2023155353A1 publication Critical patent/WO2023155353A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds

Definitions

  • the present application belongs to the technical field of image processing, and in particular relates to a depth image acquisition method, device, depth system, terminal and storage medium.
  • depth estimation has become a popular research application field.
  • Commonly used depth perception methods such as structured light, TOF, binocular, lidar, etc., after years of development, have become increasingly mature in technology and have been widely used in many fields.
  • these methods are limited by the cost and technology itself, and can only obtain reliable sparse depth point clouds, or low-resolution depth maps. Therefore, in recent years, deep completion technology based on neural network has received extensive attention.
  • Embodiments of the present application provide a depth image acquisition method, device, depth system, terminal, and storage medium, which can improve the reliability of acquired dense depth images.
  • the first aspect of the embodiment of the present application provides a method for acquiring a depth image, including:
  • the sparse depth image and the initial hidden feature perform at least one iterative optimization operation on the initial dense depth image, and confirm the target hidden feature according to the hidden features to be confirmed obtained by each iterative optimization operation ;
  • Depth estimation is performed using the target hidden features to obtain a target dense depth image of the target scene.
  • An image acquisition unit configured to acquire a color image and a sparse depth image of the target scene
  • an initial densification unit configured to extract color features and initial features of the color image, and obtain an initial dense depth image and initial hidden features according to the color features, the initial features, and the sparse depth image;
  • An iterative optimization unit configured to perform at least one iterative optimization operation on the initial dense depth image by using the color feature, the sparse depth image, and the initial hidden feature, and obtain the information to be confirmed according to each iterative optimization operation
  • Hidden features confirm target hidden features
  • the target densification unit is configured to use the hidden features of the target to perform depth estimation to obtain a target dense depth image of the target scene.
  • the third aspect of the embodiment of the present application provides a depth system, including a color module, a depth module, and the acquisition device described in the second aspect of the application, wherein:
  • the color module is used to collect a color image of a target scene
  • the depth module is used to scan the target scene to obtain point cloud data, and obtain a sparse depth image according to the point cloud data;
  • the acquisition device uses the color image and the sparse depth image to obtain a target dense depth image.
  • the fourth aspect of the embodiments of the present application provides a terminal, including a memory, a processor, and a computer program stored in the memory and operable on the processor, and the above method is implemented when the processor executes the computer program A step of.
  • a fifth aspect of the embodiments of the present application provides a computer-readable storage medium, where the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the steps of the foregoing method are implemented.
  • the sixth aspect of the embodiments of the present application provides a computer program product, which, when the computer program product runs on a terminal, enables the terminal to execute the steps of the method.
  • each iteration The optimization operation needs to refer to the sparse depth image and color features to determine the hidden features to be confirmed, that is, the process of each iterative optimization operation will be guided by the information of the RGB image and the information of the sparse depth image, which can improve the obtained dense depth image. reliability.
  • FIG. 1 is a schematic diagram of an implementation flow of a depth image acquisition method provided by an embodiment of the present application
  • FIG. 2 is a schematic diagram of a specific implementation process for determining an initial hidden feature provided by an embodiment of the present application
  • FIG. 3 is a schematic structural diagram of a depth image model provided by an embodiment of the present application.
  • Fig. 4 is a schematic structural diagram of a feedback module provided by an embodiment of the present application.
  • FIG. 5 is a schematic diagram of a specific implementation process of training a depth image model provided by an embodiment of the present application
  • FIG. 6 is a schematic structural diagram of a device for acquiring a depth image provided by an embodiment of the present application.
  • FIG. 7 is a schematic structural diagram of a terminal provided by an embodiment of the present application.
  • the existing depth completion technology generally guides the scene information of the scene through the RGB color image to realize the densification of the sparse depth image. Some methods are to directly stitch the RGB and sparse depth images, and then input the stitched images into the neural network for depth completion; other methods are to input the sparse depth images to the neural network to obtain the initial dense depth image, and then The dense depth map is fused with the RGB image to obtain a more accurate and more accurate depth completion result.
  • Fig. 1 shows a schematic flow chart of a method for acquiring a depth image provided by an embodiment of the present application.
  • the method can be applied to a terminal, and is applicable to situations where the reliability of an acquired dense depth image needs to be improved.
  • the above-mentioned terminal may be a device capable of image processing such as a computer, a smart phone, and a tablet device.
  • the above method for acquiring a depth image may include the following steps S101 to S104.
  • Step S101 acquiring a color image and a sparse depth image of a target scene.
  • the terminal can acquire the color image and point cloud data of the target scene, and then project the point cloud data onto the imaging plane of the color image to obtain a sparse depth image.
  • the terminal can obtain the color image obtained by shooting the target scene through the color camera, obtain the point cloud data obtained by scanning the target scene with the depth sensor, and project the point cloud data obtained by the depth sensor to the imaging of the color camera plane, resulting in a sparse depth image.
  • the depth sensor can include but not limited to lidar, direct time of flight (Direct Time of flight, dTof), speckle indirect measurement of time of flight (Indirect Time of flight, iTof), etc.
  • Step S102 extracting color features and initial features of the color image, and obtaining an initial dense depth image and initial hidden features according to the color features, initial features, and sparse depth image.
  • the terminal may extract its color feature and initial feature through a feature extraction algorithm; wherein, the feature extraction algorithms used for the color feature and the initial feature may be the same or different.
  • the terminal may perform a convolution operation on a color image through different convolution kernels to extract color features and initial features of the color image.
  • the terminal can use the color features and initial features acquired from the color image, as well as the sparse depth image to determine the initial hidden features for optimization feedback.
  • the initial hidden features can be used to perform feature optimization feedback on the initial dense depth image, thereby preliminarily realizing the completion of the depth image.
  • the initial hidden features in step S102 can be obtained from steps S201 to S203.
  • Step S201 performing depth estimation on initial features to obtain an initial dense depth image.
  • the terminal may perform convolution regression on the initial features to obtain an initial dense depth image. It should be noted that, the foregoing depth estimation manner may be selected according to actual conditions, and no limitation is set here.
  • step S202 the initial dense depth image and the sparse depth image are fused to obtain an initial fused feature image.
  • d sparse is the sparse depth image
  • d dense is the initial dense depth image
  • Step S203 using the initial fusion feature image, initial features and color features to determine initial hidden features.
  • the terminal may perform a convolution operation on the initial fusion feature image, the initial feature, and the color feature to determine the initial hidden feature.
  • Step S103 using the color feature, the sparse depth image and the initial hidden feature, to perform at least one iterative optimization operation on the initial dense depth image, and confirm the target hidden feature according to the hidden features to be confirmed obtained by each iterative optimization operation.
  • the color features, the initial hidden features and the sparse depth image are used as the input of the first iterative optimization operation, and the hidden features to be confirmed obtained by the first iterative optimization operation will be compared with Color features and sparse depth images are used as input for the second iterative optimization operation, and so on.
  • Each iterative optimization operation can obtain a hidden feature to be confirmed.
  • the target hidden feature used to determine the target dense depth image can be determined from the hidden features to be confirmed obtained by each iterative optimization operation. feature.
  • the target dense depth image is the depth image whose densification effect can meet the requirements.
  • the hidden features to be confirmed output by the first iterative optimization operation in step S103 can be obtained by the following steps:
  • Step S204 performing depth estimation on the initial hidden features to obtain a first dense depth image
  • Step S205 fusing the first dense depth image and the sparse depth image to obtain a first fused feature image
  • Step S206 using the first fused feature image, color features and initial hidden features to determine hidden features to be confirmed output by the first iterative optimization operation.
  • the terminal can perform depth estimation on the hidden features to be confirmed output by the first iterative optimization operation to obtain the second dense depth image, and then fuse the second dense depth image with the sparse depth image , to obtain the second fused feature image, and use the second fused feature image, color features, and hidden features to be confirmed output by the first iterative optimization operation to determine the hidden features to be confirmed output by the second iterative optimization operation.
  • N is a positive integer greater than or equal to 1.
  • step S204 to step S206 are similar to those of step S201 to step S203 respectively, and will not be repeated here.
  • N the higher the value of N, the better the densification effect of the target hidden feature output at the end of the iterative optimization operation.
  • the time consumed and the amount of calculation will increase.
  • the densification effect of the first few iterations is the best.
  • the specific value of N can be set according to actual conditions such as hardware conditions and densification requirements.
  • the terminal may calculate the difference between the hidden feature to be confirmed output by the current iterative optimization operation and the hidden feature to be confirmed output by the previous iterative optimization operation Between the error indicators, and determine whether the error indicators are within the preset error threshold range. If the error index is within the preset error threshold range, the iterative optimization operation is stopped, and the hidden feature to be confirmed obtained by the current iterative optimization operation is used as the target hidden feature. If the error index is outside the preset error threshold range, the hidden feature to be confirmed output by the current iterative optimization operation is used as the input of the next iterative optimization operation, and the next iterative optimization operation is continued.
  • the range of the error threshold can be adjusted according to the actual situation, which is not limited in this application.
  • the terminal can judge whether the error index between the hidden features to be confirmed and the initial hidden features output by the first iterative optimization operation is within the range of the error threshold, so as to determine whether The next iterative optimization operation is required. If the error index is within the range of the error threshold, the iterative optimization operation is stopped, and the hidden feature to be confirmed output by the first iterative optimization operation is used as the target hidden feature.
  • the iterative optimization operation is stopped, and the first The hidden features to be confirmed output by the N iteration optimization operation are used as the target hidden features.
  • the error indicator may also be determined based on the dense depth image corresponding to the current iterative optimization operation and the dense depth image corresponding to the previous iterative optimization operation of the current iterative optimization operation.
  • the dense depth image corresponding to the current iterative optimization operation refers to the dense depth image obtained by performing depth estimation on hidden features output by the current iterative optimization operation.
  • the terminal may subtract the dense depth image corresponding to the previous iterative optimization operation from the dense depth image obtained by the current iterative optimization operation, and then calculate a mean absolute error (Mean Absolute Error, MAE) value as an error indicator.
  • MAE mean Absolute Error
  • step S104 depth estimation is performed using hidden features of the target to obtain a target dense depth image of the target scene.
  • the terminal performs depth estimation on the target hidden features confirmed in step S103 to obtain the target dense depth image of the target scene.
  • the dense depth image is an image obtained by performing depth complementation on a sparse depth image, that is, a depth image whose densification effect can meet requirements.
  • each iteration The optimization operation needs to refer to the sparse depth image and color features to determine the hidden features to be confirmed, that is, the process of each iterative optimization operation will be guided by the information of the RGB image and the information of the sparse depth image, which can improve the obtained dense depth image. reliability.
  • each iterative optimization operation will use the hidden features to be confirmed output by the previous iterative optimization operation as a guide, so the densification degree of the dense depth image will be further improved after each iterative optimization operation.
  • Fig. 3 shows a schematic diagram of the structure of a depth image model.
  • the terminal can input the color image and the sparse depth image into the depth image model, and obtain the target dense depth image output by the depth image model.
  • the depth image model may include a feature extraction module, N feedback modules and a target depth estimation module.
  • the terminal can extract the color features and initial features of the color image through the feature extraction module, and obtain the initial dense depth image and initial hidden features through the first feedback module.
  • the terminal can perform an iterative optimization operation through the remaining feedback modules in turn, and finally through the target depth estimation module, perform depth estimation on the hidden features of the target output by the last feedback module to obtain the target dense depth image of the target scene.
  • each feedback module may include an intermediate depth estimation module, a fusion module and a sequence model module.
  • the step of the terminal performing an iterative optimization operation in a single feedback module may specifically include: performing depth estimation on the previous hidden feature (that is, the hidden feature to be confirmed output by the previous feedback module) through the intermediate depth estimation module of the current feedback module, and obtaining The dense depth image output by the current feedback module; through the fusion module of the current feedback module, the dense depth image output by the current feedback module is fused with the sparse depth image to obtain the fusion feature map of the current feedback module; through the sequence model module of the current feedback module , use the color feature, the fusion feature map of the current feedback module, and the previous hidden feature to determine the current hidden feature to be confirmed (that is, the hidden feature to be confirmed output by the current feedback module).
  • the above number of feedback modules can be set according to actual conditions.
  • the above-mentioned depth image model can also output a parameter used to characterize the densification effect. If the parameter is greater than a preset threshold, continue to input the output of the current feedback module into the next feedback module. If the parameter is smaller than the threshold, The hidden feature to be confirmed output by the current feedback module is used as the target hidden feature, and the target hidden feature is estimated in depth through the target depth estimation module, and the target dense depth image of the target scene is output.
  • the terminal Before using the depth image model, the terminal needs to train the depth image model. Furthermore, in the process of training the depth image model, the number of iterations of the iterative optimization operation through the feedback module is fixed. If the number of iterations is not fixed, the dense depth image obtained by the iterative optimization of the feedback module will follow the adjustment of the network parameters to be trained. And change, so that there are two variables in the training process, and the accurate training error cannot be obtained; while in the process of using the depth image model, the number of iterations of the feedback module may not be fixed, and the number of iterations may depend on the current iterative optimization operation. The error between the dense depth image to be optimized and the dense depth image to be optimized obtained by the previous iterative optimization operation.
  • the above-mentioned training process of the depth image model may include steps S501 to S503.
  • Step S501 acquiring a sample color image, a sample sparse depth image and a corresponding reference dense depth image.
  • step S101 for the acquisition manner of the sample color image and the sample sparse depth image, please refer to the description of step S101.
  • the reference dense depth image is an ideal dense depth image.
  • artificially synthesized depth images can be obtained, for example, it can be realized by the Unreal 4 (unreal engine 4, UE4) engine, or the depth collected by other depth sensors (such as high-precision TOF depth cameras) can also be obtained image.
  • Unreal 4 unreal engine 4, UE4
  • other depth sensors such as high-precision TOF depth cameras
  • Step S502 input the sample color image and sample sparse depth image into the network to be trained, obtain the sample dense depth image output by each feedback module in the network to be trained, and the sample target dense depth image output by the target depth estimation module in the network to be trained .
  • model structure and working process of the network to be trained can refer to the descriptions in FIG. 1 to FIG. 4 , which will not be repeated in this application.
  • Step S503 calculate the target error value according to the sample target dense depth image, each sample dense depth image, and the reference dense depth image, if the target error value is greater than the error threshold, adjust the parameters of the network to be trained to iteratively optimize the network to be trained, Until the target error value is less than or equal to the error threshold, the network to be trained is used as a depth image model.
  • the error threshold refers to the maximum value of the target error value allowed when the model converges, which can be adjusted according to the actual situation.
  • the terminal may calculate the initial error value between the sample target dense depth image and each sample dense depth image and the reference dense depth image, and then perform a weighted average on the initial error value, Get the target error value, and then ensure that the densification effect of iterative optimization is better.
  • the target error value is greater than the error threshold, it means that the network to be trained has not converged, so it is necessary to readjust the parameters of the network to be trained, and recalculate the target error value, and iterate until the target error value is less than or It is equal to the error threshold, indicating that the network to be trained has been able to output a reliable dense depth image, and the network can be used as a depth image model and put into use.
  • the training process can be implemented by using the gradient descent method, and the corresponding loss function (loss function) can be L1 norm loss function, L2 norm loss function or other loss functions.
  • FIG. 6 is a schematic structural diagram of a depth image acquisition apparatus 600 provided in an embodiment of the present application, and the depth image acquisition apparatus 600 is configured on a terminal.
  • the acquisition device 600 of the depth image may include:
  • An image acquisition unit 601 configured to acquire a color image and a sparse depth image of a target scene
  • An initial densification unit 602 configured to extract color features and initial features of the color image, and obtain an initial dense depth image and initial hidden features according to the color features, the initial features, and the sparse depth image;
  • An iterative optimization unit 603, configured to perform at least one iterative optimization operation on the initial dense depth image by using the color feature, the sparse depth image and the initial hidden feature, and obtain Confirm hidden features Confirm target hidden features;
  • the target densification unit 604 is configured to use the hidden features of the target to perform depth estimation to obtain a dense target depth image of the target scene.
  • the aforementioned depth image acquisition device 600 may include the aforementioned depth image model, please refer to FIG. 3 , the initial densification unit 602 may correspond to the feature extraction module and the first feedback module of the depth image model, and the iterative optimization unit 603 may correspond to the depth image For other feedback modules other than the first feedback module in the model, the object densification unit 604 may correspond to the object depth estimation module of the depth image model.
  • the initial densification unit 602 may be specifically configured to: perform depth estimation on the initial features to obtain an initial dense depth image; fuse the initial dense depth image with the sparse depth image , to obtain an initial fusion feature image; using the initial fusion feature image, the initial feature, and the color feature to determine the initial hidden feature.
  • the above-mentioned image acquisition unit 601 may be specifically configured to: acquire the color image and point cloud data of the target scene; project the point cloud data onto the imaging plane of the color image to obtain The sparse depth image.
  • the above-mentioned iterative optimization unit 603 may be specifically configured to: use the initial hidden features to perform at least one iterative optimization operation on the initial dense depth image, and calculate the current The error index between the hidden feature to be confirmed output by the iterative optimization operation and the hidden feature to be confirmed output by the previous iterative optimization operation, if the error index is outside the error threshold range, continue to the next iterative optimization operation until the If the error index is within the range of the error threshold, the iterative optimization operation is stopped, and the hidden feature to be confirmed output by the current iterative optimization operation is used as the target hidden feature.
  • the depth image acquisition apparatus 600 may further include a training unit, which may be used to: acquire a sample color image, a sample sparse depth image, and a corresponding reference dense depth image;
  • the sample sparse depth image is input into the network to be trained, and the sample dense depth image output by each feedback module in the network to be trained is obtained, and the sample target dense depth image output by the target depth estimation module in the network to be trained; according to calculating a target error value for the sample target dense depth image, each of the sample dense depth images, and the reference dense depth image, and adjusting the parameters of the network to be trained if the target error value is greater than an error threshold,
  • the network to be trained is optimized iteratively until the target error value is less than or equal to the error threshold, and the network to be trained is used as the depth image model.
  • the above-mentioned training unit may be specifically configured to: extract the color features and initial features of the color image through the feature extraction module in the network to be trained; A feedback module, which obtains an initial dense depth image and an initial hidden feature according to the color feature, the initial feature and the sparse depth image; performs an iterative optimization operation through other feedback modules in the network to be trained, and outputs each A sample dense depth image obtained by an iterative optimization operation.
  • the above training unit may be specifically configured to: calculate the initial error value between the sample target dense depth image and each of the sample dense depth images and the reference dense depth image; The initial error value is weighted and averaged to obtain the target error value.
  • the specific working process of the depth image acquisition apparatus 600 can refer to the corresponding process of the methods described in FIG. 1 to FIG. 5 , which will not be repeated here.
  • the embodiment of the present application also provides a depth system.
  • the system specifically includes a color module, a depth module, and the aforementioned depth image acquisition device 600, wherein the color module is used to collect a color image of the target scene; the depth module is used to The target scene is scanned to obtain point cloud data, and a sparse depth image is obtained according to the point cloud data; the obtaining device uses the color image and the sparse depth image to obtain a dense depth image of the target.
  • the color module includes a color camera
  • the depth module includes but is not limited to a laser radar, a direct time of flight (Direct Time of flight, dTof) camera, and a speckle indirect time of flight (Indirect Time of flight, iTof) camera.
  • the color module, the depth module and the acquisition device can be an integrated device or an independent device, and the data between each component can be transmitted by wire or wireless, which is not limited here.
  • the data between each component can be transmitted by wire or wireless, which is not limited here.
  • FIG. 7 it is a schematic diagram of a terminal provided in the embodiment of the present application.
  • the terminal 7 may include: a processor 70, a memory 71, and a computer program 72 stored in the memory 71 and operable on the processor 70, such as a depth image acquisition program.
  • the processor 70 executes the computer program 72, it implements the steps in the embodiments of the methods for acquiring depth images above, such as steps S101 to S104 shown in FIG. 1 .
  • the processor 70 executes the computer program 72, it realizes the functions of the modules/units in the above-mentioned device embodiments, such as the image acquisition unit 601, the initial densification unit 602, the iterative optimization unit 603 and the Target densification unit 604 .
  • the computer program can be divided into one or more modules/units, and the one or more modules/units are stored in the memory 71 and executed by the processor 70 to complete the present application.
  • the one or more modules/units may be a series of computer program instruction segments capable of accomplishing specific functions, and the instruction segments are used to describe the execution process of the computer program in the terminal.
  • the computer program can be divided into: an image acquisition unit, an initial densification unit, an iterative optimization unit and a target densification unit.
  • the image acquisition unit is used to acquire the color image and the sparse depth image of the target scene;
  • the initial densification unit is used to extract the color features and initial features of the color image, and according to the color features, the initial features and the
  • the sparse depth image acquires an initial dense depth image and initial hidden features;
  • an iterative optimization unit is configured to perform at least one iterative optimization operation on the initial dense depth image by using the color features, the sparse depth image, and the initial hidden features , and confirm the target hidden feature according to the hidden features to be confirmed obtained by each iterative optimization operation;
  • the target densification unit is configured to use the target hidden feature to perform depth estimation to obtain a target dense depth image of the target scene.
  • the terminal may include, but not limited to, a processor 70 and a memory 71 .
  • a processor 70 and a memory 71 .
  • FIG. 7 is only an example of a terminal, and does not constitute a limitation on the terminal. It may include more or less components than those shown in the figure, or combine certain components, or different components, such as the Terminals may also include input and output devices, network access devices, buses, and so on.
  • the so-called processor 70 can be a central processing unit (Central Processing Unit, CPU), and can also be other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Off-the-shelf programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • a general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
  • the storage 71 may be an internal storage unit of the terminal, such as a hard disk or memory of the terminal.
  • the memory 71 can also be an external storage device of the terminal, such as a plug-in hard disk equipped on the terminal, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, a flash memory card (Flash Card) etc.
  • the memory 71 may also include both an internal storage unit of the terminal and an external storage device.
  • the memory 71 is used to store the computer program and other programs and data required by the terminal.
  • the memory 71 can also be used to temporarily store data that has been output or will be output.
  • the disclosed device/terminal and method may be implemented in other ways.
  • the device/terminal embodiments described above are only illustrative.
  • the division of the modules or units is only a logical function division.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of devices or units may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components shown as units may or may not be physical units, that is, they may be located in one place, or may be distributed to multiple network units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units can be implemented in the form of hardware or in the form of software functional units.
  • the integrated module/unit is realized in the form of a software function unit and sold or used as an independent product, it can be stored in a computer-readable storage medium. Based on this understanding, all or part of the processes in the methods of the above embodiments in the present application can also be completed by instructing related hardware through computer programs.
  • the computer programs can be stored in a computer-readable storage medium, and the computer When the program is executed by the processor, the steps in the above-mentioned various method embodiments can be realized.
  • the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file or some intermediate form.
  • the computer-readable medium may include: any entity or device capable of carrying the computer program code, a recording medium, a USB flash drive, a removable hard disk, a magnetic disk, an optical disk, a computer memory, and a read-only memory (Read-Only Memory, ROM) , random access memory (Random Access Memory, RAM), electric carrier signal, telecommunication signal and software distribution medium, etc.
  • ROM Read-Only Memory
  • RAM Random Access Memory
  • electric carrier signal telecommunication signal and software distribution medium, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The present application is applicable to the technical field of image processing. Provided are a depth image acquisition method and apparatus, and a depth system, a terminal and a storage medium. The depth image acquisition method specifically comprises: acquiring a color image and a sparse depth image of a target scene; extracting a color feature and an initial feature of the color image, and acquiring an initial dense depth image and an initial hidden feature on the basis of the color feature, the initial feature and the sparse depth image; performing at least one iterative optimization operation, so as to confirm a target hidden feature on the basis of a hidden feature to be confirmed that is acquired by means of each iterative optimization operation; and performing depth estimation using the target hidden feature, so as to obtain a target dense depth image of the target scene. The embodiments of the present application can improve the reliability of an acquired dense depth image.

Description

深度图像的获取方法、装置、深度系统、终端和存储介质Depth image acquisition method, device, depth system, terminal and storage medium
本申请要求于2022年2月16日提交中国专利局,申请号为202210142004.8,发明名称为“深度图像的获取方法、装置、深度系统、终端和存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with the application number 202210142004.8 submitted to the China Patent Office on February 16, 2022. The contents are incorporated by reference in this application.
技术领域technical field
本申请属于图像处理技术领域,尤其涉及一种深度图像的获取方法、装置、深度系统、终端和存储介质。The present application belongs to the technical field of image processing, and in particular relates to a depth image acquisition method, device, depth system, terminal and storage medium.
背景技术Background technique
近年来,随着计算机视觉技术在自动驾驶,机器人,AR应用等领域的发展,深度估计已经成为了一个热门的研究应用领域。常用的深度感知方法,比如结构光、TOF、双目、激光雷达等,经过多年的发展,在技术上已经日趋成熟,在多个领域上得到了广泛的应用。然而这些方法受限制于成本与技术本身,只能获取到可靠的稀疏深度点云,或低分辨率的深度图。因此近年来,基于神经网络的深度补全技术得到了广泛的关注。In recent years, with the development of computer vision technology in the fields of autonomous driving, robotics, and AR applications, depth estimation has become a popular research application field. Commonly used depth perception methods, such as structured light, TOF, binocular, lidar, etc., after years of development, have become increasingly mature in technology and have been widely used in many fields. However, these methods are limited by the cost and technology itself, and can only obtain reliable sparse depth point clouds, or low-resolution depth maps. Therefore, in recent years, deep completion technology based on neural network has received extensive attention.
目前如何有效的融合RGB图和稀疏深度图仍然是一个开放的问题。而现有的深度补全技术,往往没有很好地利用稀疏深度图像,得到的稠密深度图像可靠性不足。How to effectively fuse RGB images and sparse depth images is still an open problem. However, the existing depth completion techniques often do not make good use of sparse depth images, and the resulting dense depth images are not reliable enough.
发明内容Contents of the invention
本申请实施例提供一种深度图像的获取方法、装置、深度系统、终端和存储介质,可以提高所获取到的稠密深度图像的可靠性。Embodiments of the present application provide a depth image acquisition method, device, depth system, terminal, and storage medium, which can improve the reliability of acquired dense depth images.
本申请实施例第一方面提供一种深度图像的获取方法,包括:The first aspect of the embodiment of the present application provides a method for acquiring a depth image, including:
获取目标场景的彩色图像和稀疏深度图像;Obtain a color image and a sparse depth image of the target scene;
提取所述彩色图像的彩色特征和初始特征,并根据所述彩色特征、所述初始特征及所述稀疏深度图像获取初始稠密深度图像及初始隐藏特征;extracting color features and initial features of the color image, and obtaining an initial dense depth image and initial hidden features according to the color features, the initial features, and the sparse depth image;
利用所述彩色特征、所述稀疏深度图像和所述初始隐藏特征,对所述初始稠密深度图像进行至少一次迭代优化操作,并根据每次迭代优化操作获取到的待确认隐藏特征确认目标隐藏特征;Using the color feature, the sparse depth image and the initial hidden feature, perform at least one iterative optimization operation on the initial dense depth image, and confirm the target hidden feature according to the hidden features to be confirmed obtained by each iterative optimization operation ;
利用所述目标隐藏特征进行深度估计,得到所述目标场景的目标稠密深度图像。Depth estimation is performed using the target hidden features to obtain a target dense depth image of the target scene.
本申请实施例第二方面提供的一种深度图像的获取装置,包括:A device for acquiring a depth image provided in the second aspect of the embodiment of the present application includes:
图像获取单元,用于获取目标场景的彩色图像和稀疏深度图像;An image acquisition unit, configured to acquire a color image and a sparse depth image of the target scene;
初始稠密化单元,用于提取所述彩色图像的彩色特征和初始特征,并根据所述彩色特征、所述初始特征及所述稀疏深度图像获取初始稠密深度图像及初始隐藏特征;an initial densification unit, configured to extract color features and initial features of the color image, and obtain an initial dense depth image and initial hidden features according to the color features, the initial features, and the sparse depth image;
迭代优化单元,用于利用所述彩色特征、所述稀疏深度图像和所述初始隐藏特征,对所述初始稠密深度图像进行至少一次迭代优化操作,并根据每次迭代优化操作获取到的待确认隐藏特征确认目标隐藏特征;An iterative optimization unit, configured to perform at least one iterative optimization operation on the initial dense depth image by using the color feature, the sparse depth image, and the initial hidden feature, and obtain the information to be confirmed according to each iterative optimization operation Hidden features confirm target hidden features;
目标稠密化单元,用于利用所述目标隐藏特征进行深度估计,得到所述目标场景的目标稠密深度图像。The target densification unit is configured to use the hidden features of the target to perform depth estimation to obtain a target dense depth image of the target scene.
本申请实施例第三方面提供一种深度系统,包括彩色模块、深度模块及本申请第二方面所述的获取装置,其中:The third aspect of the embodiment of the present application provides a depth system, including a color module, a depth module, and the acquisition device described in the second aspect of the application, wherein:
所述彩色模块,用于采集目标场景的彩色图像;The color module is used to collect a color image of a target scene;
所述深度模块,用于对所述目标场景进行扫描得到点云数据,并根据所述点云数据得到稀疏深度图像;The depth module is used to scan the target scene to obtain point cloud data, and obtain a sparse depth image according to the point cloud data;
所述获取装置,利用所述彩色图像及所述稀疏深度图像得到目标稠密深度图像。The acquisition device uses the color image and the sparse depth image to obtain a target dense depth image.
本申请实施例第四方面提供一种终端,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现上述方法的步骤。The fourth aspect of the embodiments of the present application provides a terminal, including a memory, a processor, and a computer program stored in the memory and operable on the processor, and the above method is implemented when the processor executes the computer program A step of.
本申请实施例第五方面提供一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现上述方法的步骤。A fifth aspect of the embodiments of the present application provides a computer-readable storage medium, where the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the steps of the foregoing method are implemented.
本申请实施例第六方面提供了一种计算机程序产品,当计算机程序产品在终端上运行时,使得终端执行时实现方法的步骤。The sixth aspect of the embodiments of the present application provides a computer program product, which, when the computer program product runs on a terminal, enables the terminal to execute the steps of the method.
在本申请的实施方式中,通过获取目标场景的彩色图像和稀疏深度图像,提取彩色图像的彩色特征和初始特征,根据彩色特征、初始特征及稀疏深度图像获取初始稠密深度图像及初始隐藏特征,并进行至少一次的迭代优化操作,以根据每次迭代优化操作获取到的待确认隐藏特征确认目标隐藏特征,并利用目标隐藏特征进行深度估计,得到目标场景的目标稠密深度图像,由于每次迭代优化操作需参考稀疏深度图像和彩色特征确定待确认隐藏特征,也即每次迭代优化操作的过程均会通过RGB图像的信息和稀疏深度图像的信息进行引导,能够提高所获取到的稠密深度图像的可靠性。In the embodiment of the present application, by acquiring the color image and sparse depth image of the target scene, extracting the color features and initial features of the color image, and obtaining the initial dense depth image and initial hidden features according to the color features, initial features and sparse depth image, And perform at least one iterative optimization operation to confirm the hidden features of the target according to the hidden features to be confirmed obtained by each iterative optimization operation, and use the hidden features of the target for depth estimation to obtain the target dense depth image of the target scene. Since each iteration The optimization operation needs to refer to the sparse depth image and color features to determine the hidden features to be confirmed, that is, the process of each iterative optimization operation will be guided by the information of the RGB image and the information of the sparse depth image, which can improve the obtained dense depth image. reliability.
附图说明Description of drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present application, the accompanying drawings that need to be used in the descriptions of the embodiments or the prior art will be briefly introduced below. Obviously, the accompanying drawings in the following description are only for the present application For some embodiments, those of ordinary skill in the art can also obtain other drawings based on these drawings without paying creative efforts.
图1是本申请实施例提供的一种深度图像的获取方法的实现流程示意图;FIG. 1 is a schematic diagram of an implementation flow of a depth image acquisition method provided by an embodiment of the present application;
图2是本申请实施例提供的确定初始隐藏特征的具体实现流程示意图;FIG. 2 is a schematic diagram of a specific implementation process for determining an initial hidden feature provided by an embodiment of the present application;
图3是本申请实施例提供的深度图像模型的结构示意图;FIG. 3 is a schematic structural diagram of a depth image model provided by an embodiment of the present application;
图4是本申请实施例提供的反馈模块的结构示意图;Fig. 4 is a schematic structural diagram of a feedback module provided by an embodiment of the present application;
图5是本申请实施例提供的训练深度图像模型的具体实现流程示意图;FIG. 5 is a schematic diagram of a specific implementation process of training a depth image model provided by an embodiment of the present application;
图6是本申请实施例提供的一种深度图像的获取装置的结构示意图;FIG. 6 is a schematic structural diagram of a device for acquiring a depth image provided by an embodiment of the present application;
图7是本申请实施例提供的终端的结构示意图。FIG. 7 is a schematic structural diagram of a terminal provided by an embodiment of the present application.
具体实施方式Detailed ways
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。基于本申请的实施例,本领域技术人员在没有做出创造性劳动的前提下所获得的所有其他实施例,都属于本申请保护。In order to make the purpose, technical solution and advantages of the present application clearer, the present application will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present application, and are not intended to limit the present application. Based on the embodiments of the present application, all other embodiments obtained by those skilled in the art without creative efforts shall belong to the protection of the present application.
现有的深度补全技术,一般是通过RGB彩色图像对场景的场景信息进行引导,实现稀疏深度图像的稠密化。一些方法是直接将RGB和稀疏深度图进行拼接操作,再将拼接好的图片输入到神经网络中进行深度补全;另外一些方法则是将稀疏深度图输入至神经网络获取初始稠密深度图,再将稠密深度图与RGB图像进行融合得到精确更高的深度补全结果。The existing depth completion technology generally guides the scene information of the scene through the RGB color image to realize the densification of the sparse depth image. Some methods are to directly stitch the RGB and sparse depth images, and then input the stitched images into the neural network for depth completion; other methods are to input the sparse depth images to the neural network to obtain the initial dense depth image, and then The dense depth map is fused with the RGB image to obtain a more accurate and more accurate depth completion result.
但是,这些方法往往没有很好地利用稀疏深度图像,得到的稠密深度图像稠密化程度不足,可靠性较低。However, these methods often do not make good use of sparse depth images, and the obtained dense depth images are not densified enough and have low reliability.
为了说明本申请的技术方案,下面通过具体实施例来进行说明。In order to illustrate the technical solution of the present application, specific examples are used below to illustrate.
图1示出了本申请实施例提供的一种深度图像的获取方法的实现流程示意图,该方法可以应用于终端上,可适用于需提高所获取到的稠密深度图像的可靠性的情形。Fig. 1 shows a schematic flow chart of a method for acquiring a depth image provided by an embodiment of the present application. The method can be applied to a terminal, and is applicable to situations where the reliability of an acquired dense depth image needs to be improved.
其中,上述终端可以是计算机、智能手机、平板设备等能够进行图像处理的设备。Wherein, the above-mentioned terminal may be a device capable of image processing such as a computer, a smart phone, and a tablet device.
具体的,上述深度图像的获取方法可以包括以下步骤S101至步骤S104。Specifically, the above method for acquiring a depth image may include the following steps S101 to S104.
步骤S101,获取目标场景的彩色图像和稀疏深度图像。Step S101, acquiring a color image and a sparse depth image of a target scene.
在本申请的一些实施方式中,终端可以获取目标场景的彩色图像和点云数 据,然后将点云数据投影至彩色图像的成像平面上,得到稀疏深度图像。In some embodiments of the present application, the terminal can acquire the color image and point cloud data of the target scene, and then project the point cloud data onto the imaging plane of the color image to obtain a sparse depth image.
更具体地,终端可以获取通过彩色相机对目标场景进行拍摄得到的彩色图像,并获取深度传感器对目标场景进行扫描得到的点云数据,将深度传感器扫描得到的点云数据投影至彩色相机的成像平面上,从而得到稀疏深度图像。其中,深度传感器可以包括但不限于激光雷达,直接测量飞行时间(Direct Time of flight,dTof),散斑间接测量飞行时间(Indirect Time of flight,iTof)等。More specifically, the terminal can obtain the color image obtained by shooting the target scene through the color camera, obtain the point cloud data obtained by scanning the target scene with the depth sensor, and project the point cloud data obtained by the depth sensor to the imaging of the color camera plane, resulting in a sparse depth image. Among them, the depth sensor can include but not limited to lidar, direct time of flight (Direct Time of flight, dTof), speckle indirect measurement of time of flight (Indirect Time of flight, iTof), etc.
需要说明的是,上述目标场景可以根据实际情况进行选择,本申请对此不作限制。It should be noted that the above target scenario may be selected according to actual conditions, which is not limited in this application.
步骤S102,提取彩色图像的彩色特征和初始特征,并根据彩色特征、初始特征及稀疏深度图像获取初始稠密深度图像及初始隐藏特征。Step S102, extracting color features and initial features of the color image, and obtaining an initial dense depth image and initial hidden features according to the color features, initial features, and sparse depth image.
在本申请的实施方式中,终端可以通过特征提取算法提取其彩色特征和初始特征;其中,彩色特征和初始特征所使用的特征提取算法可以相同,也可以不同。例如,终端可以通过不同的卷积核对彩色图像进行卷积操作,以提取彩色图像的彩色特征和初始特征。In the embodiment of the present application, the terminal may extract its color feature and initial feature through a feature extraction algorithm; wherein, the feature extraction algorithms used for the color feature and the initial feature may be the same or different. For example, the terminal may perform a convolution operation on a color image through different convolution kernels to extract color features and initial features of the color image.
进一步地,终端可以利用从彩色图像获取到的彩色特征和初始特征,以及稀疏深度图像确定用于优化反馈的初始隐藏特征。其中,初始隐藏特征可用于对初始稠密深度图像进行特征优化反馈,从而初步实现深度图像的补全。Further, the terminal can use the color features and initial features acquired from the color image, as well as the sparse depth image to determine the initial hidden features for optimization feedback. Among them, the initial hidden features can be used to perform feature optimization feedback on the initial dense depth image, thereby preliminarily realizing the completion of the depth image.
在一个实施例中,如图2所示,步骤S102中初始隐藏特征可由步骤S201至S203得到。In one embodiment, as shown in FIG. 2, the initial hidden features in step S102 can be obtained from steps S201 to S203.
步骤S201,对初始特征进行深度估计,得到初始稠密深度图像。Step S201, performing depth estimation on initial features to obtain an initial dense depth image.
在一些实施方式中,终端可以在初始特征上进行卷积回归,得到初始稠密深度图像。需要说明的是,上述深度估计方式可以根据实际情况进行选择,此处不作限制。In some implementation manners, the terminal may perform convolution regression on the initial features to obtain an initial dense depth image. It should be noted that, the foregoing depth estimation manner may be selected according to actual conditions, and no limitation is set here.
步骤S202,将初始稠密深度图像与稀疏深度图像进行融合,得到初始融合特征图像。In step S202, the initial dense depth image and the sparse depth image are fused to obtain an initial fused feature image.
具体的,终端可以通过拼接(Concat)操作实现图像的融合,初始融合特 征图像F concat(d sparse,d dense)=Concat{d sparse,d dense}。 Specifically, the terminal can realize image fusion through a concatenation (Concat) operation, and the initial fusion feature image F concat (d sparse ,d dense )=Concat{d sparse ,d dense }.
其中,d sparse为稀疏深度图像,d dense为初始稠密深度图像。 Among them, d sparse is the sparse depth image, and d dense is the initial dense depth image.
步骤S203,利用初始融合特征图像、初始特征和彩色特征,确定初始隐藏特征。Step S203, using the initial fusion feature image, initial features and color features to determine initial hidden features.
具体的,终端可以对初始融合特征图像、初始特征和彩色特征进行卷积操作确定初始隐藏特征。Specifically, the terminal may perform a convolution operation on the initial fusion feature image, the initial feature, and the color feature to determine the initial hidden feature.
步骤S103,利用彩色特征、稀疏深度图像和初始隐藏特征,对初始稠密深度图像进行至少一次迭代优化操作,并根据每次迭代优化操作分别获取到的待确认隐藏特征确认目标隐藏特征。Step S103, using the color feature, the sparse depth image and the initial hidden feature, to perform at least one iterative optimization operation on the initial dense depth image, and confirm the target hidden feature according to the hidden features to be confirmed obtained by each iterative optimization operation.
在本申请的实施方式中,在得到初始隐藏特征之后,将彩色特征、初始隐藏特征及稀疏深度图像作为第一次迭代优化操作的输入,第一次迭代优化操作得到的待确认隐藏特征将与彩色特征及稀疏深度图像作为第二次迭代优化操作的输入,以此类推。每次迭代优化操作均可得到一待确认隐藏特征,经过至少一次的迭代优化操作之后,可以从每次迭代优化操作得到的待确认隐藏特征中,确定出用于确定目标稠密深度图像的目标隐藏特征。其中,目标稠密深度图像也即稠密化效果能够满足需求的深度图像。In the embodiment of the present application, after obtaining the initial hidden features, the color features, the initial hidden features and the sparse depth image are used as the input of the first iterative optimization operation, and the hidden features to be confirmed obtained by the first iterative optimization operation will be compared with Color features and sparse depth images are used as input for the second iterative optimization operation, and so on. Each iterative optimization operation can obtain a hidden feature to be confirmed. After at least one iterative optimization operation, the target hidden feature used to determine the target dense depth image can be determined from the hidden features to be confirmed obtained by each iterative optimization operation. feature. Wherein, the target dense depth image is the depth image whose densification effect can meet the requirements.
在一些实施例中,步骤S103中第一次迭代优化操作输出的待确认隐藏特征可由以下步骤得到:In some embodiments, the hidden features to be confirmed output by the first iterative optimization operation in step S103 can be obtained by the following steps:
步骤S204,对初始隐藏特征进行深度估计,得到第一稠密深度图像;Step S204, performing depth estimation on the initial hidden features to obtain a first dense depth image;
步骤S205,将第一稠密深度图像与稀疏深度图像进行融合,得到第一融合特征图像;Step S205, fusing the first dense depth image and the sparse depth image to obtain a first fused feature image;
步骤S206,利用第一融合特征图像、彩色特征及初始隐藏特征,确定第一次迭代优化操作输出的待确认隐藏特征。Step S206, using the first fused feature image, color features and initial hidden features to determine hidden features to be confirmed output by the first iterative optimization operation.
在进行第二次迭代优化操作时,终端可以对第一次迭代优化操作输出的待确认隐藏特征进行深度估计,得到第二稠密深度图像,然后,将第二稠密深度图像与稀疏深度图像进行融合,得到第二融合特征图像,并利用第二融合特征 图像、彩色特征及第一次迭代优化操作输出的待确认隐藏特征,确定第二次迭代优化操作输出的待确认隐藏特征。依次类推,迭代N次,即可获取N个待确认隐藏特征,其中N为大于或等于1的正整数。When performing the second iterative optimization operation, the terminal can perform depth estimation on the hidden features to be confirmed output by the first iterative optimization operation to obtain the second dense depth image, and then fuse the second dense depth image with the sparse depth image , to obtain the second fused feature image, and use the second fused feature image, color features, and hidden features to be confirmed output by the first iterative optimization operation to determine the hidden features to be confirmed output by the second iterative optimization operation. By analogy and so on, iterating N times, N hidden features to be confirmed can be obtained, where N is a positive integer greater than or equal to 1.
需要说明的是,步骤S204至步骤S206的具体方法分别与步骤S201至步骤S203相似,此处不再赘述。It should be noted that the specific methods of step S204 to step S206 are similar to those of step S201 to step S203 respectively, and will not be repeated here.
此外,N的取值越大,迭代优化操作最后输出的目标隐藏特征稠密化效果越好,相应的,所需消耗的时间和计算量会增加,一般地,最初的几次迭代稠密化效果最为明显,因此,N的具体取值可以根据硬件条件、稠密化需求等实际情况进行设置。In addition, the larger the value of N, the better the densification effect of the target hidden feature output at the end of the iterative optimization operation. Correspondingly, the time consumed and the amount of calculation will increase. Generally, the densification effect of the first few iterations is the best. Obviously, therefore, the specific value of N can be set according to actual conditions such as hardware conditions and densification requirements.
在本申请的一些实施方式中,在每次通过迭代优化操作获取到待确认隐藏特征后,终端可以计算当前迭代优化操作输出的待确认隐藏特征与前一次迭代优化操作输出的待确认隐藏特征之间的误差指标,并判断误差指标是否在预设的误差阈值范围之内。若误差指标在预设的误差阈值范围之内,则停止进行迭代优化操作,将当前迭代优化操作得到的待确认隐藏特征作为目标隐藏特征。若误差指标在预设的误差阈值范围之外,则将当前迭代优化操作输出的待确认隐藏特征作为下一次迭代优化操作的输入,继续进行下一次迭代优化操作。In some embodiments of the present application, after each hidden feature to be confirmed is obtained through an iterative optimization operation, the terminal may calculate the difference between the hidden feature to be confirmed output by the current iterative optimization operation and the hidden feature to be confirmed output by the previous iterative optimization operation Between the error indicators, and determine whether the error indicators are within the preset error threshold range. If the error index is within the preset error threshold range, the iterative optimization operation is stopped, and the hidden feature to be confirmed obtained by the current iterative optimization operation is used as the target hidden feature. If the error index is outside the preset error threshold range, the hidden feature to be confirmed output by the current iterative optimization operation is used as the input of the next iterative optimization operation, and the next iterative optimization operation is continued.
其中,误差阈值范围可以根据实际情况进行调整,本申请对此不进行限制。Wherein, the range of the error threshold can be adjusted according to the actual situation, which is not limited in this application.
也就是说,在进行第一次迭代优化操作后,终端可以判断第一次迭代优化操作输出的待确认隐藏特征与初始隐藏特征之间的的误差指标是否在误差阈值范围之内,从而决定是否需要进行下一迭代优化操作。若误差指标在误差阈值范围之内,则停止进行迭代优化操作,将第一次迭代优化操作输出的待确认隐藏特征作为目标隐藏特征。That is to say, after performing the first iterative optimization operation, the terminal can judge whether the error index between the hidden features to be confirmed and the initial hidden features output by the first iterative optimization operation is within the range of the error threshold, so as to determine whether The next iterative optimization operation is required. If the error index is within the range of the error threshold, the iterative optimization operation is stopped, and the hidden feature to be confirmed output by the first iterative optimization operation is used as the target hidden feature.
否则,进行第二次迭代优化操作,并在完成第二次迭代优化操作后,判断第二次迭代优化操作输出的待确认隐藏特征与第一次迭代优化操作输出的待确认隐藏特征之间的的误差指标是否在误差阈值范围之内,从而决定是否需要进行下一迭代优化操作。Otherwise, perform the second iterative optimization operation, and after the second iterative optimization operation is completed, judge the difference between the hidden features to be confirmed output by the second iterative optimization operation and the hidden features to be confirmed output by the first iterative optimization operation Whether the error index is within the error threshold range, so as to determine whether the next iterative optimization operation is needed.
以此类推,直至第N次迭代优化操作输出的待确认隐藏特征与第N-1次迭代优化操作输出的待确认隐藏特征的误差指标在误差阈值范围之内,停止进行迭代优化操作,将第N次迭代优化操作输出的待确认隐藏特征作为目标隐藏特征。By analogy, until the error index of the hidden feature to be confirmed output by the Nth iterative optimization operation and the hidden feature to be confirmed output by the N-1th iterative optimization operation are within the error threshold range, the iterative optimization operation is stopped, and the first The hidden features to be confirmed output by the N iteration optimization operation are used as the target hidden features.
在本申请的一些实施方式中,误差指标也可以基于当前迭代优化操作对应的稠密深度图像以及当前迭代优化操作的前一次迭代优化操作对应的稠密深度图像进行确定。其中,当前迭代优化操作对应的稠密深度图像即指对当前迭代优化操作输出的隐藏特征进行深度估计得到稠密深度图像。优选地,终端可以将当前迭代优化操作得到的稠密深度图像减去前一次迭代优化操作对应的稠密深度图像,然后计算平均绝对误差(Mean Absolute Error,MAE)值作为误差指标。此时,进行迭代优化操作的次数可以兼容稠密化效果与效率。In some embodiments of the present application, the error indicator may also be determined based on the dense depth image corresponding to the current iterative optimization operation and the dense depth image corresponding to the previous iterative optimization operation of the current iterative optimization operation. Wherein, the dense depth image corresponding to the current iterative optimization operation refers to the dense depth image obtained by performing depth estimation on hidden features output by the current iterative optimization operation. Preferably, the terminal may subtract the dense depth image corresponding to the previous iterative optimization operation from the dense depth image obtained by the current iterative optimization operation, and then calculate a mean absolute error (Mean Absolute Error, MAE) value as an error indicator. At this time, the number of iterative optimization operations can be compatible with the densification effect and efficiency.
步骤S104,利用目标隐藏特征进行深度估计,得到目标场景的目标稠密深度图像。In step S104, depth estimation is performed using hidden features of the target to obtain a target dense depth image of the target scene.
在本申请的实施方式中,终端对步骤S103确认的目标隐藏特征进行深度估计,即可得到目标场景的目标稠密深度图像。该稠密深度图像即为对稀疏深度图像进行深度补全后得到的图像,也即稠密化效果能够满足需求的深度图像。In the embodiment of the present application, the terminal performs depth estimation on the target hidden features confirmed in step S103 to obtain the target dense depth image of the target scene. The dense depth image is an image obtained by performing depth complementation on a sparse depth image, that is, a depth image whose densification effect can meet requirements.
在本申请的实施方式中,通过获取目标场景的彩色图像和稀疏深度图像,提取彩色图像的彩色特征和初始特征,根据彩色特征、初始特征及稀疏深度图像获取初始稠密深度图像及初始隐藏特征,并进行至少一次的迭代优化操作,以根据每次迭代优化操作获取到的待确认隐藏特征确认目标隐藏特征,并利用目标隐藏特征进行深度估计,得到目标场景的目标稠密深度图像,由于每次迭代优化操作需参考稀疏深度图像和彩色特征确定待确认隐藏特征,也即每次迭代优化操作的过程均会通过RGB图像的信息和稀疏深度图像的信息进行引导,能够提高所获取到的稠密深度图像的可靠性。In the embodiment of the present application, by acquiring the color image and sparse depth image of the target scene, extracting the color features and initial features of the color image, and obtaining the initial dense depth image and initial hidden features according to the color features, initial features and sparse depth image, And perform at least one iterative optimization operation to confirm the hidden features of the target according to the hidden features to be confirmed obtained by each iterative optimization operation, and use the hidden features of the target for depth estimation to obtain the target dense depth image of the target scene. Since each iteration The optimization operation needs to refer to the sparse depth image and color features to determine the hidden features to be confirmed, that is, the process of each iterative optimization operation will be guided by the information of the RGB image and the information of the sparse depth image, which can improve the obtained dense depth image. reliability.
而且,每一次迭代优化操作均会使用前一次迭代优化操作输出的待确认隐藏特征作为引导,因此,每一次迭代优化操作之后稠密深度图像的稠密化程度 将进一步得到提高。Moreover, each iterative optimization operation will use the hidden features to be confirmed output by the previous iterative optimization operation as a guide, so the densification degree of the dense depth image will be further improved after each iterative optimization operation.
需要说明的是,上述深度图像的获取方法可以通过网络模型实现。图3示出了深度图像模型的结构示意图。终端可以将彩色图像和稀疏深度图像输入至深度图像模型中,获取由深度图像模型输出的目标稠密深度图像。It should be noted that, the above-mentioned method for acquiring a depth image may be implemented through a network model. Fig. 3 shows a schematic diagram of the structure of a depth image model. The terminal can input the color image and the sparse depth image into the depth image model, and obtain the target dense depth image output by the depth image model.
其中,深度图像模型可以包括特征提取模块、N个反馈模块以及目标深度估计模块。Wherein, the depth image model may include a feature extraction module, N feedback modules and a target depth estimation module.
终端可以通过特征提取模块提取彩色图像的彩色特征和初始特征,并通过第一个反馈模块获取初始稠密深度图像及初始隐藏特征。The terminal can extract the color features and initial features of the color image through the feature extraction module, and obtain the initial dense depth image and initial hidden features through the first feedback module.
接着,终端可以依次通过剩余的反馈模块分别进行一次迭代优化操作,最后通过目标深度估计模块,对最后一个反馈模块输出的目标隐藏特征进行深度估计,得到目标场景的目标稠密深度图像。Then, the terminal can perform an iterative optimization operation through the remaining feedback modules in turn, and finally through the target depth estimation module, perform depth estimation on the hidden features of the target output by the last feedback module to obtain the target dense depth image of the target scene.
如图4所示,每个反馈模块可以包括中间深度估计模块、融合模块和序列模型模块。As shown in Fig. 4, each feedback module may include an intermediate depth estimation module, a fusion module and a sequence model module.
终端在单个反馈模块内进行迭代优化操作的步骤可以具体包括:通过当前反馈模块的中间深度估计模块,对前一隐藏特征(也即前一反馈模块输出的待确认隐藏特征)进行深度估计,得到当前反馈模块输出的稠密深度图像;通过当前反馈模块的融合模块,将当前反馈模块输出的稠密深度图像与稀疏深度图像进行融合,得到当前反馈模块的融合特征图;通过当前反馈模块的序列模型模块,利用彩色特征、当前反馈模块的融合特征图、前一隐藏特征,确定当前待确认隐藏特征(也即当前反馈模块输出的待确认隐藏特征)。The step of the terminal performing an iterative optimization operation in a single feedback module may specifically include: performing depth estimation on the previous hidden feature (that is, the hidden feature to be confirmed output by the previous feedback module) through the intermediate depth estimation module of the current feedback module, and obtaining The dense depth image output by the current feedback module; through the fusion module of the current feedback module, the dense depth image output by the current feedback module is fused with the sparse depth image to obtain the fusion feature map of the current feedback module; through the sequence model module of the current feedback module , use the color feature, the fusion feature map of the current feedback module, and the previous hidden feature to determine the current hidden feature to be confirmed (that is, the hidden feature to be confirmed output by the current feedback module).
上述反馈模块数量可以根据实际情况进行设置。并且,上述深度图像模型还可以输出一个用于表征稠密化效果的参数,若该参数大于预先设定好的阈值,则继续将当前反馈模块的输出输入下一个反馈模块中,若参数小于阈值,则将当前反馈模块输出的待确认隐藏特征作为目标隐藏特征,并通过目标深度估计模块,对目标隐藏特征进行深度估计,输出目标场景的目标稠密深度图像。The above number of feedback modules can be set according to actual conditions. Moreover, the above-mentioned depth image model can also output a parameter used to characterize the densification effect. If the parameter is greater than a preset threshold, continue to input the output of the current feedback module into the next feedback module. If the parameter is smaller than the threshold, The hidden feature to be confirmed output by the current feedback module is used as the target hidden feature, and the target hidden feature is estimated in depth through the target depth estimation module, and the target dense depth image of the target scene is output.
需要说明的是,上述深度图像模型的具体工作过程可以参考前述图1和图 2所示方法的描述,本申请对此不进行赘述。It should be noted that, for the specific working process of the above-mentioned depth image model, reference may be made to the description of the method shown in Figure 1 and Figure 2 above, which will not be repeated in this application.
在利用深度图像模型进行之前,终端需要对深度图像模型进行训练。进一步地,在训练深度图像模型的过程中,通过反馈模块进行迭代优化操作的迭代次数是固定的,若迭代次数不固定,反馈模块迭代优化得到的稠密深度图像将随着待训练网络参数的调整而改变,以致训练过程中存在两个变量,得不到准确的训练误差;而在深度图像模型的使用过程中,则反馈模块的迭代次数可不固定,迭代次数可以取决于当前迭代优化操作得到的待优化稠密深度图像与前一迭代优化操作得到的待优化稠密深度图像的误差。Before using the depth image model, the terminal needs to train the depth image model. Furthermore, in the process of training the depth image model, the number of iterations of the iterative optimization operation through the feedback module is fixed. If the number of iterations is not fixed, the dense depth image obtained by the iterative optimization of the feedback module will follow the adjustment of the network parameters to be trained. And change, so that there are two variables in the training process, and the accurate training error cannot be obtained; while in the process of using the depth image model, the number of iterations of the feedback module may not be fixed, and the number of iterations may depend on the current iterative optimization operation. The error between the dense depth image to be optimized and the dense depth image to be optimized obtained by the previous iterative optimization operation.
具体的,如图5所示,上述深度图像模型的训练过程可以包括步骤S501至步骤S503。Specifically, as shown in FIG. 5 , the above-mentioned training process of the depth image model may include steps S501 to S503.
步骤S501,获取样本彩色图像、样本稀疏深度图像和对应的参考稠密深度图像。Step S501, acquiring a sample color image, a sample sparse depth image and a corresponding reference dense depth image.
其中,样本彩色图像和样本稀疏深度图像的获取方式可以参考步骤S101的描述。Wherein, for the acquisition manner of the sample color image and the sample sparse depth image, please refer to the description of step S101.
参考稠密深度图像是理想的稠密深度图像。在一些实施方式中,可以获取由人工合成的深度图像,例如可以通过虚幻4(unreal engine 4,UE4)引擎实现,或者,也可以获取其它深度传感器(比如高精度的TOF深度相机)采集的深度图像。The reference dense depth image is an ideal dense depth image. In some implementations, artificially synthesized depth images can be obtained, for example, it can be realized by the Unreal 4 (unreal engine 4, UE4) engine, or the depth collected by other depth sensors (such as high-precision TOF depth cameras) can also be obtained image.
步骤S502,将样本彩色图像、样本稀疏深度图像输入待训练网络中,获取待训练网络中每一反馈模块输出的样本稠密深度图像,以及待训练网络中目标深度估计模块输出的样本目标稠密深度图像。Step S502, input the sample color image and sample sparse depth image into the network to be trained, obtain the sample dense depth image output by each feedback module in the network to be trained, and the sample target dense depth image output by the target depth estimation module in the network to be trained .
其中,待训练网络的模型结构和工作过程可以参看图1至图4的描述,本申请对此不进行赘述。Wherein, the model structure and working process of the network to be trained can refer to the descriptions in FIG. 1 to FIG. 4 , which will not be repeated in this application.
步骤S503,根据样本目标稠密深度图像、每个样本稠密深度图像,以及参考稠密深度图像,计算目标误差值,若目标误差值大于误差阈值,则调整待训练网络的参数对待训练网络进行迭代优化,直至目标误差值小于或等于误差阈 值,将待训练网络作为深度图像模型。Step S503, calculate the target error value according to the sample target dense depth image, each sample dense depth image, and the reference dense depth image, if the target error value is greater than the error threshold, adjust the parameters of the network to be trained to iteratively optimize the network to be trained, Until the target error value is less than or equal to the error threshold, the network to be trained is used as a depth image model.
其中,误差阈值指认为模型收敛时所允许的目标误差值的最大值,可以根据实际情况进行调整。Among them, the error threshold refers to the maximum value of the target error value allowed when the model converges, which can be adjusted according to the actual situation.
具体的,在本申请的一些实施方式中,终端可以计算样本目标稠密深度图像和每个样本稠密深度图像分别与参考稠密深度图像之间的初始误差值,然后,对初始误差值进行加权平均,得到目标误差值,进而保证迭代优化的稠密化效果更好。Specifically, in some embodiments of the present application, the terminal may calculate the initial error value between the sample target dense depth image and each sample dense depth image and the reference dense depth image, and then perform a weighted average on the initial error value, Get the target error value, and then ensure that the densification effect of iterative optimization is better.
通过与预设的误差阈值进行比对,如果目标误差值大于误差阈值,说明待训练网络未收敛,因此需要重新调整待训练网络的参数,并重新计算目标误差值,迭代直至目标误差值小于或等于误差阈值,说明待训练网络已能够输出可靠的稠密深度图像,可以将该网络作为深度图像模型并投入使用。By comparing with the preset error threshold, if the target error value is greater than the error threshold, it means that the network to be trained has not converged, so it is necessary to readjust the parameters of the network to be trained, and recalculate the target error value, and iterate until the target error value is less than or It is equal to the error threshold, indicating that the network to be trained has been able to output a reliable dense depth image, and the network can be used as a depth image model and put into use.
需要说明的是,上述样本彩色图像、样本稀疏深度图像和对应的参考稠密深度图像的数量可以为多个,每次迭代训练的过程,可以抽取其中的任意一个或多个进行训练。训练的过程可以采用梯度下降法实现,对应的损失函数(loss function)可以为L1范数损失函数、L2范数损失函数或其他损失函数。It should be noted that there may be multiple sample color images, sample sparse depth images, and corresponding reference dense depth images, and any one or more of them may be selected for training in each iterative training process. The training process can be implemented by using the gradient descent method, and the corresponding loss function (loss function) can be L1 norm loss function, L2 norm loss function or other loss functions.
需要说明的是,对于前述的各方法实施例,为了简单描述,故将其都表述为一系列的动作组合,但是本领域技术人员应该知悉,本申请并不受所描述的动作顺序的限制,因为根据本申请,某些步骤可以采用其它顺序进行。It should be noted that for the foregoing method embodiments, for the sake of simple description, they are expressed as a series of action combinations, but those skilled in the art should know that the present application is not limited by the described action sequence. Because of this application, certain steps may be performed in other orders.
如图6所示为本申请实施例提供的一种深度图像的获取装置600的结构示意图,所述深度图像的获取装置600配置于终端上。FIG. 6 is a schematic structural diagram of a depth image acquisition apparatus 600 provided in an embodiment of the present application, and the depth image acquisition apparatus 600 is configured on a terminal.
具体的,所述深度图像的获取装置600可以包括:Specifically, the acquisition device 600 of the depth image may include:
图像获取单元601,用于获取目标场景的彩色图像和稀疏深度图像;An image acquisition unit 601, configured to acquire a color image and a sparse depth image of a target scene;
初始稠密化单元602,用于提取所述彩色图像的彩色特征和初始特征,并根据所述彩色特征、所述初始特征及所述稀疏深度图像获取初始稠密深度图像及初始隐藏特征;An initial densification unit 602, configured to extract color features and initial features of the color image, and obtain an initial dense depth image and initial hidden features according to the color features, the initial features, and the sparse depth image;
迭代优化单元603,用于利用所述彩色特征、所述稀疏深度图像和所述初 始隐藏特征,对所述初始稠密深度图像进行至少一次迭代优化操作,并根据每次迭代优化操作获取到的待确认隐藏特征确认目标隐藏特征;An iterative optimization unit 603, configured to perform at least one iterative optimization operation on the initial dense depth image by using the color feature, the sparse depth image and the initial hidden feature, and obtain Confirm hidden features Confirm target hidden features;
目标稠密化单元604,用于利用所述目标隐藏特征进行深度估计,得到所述目标场景的目标稠密深度图像。The target densification unit 604 is configured to use the hidden features of the target to perform depth estimation to obtain a dense target depth image of the target scene.
其中,上述深度图像的获取装置600可以包括前述深度图像模型,请参考图3,初始稠密化单元602可以对应深度图像模型的特征提取模块及第一个反馈模块,迭代优化单元603可以对应深度图像模型中第一个反馈模块以外的其他反馈模块,目标稠密化单元604可以对应深度图像模型的目标深度估计模块。Wherein, the aforementioned depth image acquisition device 600 may include the aforementioned depth image model, please refer to FIG. 3 , the initial densification unit 602 may correspond to the feature extraction module and the first feedback module of the depth image model, and the iterative optimization unit 603 may correspond to the depth image For other feedback modules other than the first feedback module in the model, the object densification unit 604 may correspond to the object depth estimation module of the depth image model.
在本申请的一些实施方式中,上述初始稠密化单元602可以具体用于:对所述初始特征进行深度估计,得到初始稠密深度图像;将所述初始稠密深度图像与所述稀疏深度图像进行融合,得到初始融合特征图像;利用所述初始融合特征图像、所述初始特征和所述彩色特征,确定所述初始隐藏特征。In some implementations of the present application, the initial densification unit 602 may be specifically configured to: perform depth estimation on the initial features to obtain an initial dense depth image; fuse the initial dense depth image with the sparse depth image , to obtain an initial fusion feature image; using the initial fusion feature image, the initial feature, and the color feature to determine the initial hidden feature.
在本申请的一些实施方式中,上述图像获取单元601可以具体用于:获取所述目标场景的彩色图像和点云数据;将所述点云数据投影至所述彩色图像的成像平面上,得到所述稀疏深度图像。In some embodiments of the present application, the above-mentioned image acquisition unit 601 may be specifically configured to: acquire the color image and point cloud data of the target scene; project the point cloud data onto the imaging plane of the color image to obtain The sparse depth image.
在本申请的一些实施方式中,上述迭代优化单元603可以具体用于:利用所述初始隐藏特征对所述初始稠密深度图像进行至少一次迭代优化操作,在每次进行迭代优化操作后,计算当前迭代优化操作输出的待确认隐藏特征与前一次迭代优化操作输出的待确认隐藏特征之间的误差指标,若所述误差指标位于误差阈值范围之外,则继续进行下一次迭代优化操作,直至所述误差指标位于所述误差阈值范围之内,停止进行迭代优化操作,将当前迭代优化操作输出的待确认隐藏特征作为目标隐藏特征。In some embodiments of the present application, the above-mentioned iterative optimization unit 603 may be specifically configured to: use the initial hidden features to perform at least one iterative optimization operation on the initial dense depth image, and calculate the current The error index between the hidden feature to be confirmed output by the iterative optimization operation and the hidden feature to be confirmed output by the previous iterative optimization operation, if the error index is outside the error threshold range, continue to the next iterative optimization operation until the If the error index is within the range of the error threshold, the iterative optimization operation is stopped, and the hidden feature to be confirmed output by the current iterative optimization operation is used as the target hidden feature.
在本申请的一些实施方式中,上述深度图像的获取装置600还可以包括训练单元,可以用于:获取样本彩色图像、样本稀疏深度图像和对应的参考稠密深度图像;将所述样本彩色图像、所述样本稀疏深度图像输入待训练网络中,获取所述待训练网络中每一反馈模块输出的样本稠密深度图像,以及所述待训 练网络中目标深度估计模块输出的样本目标稠密深度图像;根据所述样本目标稠密深度图像、每个所述样本稠密深度图像,以及所述参考稠密深度图像,计算目标误差值,若所述目标误差值大于误差阈值,则调整所述待训练网络的参数,以对所述待训练网络进行迭代优化,直至所述目标误差值小于或等于所述误差阈值,将所述待训练网络作为所述深度图像模型。In some embodiments of the present application, the depth image acquisition apparatus 600 may further include a training unit, which may be used to: acquire a sample color image, a sample sparse depth image, and a corresponding reference dense depth image; The sample sparse depth image is input into the network to be trained, and the sample dense depth image output by each feedback module in the network to be trained is obtained, and the sample target dense depth image output by the target depth estimation module in the network to be trained; according to calculating a target error value for the sample target dense depth image, each of the sample dense depth images, and the reference dense depth image, and adjusting the parameters of the network to be trained if the target error value is greater than an error threshold, The network to be trained is optimized iteratively until the target error value is less than or equal to the error threshold, and the network to be trained is used as the depth image model.
在本申请的一些实施方式中,上述训练单元可以具体用于:通过所述待训练网络中的特征提取模块,提取所述彩色图像的彩色特征和初始特征;通过所述待训练网络中的第一个反馈模块,根据所述彩色特征、所述初始特征及所述稀疏深度图像获取初始稠密深度图像及初始隐藏特征;通过所述待训练网络中其他反馈模块分别进行一次迭代优化操作,输出每一次迭代优化操作得到的样本稠密深度图像。In some embodiments of the present application, the above-mentioned training unit may be specifically configured to: extract the color features and initial features of the color image through the feature extraction module in the network to be trained; A feedback module, which obtains an initial dense depth image and an initial hidden feature according to the color feature, the initial feature and the sparse depth image; performs an iterative optimization operation through other feedback modules in the network to be trained, and outputs each A sample dense depth image obtained by an iterative optimization operation.
在本申请的一些实施方式中,上述训练单元可以具体用于:计算所述样本目标稠密深度图像和每个所述样本稠密深度图像分别与所述参考稠密深度图像之间的初始误差值;对所述初始误差值进行加权平均,得到所述目标误差值。In some embodiments of the present application, the above training unit may be specifically configured to: calculate the initial error value between the sample target dense depth image and each of the sample dense depth images and the reference dense depth image; The initial error value is weighted and averaged to obtain the target error value.
需要说明的是,为描述的方便和简洁,上述深度图像的获取装置600的具体工作过程,可以参考图1至图5所述方法的对应过程,在此不再赘述。It should be noted that, for the convenience and brevity of description, the specific working process of the depth image acquisition apparatus 600 can refer to the corresponding process of the methods described in FIG. 1 to FIG. 5 , which will not be repeated here.
本申请实施例还提供的一种深度系统,系统具体包括彩色模块、深度模块、及前述的深度图像的获取装置600,其中,彩色模块,用于采集目标场景的彩色图像;深度模块,用于对目标场景进行扫描得到点云数据,并根据点云数据得到稀疏深度图像;获取装置,利用所述彩色图像及所述稀疏深度图像得到目标稠密深度图像。需要说明的是,彩色模块包括彩色相机,深度模块包括但不限于激光雷达,直接测量飞行时间(Direct Time of flight,dTof)相机,散斑间接测量飞行时间(Indirect Time of flight,iTof)相机中的任一一种;彩色模块、深度模块及获取装置可为一体化装置或独立设置,各元件间的数据可通过有线或无线传输,此处不作限制。所述深度系统的具体工作过程可以参看图1至图6的描述,本申请对此不进行赘述。The embodiment of the present application also provides a depth system. The system specifically includes a color module, a depth module, and the aforementioned depth image acquisition device 600, wherein the color module is used to collect a color image of the target scene; the depth module is used to The target scene is scanned to obtain point cloud data, and a sparse depth image is obtained according to the point cloud data; the obtaining device uses the color image and the sparse depth image to obtain a dense depth image of the target. It should be noted that the color module includes a color camera, and the depth module includes but is not limited to a laser radar, a direct time of flight (Direct Time of flight, dTof) camera, and a speckle indirect time of flight (Indirect Time of flight, iTof) camera. Any one of them; the color module, the depth module and the acquisition device can be an integrated device or an independent device, and the data between each component can be transmitted by wire or wireless, which is not limited here. For the specific working process of the depth system, reference can be made to the descriptions in FIG. 1 to FIG. 6 , which will not be repeated in this application.
如图7所示,为本申请实施例提供的一种终端的示意图。该终端7可以包括:处理器70、存储器71以及存储在所述存储器71中并可在所述处理器70上运行的计算机程序72,例如深度图像的获取程序。所述处理器70执行所述计算机程序72时实现上述各个深度图像的获取方法实施例中的步骤,例如图1所示的步骤S101至S104。或者,所述处理器70执行所述计算机程序72时实现上述各装置实施例中各模块/单元的功能,例如图6所示的图像获取单元601、初始稠密化单元602、迭代优化单元603和目标稠密化单元604。As shown in FIG. 7 , it is a schematic diagram of a terminal provided in the embodiment of the present application. The terminal 7 may include: a processor 70, a memory 71, and a computer program 72 stored in the memory 71 and operable on the processor 70, such as a depth image acquisition program. When the processor 70 executes the computer program 72, it implements the steps in the embodiments of the methods for acquiring depth images above, such as steps S101 to S104 shown in FIG. 1 . Alternatively, when the processor 70 executes the computer program 72, it realizes the functions of the modules/units in the above-mentioned device embodiments, such as the image acquisition unit 601, the initial densification unit 602, the iterative optimization unit 603 and the Target densification unit 604 .
所述计算机程序可以被分割成一个或多个模块/单元,所述一个或者多个模块/单元被存储在所述存储器71中,并由所述处理器70执行,以完成本申请。所述一个或多个模块/单元可以是能够完成特定功能的一系列计算机程序指令段,该指令段用于描述所述计算机程序在所述终端中的执行过程。The computer program can be divided into one or more modules/units, and the one or more modules/units are stored in the memory 71 and executed by the processor 70 to complete the present application. The one or more modules/units may be a series of computer program instruction segments capable of accomplishing specific functions, and the instruction segments are used to describe the execution process of the computer program in the terminal.
例如,所述计算机程序可以被分割成:图像获取单元、初始稠密化单元、迭代优化单元和目标稠密化单元。For example, the computer program can be divided into: an image acquisition unit, an initial densification unit, an iterative optimization unit and a target densification unit.
各单元具体功能如下:The specific functions of each unit are as follows:
图像获取单元,用于获取目标场景的彩色图像和稀疏深度图像;初始稠密化单元,用于提取所述彩色图像的彩色特征和初始特征,并根据所述彩色特征、所述初始特征及所述稀疏深度图像获取初始稠密深度图像及初始隐藏特征;迭代优化单元,用于利用所述彩色特征、所述稀疏深度图像和所述初始隐藏特征,对所述初始稠密深度图像进行至少一次迭代优化操作,并根据每次迭代优化操作获取到的待确认隐藏特征确认目标隐藏特征;目标稠密化单元,用于利用所述目标隐藏特征进行深度估计,得到所述目标场景的目标稠密深度图像。The image acquisition unit is used to acquire the color image and the sparse depth image of the target scene; the initial densification unit is used to extract the color features and initial features of the color image, and according to the color features, the initial features and the The sparse depth image acquires an initial dense depth image and initial hidden features; an iterative optimization unit is configured to perform at least one iterative optimization operation on the initial dense depth image by using the color features, the sparse depth image, and the initial hidden features , and confirm the target hidden feature according to the hidden features to be confirmed obtained by each iterative optimization operation; the target densification unit is configured to use the target hidden feature to perform depth estimation to obtain a target dense depth image of the target scene.
所述终端可包括,但不仅限于,处理器70、存储器71。本领域技术人员可以理解,图7仅仅是终端的示例,并不构成对终端的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如所述终端还可以包括输入输出设备、网络接入设备、总线等。The terminal may include, but not limited to, a processor 70 and a memory 71 . Those skilled in the art can understand that FIG. 7 is only an example of a terminal, and does not constitute a limitation on the terminal. It may include more or less components than those shown in the figure, or combine certain components, or different components, such as the Terminals may also include input and output devices, network access devices, buses, and so on.
所称处理器70可以是中央处理单元(Central Processing Unit,CPU),还可 以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The so-called processor 70 can be a central processing unit (Central Processing Unit, CPU), and can also be other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Off-the-shelf programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
所述存储器71可以是所述终端的内部存储单元,例如终端的硬盘或内存。所述存储器71也可以是所述终端的外部存储设备,例如所述终端上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器71还可以既包括所述终端的内部存储单元也包括外部存储设备。所述存储器71用于存储所述计算机程序以及所述终端所需的其他程序和数据。所述存储器71还可以用于暂时地存储已经输出或者将要输出的数据。The storage 71 may be an internal storage unit of the terminal, such as a hard disk or memory of the terminal. The memory 71 can also be an external storage device of the terminal, such as a plug-in hard disk equipped on the terminal, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, a flash memory card (Flash Card) etc. Further, the memory 71 may also include both an internal storage unit of the terminal and an external storage device. The memory 71 is used to store the computer program and other programs and data required by the terminal. The memory 71 can also be used to temporarily store data that has been output or will be output.
需要说明的是,为描述的方便和简洁,上述终端的结构还可以参考方法实施例中对结构的具体描述,在此不再赘述。It should be noted that, for the convenience and brevity of description, the structure of the terminal above can also refer to the specific description of the structure in the method embodiment, and details are not repeated here.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。实施例中的各功能单元、模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中,上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。另外,各功能单元、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述系统中单元、模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that for the convenience and brevity of description, only the division of the above-mentioned functional units and modules is used for illustration. In practical applications, the above-mentioned functions can be assigned to different functional units, Completion of modules means that the internal structure of the device is divided into different functional units or modules to complete all or part of the functions described above. Each functional unit and module in the embodiment can be integrated into one processing unit, or each unit can exist separately physically, or two or more units can be integrated into one unit, and the above-mentioned integrated units can either adopt hardware It can also be implemented in the form of software functional units. In addition, the specific names of the functional units and modules are only for the convenience of distinguishing each other, and are not used to limit the protection scope of the present application. For the specific working processes of the units and modules in the above system, reference may be made to the corresponding processes in the aforementioned method embodiments, and details will not be repeated here.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the above-mentioned embodiments, the descriptions of each embodiment have their own emphases, and for parts that are not detailed or recorded in a certain embodiment, refer to the relevant descriptions of other embodiments.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对各个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。Those skilled in the art can appreciate that the units and algorithm steps of the examples described in conjunction with the embodiments disclosed herein can be implemented by electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are executed by hardware or software depends on the specific application and design constraints of the technical solution. Those skilled in the art may use different methods to implement the described functions for each specific application, but such implementation should not be regarded as exceeding the scope of the present application.
在本申请所提供的实施例中,应该理解到,所揭露的装置/终端和方法,可以通过其它的方式实现。例如,以上所描述的装置/终端实施例仅仅是示意性的,例如,所述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,装置或单元的间接耦合或通讯连接,可以是电性,机械或其它的形式。In the embodiments provided in this application, it should be understood that the disclosed device/terminal and method may be implemented in other ways. For example, the device/terminal embodiments described above are only illustrative. For example, the division of the modules or units is only a logical function division. In actual implementation, there may be other division methods, such as multiple units or Components may be combined or integrated into another system, or some features may be omitted, or not implemented. In another point, the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of devices or units may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components shown as units may or may not be physical units, that is, they may be located in one place, or may be distributed to multiple network units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit. The above-mentioned integrated units can be implemented in the form of hardware or in the form of software functional units.
所述集成的模块/单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请实现上述实施例方法中的全部或部分流程,也可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一计算机可读存储介质中,该计算机程序在被处理器执行时,可实现上述各个方法实施例的步骤。其中,所述计算机程序包括计算机程序代码,所述计算机程序代码可以为源代码形式、 对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质可以包括:能够携带所述计算机程序代码的任何实体或装置、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、电载波信号、电信信号以及软件分发介质等。需要说明的是,所述计算机可读介质包含的内容可以根据司法管辖区内立法和专利实践的要求进行适当的增减,例如在某些司法管辖区,根据立法和专利实践,计算机可读介质不包括电载波信号和电信信号。If the integrated module/unit is realized in the form of a software function unit and sold or used as an independent product, it can be stored in a computer-readable storage medium. Based on this understanding, all or part of the processes in the methods of the above embodiments in the present application can also be completed by instructing related hardware through computer programs. The computer programs can be stored in a computer-readable storage medium, and the computer When the program is executed by the processor, the steps in the above-mentioned various method embodiments can be realized. Wherein, the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file or some intermediate form. The computer-readable medium may include: any entity or device capable of carrying the computer program code, a recording medium, a USB flash drive, a removable hard disk, a magnetic disk, an optical disk, a computer memory, and a read-only memory (Read-Only Memory, ROM) , random access memory (Random Access Memory, RAM), electric carrier signal, telecommunication signal and software distribution medium, etc. It should be noted that the content contained in the computer-readable medium may be appropriately increased or decreased according to the requirements of legislation and patent practice in the jurisdiction. For example, in some jurisdictions, computer-readable media Excludes electrical carrier signals and telecommunication signals.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-described embodiments are only used to illustrate the technical solutions of the present application, rather than to limit them; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that: it can still implement the foregoing embodiments Modifications to the technical solutions described in the examples, or equivalent replacements for some of the technical features; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of the various embodiments of the application, and should be included in the Within the protection scope of this application.

Claims (11)

  1. 一种深度图像的获取方法,其特征在于,包括:A method for acquiring a depth image, comprising:
    获取目标场景的彩色图像和稀疏深度图像;Obtain a color image and a sparse depth image of the target scene;
    提取所述彩色图像的彩色特征和初始特征,并根据所述彩色特征、所述初始特征及所述稀疏深度图像获取初始稠密深度图像及初始隐藏特征;extracting color features and initial features of the color image, and obtaining an initial dense depth image and initial hidden features according to the color features, the initial features, and the sparse depth image;
    利用所述彩色特征、所述稀疏深度图像和所述初始隐藏特征,对所述初始稠密深度图像进行至少一次迭代优化操作,并根据每次迭代优化操作获取到的待确认隐藏特征确认目标隐藏特征;Using the color feature, the sparse depth image and the initial hidden feature, perform at least one iterative optimization operation on the initial dense depth image, and confirm the target hidden feature according to the hidden features to be confirmed obtained by each iterative optimization operation ;
    利用所述目标隐藏特征进行深度估计,得到所述目标场景的目标稠密深度图像。Depth estimation is performed using the target hidden features to obtain a target dense depth image of the target scene.
  2. 如权利要求1所述的深度图像的获取方法,其特征在于,所述根据所述彩色特征、所述初始特征及所述稀疏深度图像获取初始稠密深度图像及初始隐藏特征,包括:The method for obtaining a depth image according to claim 1, wherein said obtaining an initial dense depth image and an initial hidden feature according to said color feature, said initial feature, and said sparse depth image comprises:
    对所述初始特征进行深度估计,得到初始稠密深度图像;performing depth estimation on the initial features to obtain an initial dense depth image;
    将所述初始稠密深度图像与所述稀疏深度图像进行融合,得到初始融合特征图像;Fusing the initial dense depth image with the sparse depth image to obtain an initial fusion feature image;
    利用所述初始融合特征图像、所述初始特征和所述彩色特征,确定所述初始隐藏特征。The initial hidden features are determined using the initial fused feature image, the initial features, and the color features.
  3. 如权利要求1或2所述的深度图像的获取方法,其特征在于,所述获取目标场景的彩色图像和稀疏深度图像,包括:The method for acquiring a depth image according to claim 1 or 2, wherein the acquisition of the color image and the sparse depth image of the target scene comprises:
    获取所述目标场景的彩色图像和点云数据;Obtain the color image and point cloud data of the target scene;
    将所述点云数据投影至所述彩色图像的成像平面上,得到所述稀疏深度图像。The point cloud data is projected onto the imaging plane of the color image to obtain the sparse depth image.
  4. 如权利要求1或2所述的深度图像的获取方法,其特征在于,所述利用所述彩色特征、所述稀疏深度图像和所述初始隐藏特征,对所述初始稠密深度图像进行至少一次迭代优化操作,并根据每次迭代优化操作获取到的待确认隐藏特征确认目标隐藏特征,包括:The method for acquiring a depth image according to claim 1 or 2, wherein the initial dense depth image is iterated at least once by using the color feature, the sparse depth image and the initial hidden feature Optimize the operation, and confirm the target hidden features according to the hidden features to be confirmed obtained by each iterative optimization operation, including:
    利用所述初始隐藏特征对所述初始稠密深度图像进行至少一次迭代优化操作,在每次进行迭代优化操作后,计算当前迭代优化操作输出的待确认隐藏特征与前一次迭代优化操作输出的待确认隐藏特征之间的误差指标,若所述误差指标位于误差阈值范围之外,则继续进行下一次迭代优化操作,直至所述误差指标位于所述误差阈值范围之内,停止进行迭代优化操作,将当前迭代优化操作输出的待确认隐藏特征作为目标隐藏特征。Using the initial hidden features to perform at least one iterative optimization operation on the initial dense depth image, after each iterative optimization operation, calculate the hidden features to be confirmed output by the current iterative optimization operation and the unconfirmed hidden features output by the previous iterative optimization operation The error index between the hidden features, if the error index is outside the error threshold range, continue the next iterative optimization operation until the error index is within the error threshold range, stop the iterative optimization operation, and set The hidden features to be confirmed output by the current iterative optimization operation are used as the target hidden features.
  5. 如权利要求1或2所述的深度图像的获取方法,其特征在于,所述深度图像的获取方法由预先训练得到的深度图像模型执行;The method for obtaining a depth image according to claim 1 or 2, wherein the method for obtaining a depth image is performed by a pre-trained depth image model;
    其中,所述深度图像模型的训练过程包括:Wherein, the training process of the depth image model includes:
    获取样本彩色图像、样本稀疏深度图像和对应的参考稠密深度图像;Obtain the sample color image, the sample sparse depth image and the corresponding reference dense depth image;
    将所述样本彩色图像、所述样本稀疏深度图像输入待训练网络中,获取所述待训练网络中每一反馈模块输出的样本稠密深度图像,以及所述待训练网络中目标深度估计模块输出的样本目标稠密深度图像;Input the sample color image and the sample sparse depth image into the network to be trained, obtain the sample dense depth image output by each feedback module in the network to be trained, and the target depth estimation module output in the network to be trained Sample target dense depth image;
    根据所述样本目标稠密深度图像、每个所述样本稠密深度图像,以及所述参考稠密深度图像,计算目标误差值,若所述目标误差值大于误差阈值,则调整所述待训练网络的参数,以对所述待训练网络进行迭代优化,直至所述目标误差值小于或等于所述误差阈值,将所述待训练网络作为所述深度图像模型。Calculate a target error value based on the sample target dense depth image, each of the sample dense depth images, and the reference dense depth image, and adjust the parameters of the network to be trained if the target error value is greater than an error threshold , to iteratively optimize the network to be trained until the target error value is less than or equal to the error threshold, and use the network to be trained as the depth image model.
  6. 如权利要求5所述的深度图像的获取方法,其特征在于,所述获取所述待训练网络中每一反馈模块输出的样本稠密深度图像,包括:The method for obtaining a depth image according to claim 5, wherein said obtaining the sample-dense depth image output by each feedback module in the network to be trained comprises:
    通过所述待训练网络中的特征提取模块,提取所述彩色图像的彩色特征和 初始特征;By the feature extraction module in the network to be trained, extract the color features and initial features of the color image;
    通过所述待训练网络中的第一个反馈模块,根据所述彩色特征、所述初始特征及所述稀疏深度图像获取初始稠密深度图像及初始隐藏特征;Obtain an initial dense depth image and an initial hidden feature according to the color feature, the initial feature, and the sparse depth image through the first feedback module in the network to be trained;
    通过所述待训练网络中其他反馈模块分别进行一次迭代优化操作,输出每一次迭代优化操作得到的样本稠密深度图像。Perform an iterative optimization operation through other feedback modules in the network to be trained, and output a sample-dense depth image obtained by each iterative optimization operation.
  7. 如权利要求5所述的深度图像的获取方法,其特征在于,所述根据所述样本目标稠密深度图像、每个所述样本稠密深度图像,以及所述参考稠密深度图像,计算目标误差值,包括:The method for acquiring a depth image according to claim 5, wherein the target error value is calculated according to the sample target dense depth image, each of the sample dense depth images, and the reference dense depth image, include:
    计算所述样本目标稠密深度图像和每个所述样本稠密深度图像分别与所述参考稠密深度图像之间的初始误差值;calculating an initial error value between the sample target dense depth image and each of the sample dense depth images and the reference dense depth image;
    对所述初始误差值进行加权平均,得到所述目标误差值。The weighted average is performed on the initial error value to obtain the target error value.
  8. 一种深度图像的获取装置,其特征在于,包括:A device for acquiring a depth image, characterized in that it comprises:
    图像获取单元,用于获取目标场景的彩色图像和稀疏深度图像;An image acquisition unit, configured to acquire a color image and a sparse depth image of the target scene;
    初始稠密化单元,用于提取所述彩色图像的彩色特征和初始特征,并根据所述彩色特征、所述初始特征及所述稀疏深度图像获取初始稠密深度图像及初始隐藏特征;an initial densification unit, configured to extract color features and initial features of the color image, and obtain an initial dense depth image and initial hidden features according to the color features, the initial features, and the sparse depth image;
    迭代优化单元,用于利用所述彩色特征、所述稀疏深度图像和所述初始隐藏特征,对所述初始稠密深度图像进行至少一次迭代优化操作,并根据每次迭代优化操作获取到的待确认隐藏特征确认目标隐藏特征;An iterative optimization unit, configured to perform at least one iterative optimization operation on the initial dense depth image by using the color feature, the sparse depth image, and the initial hidden feature, and obtain the information to be confirmed according to each iterative optimization operation Hidden features confirm target hidden features;
    目标稠密化单元,用于利用所述目标隐藏特征进行深度估计,得到所述目标场景的目标稠密深度图像。The target densification unit is configured to use the hidden features of the target to perform depth estimation to obtain a target dense depth image of the target scene.
  9. 一种深度系统,其特征在于,包括彩色模块、深度模块及如权利要求8所述的获取装置,其中:A depth system, characterized in that it comprises a color module, a depth module and the acquisition device according to claim 8, wherein:
    所述彩色模块,用于采集目标场景的彩色图像;The color module is used to collect a color image of a target scene;
    所述深度模块,用于对所述目标场景进行扫描得到点云数据,并根据所述点云数据得到稀疏深度图像;The depth module is used to scan the target scene to obtain point cloud data, and obtain a sparse depth image according to the point cloud data;
    所述获取装置,利用所述彩色图像及所述稀疏深度图像得到目标稠密深度图像。The acquisition device uses the color image and the sparse depth image to obtain a target dense depth image.
  10. 一种终端,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如权利要求1至7任一项所述获取方法的步骤。A terminal, comprising a memory, a processor, and a computer program stored in the memory and operable on the processor, characterized in that, when the processor executes the computer program, the computer program according to claims 1 to 7 is implemented. The steps of any one of the acquisition methods.
  11. 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至7任一项所述获取方法的步骤。A computer-readable storage medium, the computer-readable storage medium stores a computer program, wherein when the computer program is executed by a processor, the steps of the acquisition method according to any one of claims 1 to 7 are realized.
PCT/CN2022/100593 2022-02-16 2022-06-23 Depth image acquisition method and apparatus, and depth system, terminal and storage medium WO2023155353A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210142004.8 2022-02-16
CN202210142004.8A CN114638869A (en) 2022-02-16 2022-02-16 Method and device for acquiring depth image, depth system, terminal and storage medium

Publications (1)

Publication Number Publication Date
WO2023155353A1 true WO2023155353A1 (en) 2023-08-24

Family

ID=81945910

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/100593 WO2023155353A1 (en) 2022-02-16 2022-06-23 Depth image acquisition method and apparatus, and depth system, terminal and storage medium

Country Status (2)

Country Link
CN (1) CN114638869A (en)
WO (1) WO2023155353A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117422629A (en) * 2023-12-19 2024-01-19 华南理工大学 Instance-aware monocular semantic scene completion method, medium and device

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114638869A (en) * 2022-02-16 2022-06-17 奥比中光科技集团股份有限公司 Method and device for acquiring depth image, depth system, terminal and storage medium
CN115100360B (en) * 2022-07-28 2023-12-01 中国电信股份有限公司 Image generation method and device, storage medium and electronic equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112102472A (en) * 2020-09-01 2020-12-18 北京航空航天大学 Sparse three-dimensional point cloud densification method
US20200410699A1 (en) * 2018-03-13 2020-12-31 Magic Leap, Inc. Image-enhanced depth sensing using machine learning
WO2021013334A1 (en) * 2019-07-22 2021-01-28 Toyota Motor Europe Depth maps prediction system and training method for such a system
CN112541482A (en) * 2020-12-25 2021-03-23 北京百度网讯科技有限公司 Deep information completion model training method, device, equipment and storage medium
CN112560875A (en) * 2020-12-25 2021-03-26 北京百度网讯科技有限公司 Deep information completion model training method, device, equipment and storage medium
CN114638869A (en) * 2022-02-16 2022-06-17 奥比中光科技集团股份有限公司 Method and device for acquiring depth image, depth system, terminal and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200410699A1 (en) * 2018-03-13 2020-12-31 Magic Leap, Inc. Image-enhanced depth sensing using machine learning
WO2021013334A1 (en) * 2019-07-22 2021-01-28 Toyota Motor Europe Depth maps prediction system and training method for such a system
CN112102472A (en) * 2020-09-01 2020-12-18 北京航空航天大学 Sparse three-dimensional point cloud densification method
CN112541482A (en) * 2020-12-25 2021-03-23 北京百度网讯科技有限公司 Deep information completion model training method, device, equipment and storage medium
CN112560875A (en) * 2020-12-25 2021-03-26 北京百度网讯科技有限公司 Deep information completion model training method, device, equipment and storage medium
CN114638869A (en) * 2022-02-16 2022-06-17 奥比中光科技集团股份有限公司 Method and device for acquiring depth image, depth system, terminal and storage medium

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117422629A (en) * 2023-12-19 2024-01-19 华南理工大学 Instance-aware monocular semantic scene completion method, medium and device
CN117422629B (en) * 2023-12-19 2024-04-26 华南理工大学 Instance-aware monocular semantic scene completion method, medium and device

Also Published As

Publication number Publication date
CN114638869A (en) 2022-06-17

Similar Documents

Publication Publication Date Title
WO2023155353A1 (en) Depth image acquisition method and apparatus, and depth system, terminal and storage medium
CN111563923B (en) Method for obtaining dense depth map and related device
US11532094B2 (en) Systems and methods for three-dimensional pose determination
US20220383549A1 (en) Multi-mode three-dimensional scanning method and system
US9189889B2 (en) Method for building a three-dimensional model and apparatus thereof
US8588516B2 (en) Interpolation image generation apparatus, reconstructed image generation apparatus, method of generating interpolation image, and computer-readable recording medium storing program
JP2017520050A (en) Local adaptive histogram flattening
CN109672871B (en) White balance information synchronization method, white balance information synchronization device and computer readable medium
KR20190112894A (en) Method and apparatus for 3d rendering
CN110378944B (en) Depth map processing method and device and electronic equipment
CN112862877A (en) Method and apparatus for training image processing network and image processing
WO2022135588A1 (en) Image correction method, apparatus and system, and electronic device
CN114494388B (en) Three-dimensional image reconstruction method, device, equipment and medium in large-view-field environment
WO2023142352A1 (en) Depth image acquisition method and device, terminal, imaging system and medium
WO2021115061A1 (en) Image segmentation method and apparatus, and server
WO2020124374A1 (en) Image processing method, terminal device and storage medium
WO2022036539A1 (en) Color consistency correction method and device for multiple cameras
CN116612468A (en) Three-dimensional target detection method based on multi-mode fusion and depth attention mechanism
CN114078093A (en) Image correction method, intelligent terminal and storage medium
CN110060264B (en) Neural network training method, video frame processing method, device and system
CN112184828B (en) Laser radar and camera external parameter calibration method and device and automatic driving vehicle
CN113269823A (en) Depth data acquisition method and device, storage medium and electronic equipment
US20230384085A1 (en) Phase unwrapping method based on multi-view constraints of light field and related components
CN116778091A (en) Deep learning multi-view three-dimensional reconstruction algorithm based on path aggregation
JP7321772B2 (en) Image processing device, image processing method, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22926662

Country of ref document: EP

Kind code of ref document: A1