WO2019219012A1 - Three-dimensional reconstruction method and device uniting rigid motion and non-rigid deformation - Google Patents

Three-dimensional reconstruction method and device uniting rigid motion and non-rigid deformation Download PDF

Info

Publication number
WO2019219012A1
WO2019219012A1 PCT/CN2019/086889 CN2019086889W WO2019219012A1 WO 2019219012 A1 WO2019219012 A1 WO 2019219012A1 CN 2019086889 W CN2019086889 W CN 2019086889W WO 2019219012 A1 WO2019219012 A1 WO 2019219012A1
Authority
WO
WIPO (PCT)
Prior art keywords
rigid
model
dimensional
deformation
motion
Prior art date
Application number
PCT/CN2019/086889
Other languages
French (fr)
Chinese (zh)
Inventor
刘烨斌
戴琼海
徐枫
方璐
Original Assignee
清华大学
清华大学深圳研究生院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 清华大学, 清华大学深圳研究生院 filed Critical 清华大学
Publication of WO2019219012A1 publication Critical patent/WO2019219012A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/521Depth or shape recovery from laser ranging, e.g. using interferometry; from the projection of structured light
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images

Definitions

  • the invention relates to the technical field of computer vision and computer graphics, in particular to a three-dimensional reconstruction method and device for joint rigid motion and non-rigid deformation.
  • Dynamic object 3D reconstruction is a key issue in the field of computer graphics and computer vision.
  • High-quality dynamic object 3D models such as human body, animal, human face, human hand, etc.
  • the acquisition of high-quality 3D models usually relies on expensive laser scanners or multi-camera array systems.
  • the accuracy is high, there are also some shortcomings: First, the object is required to remain absolutely still during the scanning process. Movement will lead to obvious errors in the scanning results. Second, the fraud is expensive and difficult to spread to the daily lives of ordinary people, often applied to large companies or national statistical departments. Third, the speed is slow, and it often takes at least 10 minutes to several hours to reconstruct a 3D model. The cost of reconstructing a dynamic model sequence is greater.
  • the existing reconstruction method concentrates on solving the rigid motion information of the object first, obtaining the approximation of the object, and reconstructing the non-rigid surface motion information.
  • this reconstruction method requires obtaining a three-dimensional model of the key frame of the object in advance.
  • the existing frame-by-frame dynamic fusion surface reconstruction method can realize dynamic three-dimensional reconstruction without template, the robustness of tracking reconstruction is low only by using the non-rigid surface deformation method.
  • the present invention aims to solve at least one of the technical problems in the related art to some extent.
  • an object of the present invention is to propose a three-dimensional reconstruction method combining joint rigid motion and non-rigid deformation, which can effectively improve the real-time, robustness and accuracy of reconstruction, and has strong scalability and is easy to implement.
  • Another object of the present invention is to provide a three-dimensional reconstruction apparatus that combines rigid motion and non-rigid deformation.
  • an embodiment of the present invention provides a three-dimensional reconstruction method combining joint rigid motion and non-rigid deformation, including the following steps: performing depth camera-based photographing on a target object to obtain a single depth image;
  • the skeleton extraction algorithm performs three-dimensional skeleton extraction on the depth point cloud; transforms the single depth image into a three-dimensional point cloud, and acquires a matching point pair between the three-dimensional point cloud and the reconstructed model vertex; and according to the matching point pair
  • the three-dimensional skeleton information establishes an energy function, and solves a non-rigid motion position transformation parameter of each vertex on the reconstruction model and optimizes an object skeleton parameter; performing a GPU (Graphics Processing Unit) optimization solution on the energy function to Obtaining a non-rigid deformation of each surface vertex, deforming the reconstructed three-dimensional model of the previous frame according to the solution result, so that the deformation model is aligned with the current frame three-dimensional point cloud; and the current frame
  • the three-dimensional reconstruction method of the joint rigid motion and the non-rigid deformation combines the three-dimensional information of the dynamic object surface frame by frame by real-time non-rigid alignment method, in order to achieve robust tracking, realizing the three-dimensional frame without the first frame Robust real-time dynamic 3D reconstruction under template conditions can effectively improve the real-time, robustness and accuracy of reconstruction, and it is scalable and easy to implement.
  • the three-dimensional reconstruction method of the joint rigid motion and the non-rigid deformation may further have the following additional technical features:
  • the converting the single depth image into a three-dimensional point cloud further comprises: projecting the single depth image into the three-dimensional space by using an internal parameter matrix of the depth camera, Generating the three-dimensional point cloud, wherein the depth map projection formula is:
  • u, v are pixel coordinates
  • d(u, v) is a depth value at a pixel (u, v) position on the depth image
  • the energy function is:
  • E t is the total energy term
  • E n is the non-rigid surface deformation constraint term
  • E s is the rigid skeleton motion constraint term
  • E j is the rigid skeleton recognition constraint term
  • E g is the local rigid motion constraint term
  • ⁇ n , ⁇ s , ⁇ j and ⁇ g are weight coefficients corresponding to respective constraint terms, respectively.
  • u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair
  • c i represents the i-th element in the set of matching point pairs
  • the non-rigid surface deformation constraint item with Representing the vertex coordinates of the model and its normal direction driven by the non-rigid deformation, respectively, in the rigid skeleton motion constraint with Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton, with Representing the model vertex coordinates driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation, where i represents the i-th vertex on the model.
  • the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
  • a deformation matrix acting on the vertex v i including two parts of rotation and translation; Is the rotating portion of the deformation matrix; a set of bones that have a driving effect on the vertex v i ; ⁇ i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j bones themselves, Is the rotating part of the deformation matrix.
  • another embodiment of the present invention provides a three-dimensional reconstruction apparatus combining rigid motion and non-rigid deformation, comprising: a photographing module for performing depth camera-based photographing on a target object to obtain a single depth
  • An image extraction module is configured to perform three-dimensional skeleton extraction on the depth point cloud by using a three-dimensional skeleton extraction algorithm; and the matching module converts the single depth image into a three-dimensional point cloud, and acquires between the three-dimensional point cloud and the reconstructed model vertex a matching point pair;
  • a solution module configured to establish an energy function according to the pair of matching points and the three-dimensional skeleton information, and solve a non-rigid motion position transformation parameter of each vertex on the reconstruction model and optimize an object skeleton parameter;
  • performing GPU optimization on the energy function to obtain a non-rigid deformation of each surface vertex, and deforming the reconstructed three-dimensional model of the previous frame according to the solution result, so that the deformation model is aligned with the current
  • the joint rigid motion and non-rigid deformation three-dimensional reconstruction device of the embodiment of the invention combines the three-dimensional information of the dynamic object surface frame by frame by real-time non-rigid alignment method, in order to achieve robust tracking, realizing the three-dimensional frame without the first frame Robust real-time dynamic 3D reconstruction under template conditions can effectively improve the real-time, robustness and accuracy of reconstruction, and it is scalable and easy to implement.
  • the combined rigid motion and non-rigid deformation three-dimensional reconstruction apparatus may further have the following additional technical features:
  • the matching module is further configured to project the single depth image into a three-dimensional space by using an internal parameter matrix of a depth camera to generate the three-dimensional point cloud, wherein the depth map
  • the projection formula is:
  • u, v are pixel coordinates
  • d(u, v) is a depth value at a pixel (u, v) position on the depth image
  • the energy function is:
  • E t is the total energy term
  • E n is the non-rigid surface deformation constraint term
  • E s is the rigid skeleton motion constraint term
  • E j is the rigid skeleton recognition constraint term
  • E g is the local rigid motion constraint term
  • ⁇ n , ⁇ s , ⁇ j and ⁇ g are weight coefficients corresponding to respective constraint terms, respectively.
  • u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair
  • c i represents the i-th element in the set of matching point pairs
  • the non-rigid surface deformation constraint item with Representing the vertex coordinates of the model and its normal direction driven by the non-rigid deformation, respectively, in the rigid skeleton motion constraint with Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton, with Representing the model vertex coordinates driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation, where i represents the i-th vertex on the model.
  • the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
  • a deformation matrix acting on the vertex v i including two parts of rotation and translation; Is the rotating portion of the deformation matrix; a set of bones that have a driving effect on the vertex v i ; ⁇ i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j skeletons themselves, rot(T bj ) is the rotation part of the deformation matrix.
  • FIG. 1 is a flow chart of a three-dimensional reconstruction method for joint rigid motion and non-rigid deformation according to an embodiment of the present invention
  • FIG. 2 is a flow chart of a three-dimensional reconstruction method combining joint rigid motion and non-rigid deformation according to an embodiment of the present invention
  • FIG. 3 is a schematic structural view of a three-dimensional reconstruction apparatus combining rigid motion and non-rigid deformation according to an embodiment of the present invention.
  • a three-dimensional reconstruction method and apparatus for joint rigid motion and non-rigid deformation according to an embodiment of the present invention will be described below with reference to the accompanying drawings.
  • a three-dimensional reconstruction method for joint rigid motion and non-rigid deformation according to an embodiment of the present invention will be described with reference to the accompanying drawings. .
  • FIG. 1 is a flow chart of a three-dimensional reconstruction method for joint rigid motion and non-rigid deformation according to an embodiment of the present invention.
  • the three-dimensional reconstruction method of the joint rigid motion and the non-rigid deformation includes the following steps:
  • step S101 depth camera-based photographing is performed on the target object to obtain a single depth image.
  • the real-time video frame rate depth point cloud acquisition is performed, and the dynamic object is subjected to depth map shooting to obtain a frame-by-frame depth point cloud.
  • a dynamic object is photographed using a depth camera to obtain a continuous single depth image sequence. Transform a single depth image into a set of 3D point clouds.
  • step S102 the depth point cloud is subjected to three-dimensional skeleton extraction by a three-dimensional skeleton extraction algorithm.
  • the 3D skeleton extraction is performed by the skeleton recognition algorithm, and the three-dimensional rigid skeleton information of the current frame of the object is extracted by the existing skeleton recognition algorithm.
  • the object 3D skeleton stealing is implemented by KinectSDK.
  • step S103 the single depth image is transformed into a three-dimensional point cloud, and a matching point pair between the three-dimensional point cloud and the reconstructed model vertex is acquired.
  • converting the single depth image into the three-dimensional point cloud further includes: projecting the single depth image into the three-dimensional space by using an internal parameter matrix of the depth camera to generate a three-dimensional point cloud, wherein
  • the depth map projection formula is:
  • u, v are pixel coordinates
  • d(u, v) is a depth value at a pixel (u, v) position on the depth image
  • the object is photographed by the depth camera to obtain a depth image, and the depth map is transformed into a set of three-dimensional point clouds.
  • the depth map is projected into the three-dimensional space to generate a set of three-dimensional point clouds.
  • the depth map projection formula is:
  • u, v are pixel coordinates
  • d(u, v) is a depth value at a pixel (u, v) position on the depth image
  • the internal parameter matrix of the depth camera is acquired, and the depth map is projected into the three-dimensional space according to the internal reference matrix and transformed into a set of three-dimensional point clouds.
  • the formula of the transformation is: Where u, v are pixel coordinates, and d(u, v) is a depth value at a pixel (u, v) position on the depth image,
  • u, v are pixel coordinates
  • d(u, v) is a depth value at a pixel (u, v) position on the depth image
  • the vertices of the three-dimensional model are projected onto the depth image using a camera projection formula to obtain matching point pairs.
  • the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
  • a deformation matrix acting on the vertex v i including two parts of rotation and translation; Is the rotating portion of the deformation matrix; a set of bones that have a driving effect on the vertex v i ; ⁇ i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j skeletons themselves, rot(T bj ) is the rotation part of the deformation matrix.
  • step S104 an energy function is established according to the matching point pair and the three-dimensional skeleton information, and the non-rigid motion position transformation parameters of each vertex on the reconstruction model are solved and the object skeleton parameters are optimized.
  • the energy function is established, and an energy function is established according to the current frame matching point pair information and the extracted three-dimensional rigid skeleton information of the current frame.
  • a single depth camera such as a Microsoft Kinect depth camera, an IphoneX depth camera, an Obi immersion depth camera, etc., captures a dynamic scene, and obtains real-time depth image data (video frame rate, 20 frames/second or more) transmitted to the computer.
  • the three-dimensional geometric information of the dynamic object is calculated by the computer in real time, the three-dimensional model of the object at the same frame rate is reconstructed, and the three-dimensional skeleton information of the object is output.
  • the energy function is:
  • E t is the total energy term
  • E n is the non-rigid surface deformation constraint term
  • E s is the rigid skeleton motion constraint term
  • E j is the rigid skeleton recognition constraint term
  • E g is the local rigid motion constraint term
  • ⁇ n , ⁇ s , ⁇ j and ⁇ g are weight coefficients corresponding to respective constraint terms, respectively.
  • u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair
  • c i represents the i-th element in the matching point pair set
  • the non-rigid surface deformation constraint item with Representing the vertex coordinates of the model and its normal direction driven by non-rigid deformation, respectively, in the rigid skeleton motion constraint with Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton, with Representing the vertex coordinates of the model driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation.
  • i represents the i-th vertex on the model.
  • the rigid motion constraint term E s and the non-rigid motion constraint term E n are simultaneously used to perform an optimal solution of the object motion, and a single depth image is used to perform the object rigid skeleton constraint term E j to constrain the solved rigid motion.
  • the surface non-rigid constraint E n ensures that the model after the non-rigid deformation is aligned with the three-dimensional point cloud obtained from the depth map as much as possible; with Representing the vertex coordinates of the model and its normal direction after being driven by non-rigid deformation, with Represents the vertex coordinates of the model and its normal direction driven by the object skeleton motion.
  • the rigid skeleton motion constraint E s ensures that the rigid deformation model driven by the skeleton motion is aligned with the three-dimensional point cloud obtained from the depth map as much as possible.
  • i represents the i-th vertex on the model
  • It represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j , that is, to ensure that the non-rigid driving effects of adjacent vertices on the model are as uniform as possible.
  • Is a robust penalty function Representing the driving effect of the rigid skeleton motion on the surface vertices v i and v j of the model respectively.
  • the value of the robust penalty function is small, when the two phases are When the neighboring vertex is less affected by the skeleton motion driving effect, the robust penalty function value is larger.
  • the robust penalty function the model can be subjected to local rigid constrained motion while ensuring a large amplitude of reasonable non-rigid motion. Can be well solved, so that the model is more accurately aligned with the 3D point cloud.
  • step S105 GPU optimization is performed on the energy function to obtain a non-rigid deformation of each surface vertex, and the reconstructed three-dimensional model of the previous frame is deformed according to the solution result, so that the deformation model is aligned with the current frame three-dimensional point cloud.
  • the GPU optimization of the energy function is performed to solve the non-rigid motion position transformation parameters of each vertex on the reconstructed model, and the three-dimensional rigid motion information of the object is optimized, and the previous frame is reconstructed according to the solution result.
  • the model is deformed to align with the current frame 3D point cloud.
  • the energy function is solved, and the reconstructed model is aligned with the three-dimensional point cloud according to the solution result.
  • the non-rigid motion position transformation parameters and the object skeleton motion parameters of each vertex on the reconstruction model are solved.
  • the information obtained by the final solution is the transformation matrix of each 3D model vertex and the object skeleton motion parameters, that is, the individual transformation matrix of each bone.
  • the method of the embodiment of the present invention approximates the deformation equation by using an exponential mapping method:
  • I is a four-dimensional unit matrix
  • the linearization of bone movement is the same as that of non-rigid motion.
  • step S106 the current frame three-dimensional point cloud and the deformation model are merged to obtain an updated model of the current frame to enter an iteration of the next frame.
  • Poisson fusion is performed on the aligned model and the point cloud to obtain a relatively complete three-dimensional model of the new frame.
  • the point cloud and the three-dimensional model are merged to obtain an updated model of the current frame.
  • Update and complete the 3D model aligned with the depth point cloud fuse the newly obtained depth information into the 3D model, update the surface vertex position of the 3D model or add new vertices to the 3D model to make it more consistent with the current depth image expression.
  • the core function of the embodiment of the present invention is to receive a depth image code stream in real time and calculate a three-dimensional model of each frame in real time.
  • the time-varying three-dimensional model of the dynamic object is calculated by using the large-scale rigid skeleton motion of the object and the small-scale surface non-rigid deformation information.
  • the method of the embodiment of the invention is accurate and can realize high-precision reconstruction of the dynamic object in real time. Since the method is a real-time reconstruction method and only needs to provide a single depth camera input, the system has the advantages of simple equipment, convenient deployment and scalability, and the like.
  • the required input information is very easy to acquire and a dynamic 3D model can be obtained in real time.
  • the method is accurate, robust, simple and easy to operate, and runs at a real-time speed. It has broad application prospects and can be quickly implemented on hardware systems such as PC (personal computer) or workstations.
  • the three-dimensional information of the dynamic object surface is merged frame by frame by real-time non-rigid alignment method, in order to achieve robust tracking, realizing the key without the first frame Robust real-time dynamic 3D reconstruction under frame 3D template conditions can effectively improve the real-time, robustness and accuracy of reconstruction, and it is scalable and easy to implement.
  • FIG. 3 is a schematic view showing the structure of a three-dimensional reconstruction apparatus combining rigid motion and non-rigid deformation according to an embodiment of the present invention.
  • the joint rigid motion and non-rigid deformation 3D reconstruction apparatus 10 includes a photographing module 100, an extraction module 200, a matching module 300, a solution module 400, a solution module 500, and a model update module 600.
  • the shooting module 100 is configured to perform depth camera-based shooting on the target object to obtain a single depth image.
  • the extraction module 200 is configured to perform three-dimensional skeleton extraction on the depth point cloud by using a three-dimensional skeleton extraction algorithm.
  • the matching module 300 transforms the single depth image into a three-dimensional point cloud, and acquires a matching point pair between the three-dimensional point cloud and the reconstructed model vertex.
  • the solving module 400 is configured to establish an energy function according to the matching point pair and the three-dimensional skeleton information, and solve the non-rigid motion position transformation parameter of each vertex on the reconstruction model and optimize the object skeleton parameter.
  • the solving module 500 is configured to perform GPU optimization on the energy function to obtain a non-rigid deformation of each surface vertex, and deform the reconstructed three-dimensional model of the previous frame according to the solution result, so that the deformation model is aligned with the current frame three-dimensional point cloud.
  • the model update module 600 is configured to fuse the current frame three-dimensional point cloud and the deformation model to obtain an updated model of the current frame to enter an iteration of the next frame.
  • the device 10 of the embodiment of the invention can effectively improve the real-time, robustness and accuracy of the reconstruction, has strong scalability, and is simple and easy to implement.
  • the matching module 300 is further configured to project a single depth image into the three-dimensional space by using an internal parameter matrix of the depth camera to generate a three-dimensional point cloud, wherein the depth map projection formula is:
  • u, v are pixel coordinates
  • d(u, v) is a depth value at a pixel (u, v) position on the depth image
  • the energy function is:
  • E t is the total energy term
  • E n is the non-rigid surface deformation constraint term
  • E s is the rigid skeleton motion constraint term
  • E j is the rigid skeleton recognition constraint term
  • E g is the local rigid motion constraint term
  • ⁇ n , ⁇ s , ⁇ j and ⁇ g are weight coefficients corresponding to respective constraint terms, respectively.
  • u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair
  • c i represents the i-th element in the matching point pair set
  • the non-rigid surface deformation constraint item with Representing the vertex coordinates of the model and its normal direction driven by non-rigid deformation, respectively, in the rigid skeleton motion constraint with Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton, with Representing the vertex coordinates of the model driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation.
  • i represents the i-th vertex on the model.
  • the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
  • a deformation matrix acting on the vertex v i including two parts of rotation and translation; Is the rotating portion of the deformation matrix; a set of bones that have a driving effect on the vertex v i ; ⁇ i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j skeletons themselves, rot(T bj ) is the rotation part of the deformation matrix.
  • the three-dimensional information of the dynamic object surface is merged frame by frame by real-time non-rigid alignment method, in order to achieve robust tracking, the key in the first frame is realized.
  • Robust real-time dynamic 3D reconstruction under frame 3D template conditions can effectively improve the real-time, robustness and accuracy of reconstruction, and it is scalable and easy to implement.
  • first and second are used for descriptive purposes only and are not to be construed as indicating or implying a relative importance or implicitly indicating the number of technical features indicated.
  • features defining “first” or “second” may include at least one of the features, either explicitly or implicitly.
  • the meaning of "a plurality” is at least two, such as two, three, etc., unless specifically defined otherwise.
  • the terms “installation”, “connected”, “connected”, “fixed” and the like shall be understood broadly, and may be either a fixed connection or a detachable connection, unless explicitly stated and defined otherwise. , or integrated; can be mechanical or electrical connection; can be directly connected, or indirectly connected through an intermediate medium, can be the internal communication of two elements or the interaction of two elements, unless otherwise specified Limited.
  • the specific meanings of the above terms in the present invention can be understood on a case-by-case basis.
  • the first feature "on” or “under” the second feature may be a direct contact of the first and second features, or the first and second features may be indirectly through an intermediate medium, unless otherwise explicitly stated and defined. contact.
  • the first feature "above”, “above” and “above” the second feature may be that the first feature is directly above or above the second feature, or merely that the first feature level is higher than the second feature.
  • the first feature “below”, “below” and “below” the second feature may be that the first feature is directly below or obliquely below the second feature, or merely that the first feature level is less than the second feature.

Abstract

A three-dimensional reconstruction method and device uniting a rigid motion and a non-rigid deformation. The method comprises: performing a depth camera-based photography on a target object to obtain a single depth image (S101); performing three-dimensional framework extraction on a depth point cloud by means of a three-dimensional framework extraction algorithm (S102); obtaining a matching point pair between a three-dimensional point cloud and vertexes of a reconstruction model; establishing an energy function according to the matching point pair and three-dimensional framework information, solving a non-rigid motion position conversion parameter of each vertex on the reconstruction model, and optimizing an object framework parameter (S104); performing GPU optimal solution on the energy function to obtain the non-rigid deformation of each surface vertex, and deforming the previous frame of reconstruction three-dimensional model according to the solving result, so that the deformed model is aligned with the current frame of three-dimensional point cloud (S105); and obtaining the current frame of updated model to enter iteration of the next frame. The method can effectively improve the real-time performance, robustness, and accuracy of reconstruction, is high in expandability, and simple and easy to be implemented.

Description

联合刚性运动和非刚性形变的三维重建方法及装置Three-dimensional reconstruction method and device for joint rigid motion and non-rigid deformation
相关申请的交叉引用Cross-reference to related applications
本申请要求清华大学于2018年05月15日提交的、发明名称为“联合刚性运动和非刚性形变的三维重建方法及装置”的、中国专利申请号“201810460091.5”的优先权。The present application claims the priority of the Chinese Patent Application No. 201810460091.5, filed on May 15, 2018, the entire disclosure of which is incorporated herein by reference.
技术领域Technical field
本发明涉及计算机视觉和计算机图形学技术领域,特别涉及一种联合刚性运动和非刚性形变的三维重建方法及装置。The invention relates to the technical field of computer vision and computer graphics, in particular to a three-dimensional reconstruction method and device for joint rigid motion and non-rigid deformation.
背景技术Background technique
动态对象三维重建是计算机图形学和计算机视觉领域的重点问题。高质量的动态对象三维模型,如人体,动物,人脸,人手部等,在影视娱乐、体育游戏、虚拟现实等领域有着广泛的应用前景和重要的应用价值。但是高质量三维模型的获取通常依靠价格昂贵的激光扫描仪或者多相机阵列系统来实现,虽然精度较高,但是也显著存在着一些缺点:第一,扫描过程中要求对象保持绝对静止,微小的移动就会导致扫描结果存在明显的误差;第二,造假昂贵,很难普及到普通民众日常生活中,往往应用于大公司或国家统计部门。第三,速度慢,往往重建一个三维模型需要至少10分钟到数小时的时间,重建动态模型序列的代价更大。Dynamic object 3D reconstruction is a key issue in the field of computer graphics and computer vision. High-quality dynamic object 3D models, such as human body, animal, human face, human hand, etc., have broad application prospects and important application value in the fields of film and television entertainment, sports games, virtual reality and so on. However, the acquisition of high-quality 3D models usually relies on expensive laser scanners or multi-camera array systems. Although the accuracy is high, there are also some shortcomings: First, the object is required to remain absolutely still during the scanning process. Movement will lead to obvious errors in the scanning results. Second, the fraud is expensive and difficult to spread to the daily lives of ordinary people, often applied to large companies or national statistical departments. Third, the speed is slow, and it often takes at least 10 minutes to several hours to reconstruct a 3D model. The cost of reconstructing a dynamic model sequence is greater.
从技术角度,现有的重建方法要么集中在先求解对象的刚性运动信息,获得对象的逼近,进而重建非刚性表面运动信息。但这种重建方法需要事先获得对象的关键帧三维模型。另一方面,现有的逐帧动态融合表面的重建方法虽然可实现无模板的动态三维重建,但仅仅使用非刚性表面形变方法,跟踪重建的鲁棒性低。From a technical point of view, the existing reconstruction method concentrates on solving the rigid motion information of the object first, obtaining the approximation of the object, and reconstructing the non-rigid surface motion information. However, this reconstruction method requires obtaining a three-dimensional model of the key frame of the object in advance. On the other hand, although the existing frame-by-frame dynamic fusion surface reconstruction method can realize dynamic three-dimensional reconstruction without template, the robustness of tracking reconstruction is low only by using the non-rigid surface deformation method.
发明内容Summary of the invention
本发明旨在至少在一定程度上解决相关技术中的技术问题之一。The present invention aims to solve at least one of the technical problems in the related art to some extent.
为此,本发明的一个目的在于提出一种联合刚性运动和非刚性形变的三维重建方法,该方法可以有效提高重建的实时性、鲁棒性和准确性,扩展性强,简单易实现。Therefore, an object of the present invention is to propose a three-dimensional reconstruction method combining joint rigid motion and non-rigid deformation, which can effectively improve the real-time, robustness and accuracy of reconstruction, and has strong scalability and is easy to implement.
本发明的另一个目的在于提出一种联合刚性运动和非刚性形变的三维重建装置。Another object of the present invention is to provide a three-dimensional reconstruction apparatus that combines rigid motion and non-rigid deformation.
为达到上述目的,本发明一方面实施例提出了一种联合刚性运动和非刚性形变的三维重建方法,包括以下步骤:对目标对象进行基于深度相机的拍摄,以得到单张深度图像; 通过三维骨架提取算法对深度点云进行三维骨架提取;将所述单张深度图像变换为三维点云,并获取所述三维点云和重建模型顶点之间的匹配点对;根据所述匹配点对和三维骨架信息建立能量函数,并求解所述重建模型上每一个顶点的非刚性运动位置变换参数并优化对象骨架参数;对所述能量函数进行GPU(Graphics Processing Unit,图形处理器)优化求解,以获得每个表面顶点的非刚性形变,根据求解结果将前一帧的重建三维模型进行形变,使得形变模型与当前帧三维点云进行对齐;融合当前帧三维点云与所述形变模型,以获得当前帧的更新后的模型,以进入下一帧的迭代。In order to achieve the above object, an embodiment of the present invention provides a three-dimensional reconstruction method combining joint rigid motion and non-rigid deformation, including the following steps: performing depth camera-based photographing on a target object to obtain a single depth image; The skeleton extraction algorithm performs three-dimensional skeleton extraction on the depth point cloud; transforms the single depth image into a three-dimensional point cloud, and acquires a matching point pair between the three-dimensional point cloud and the reconstructed model vertex; and according to the matching point pair The three-dimensional skeleton information establishes an energy function, and solves a non-rigid motion position transformation parameter of each vertex on the reconstruction model and optimizes an object skeleton parameter; performing a GPU (Graphics Processing Unit) optimization solution on the energy function to Obtaining a non-rigid deformation of each surface vertex, deforming the reconstructed three-dimensional model of the previous frame according to the solution result, so that the deformation model is aligned with the current frame three-dimensional point cloud; and the current frame three-dimensional point cloud and the deformation model are merged to obtain The updated model of the current frame to enter the iteration of the next frame.
本发明实施例的联合刚性运动和非刚性形变的三维重建方法,通过实时非刚性对齐的方法,逐帧地融合动态对象表面三维信息,为了实现鲁棒地跟踪,实现在无首帧关键帧三维模板条件下的鲁棒性实时动态三维重建,从而可以有效提高重建的实时性、鲁棒性和准确性,扩展性强,简单易实现。The three-dimensional reconstruction method of the joint rigid motion and the non-rigid deformation according to the embodiment of the present invention combines the three-dimensional information of the dynamic object surface frame by frame by real-time non-rigid alignment method, in order to achieve robust tracking, realizing the three-dimensional frame without the first frame Robust real-time dynamic 3D reconstruction under template conditions can effectively improve the real-time, robustness and accuracy of reconstruction, and it is scalable and easy to implement.
另外,根据本发明上述实施例的联合刚性运动和非刚性形变的三维重建方法还可以具有以下附加的技术特征:In addition, the three-dimensional reconstruction method of the joint rigid motion and the non-rigid deformation according to the above-described embodiments of the present invention may further have the following additional technical features:
进一步地,在本发明的一个实施例中,所述将所述单张深度图像变换为三维点云,进一步包括:通过深度相机的内参矩阵将所述单张深度图像投影到三维空间中,以生成所述三维点云,其中,深度图投影公式为:Further, in an embodiment of the present invention, the converting the single depth image into a three-dimensional point cloud further comprises: projecting the single depth image into the three-dimensional space by using an internal parameter matrix of the depth camera, Generating the three-dimensional point cloud, wherein the depth map projection formula is:
Figure PCTCN2019086889-appb-000001
Figure PCTCN2019086889-appb-000001
其中,u,v为像素坐标,d(u,v)为深度图像上像素(u,v)位置处的深度值,
Figure PCTCN2019086889-appb-000002
为所述深度相机的内参矩阵。
Where u, v are pixel coordinates, and d(u, v) is a depth value at a pixel (u, v) position on the depth image,
Figure PCTCN2019086889-appb-000002
Is the internal reference matrix of the depth camera.
进一步地,在本发明的一个实施例中,所述能量函数为:Further, in an embodiment of the invention, the energy function is:
E t=λ nE nsE sjE jgE gbE bE tn E ns E sj E jg E gb E b ,
其中,E t为总能量项,E n为非刚性表面形变约束项,E s为刚性骨架运动约束项,E j为刚性骨架识别约束项,E g为局部刚性运动约束项,λ n、λ s、λ j和λ g分别为对应各个约束项的权重系数。 Where E t is the total energy term, E n is the non-rigid surface deformation constraint term, E s is the rigid skeleton motion constraint term, E j is the rigid skeleton recognition constraint term, E g is the local rigid motion constraint term, λ n , λ s , λ j and λ g are weight coefficients corresponding to respective constraint terms, respectively.
进一步地,在本发明的一个实施例中,其中,Further, in an embodiment of the present invention, wherein
Figure PCTCN2019086889-appb-000003
Figure PCTCN2019086889-appb-000003
Figure PCTCN2019086889-appb-000004
Figure PCTCN2019086889-appb-000004
Figure PCTCN2019086889-appb-000005
Figure PCTCN2019086889-appb-000005
Figure PCTCN2019086889-appb-000006
Figure PCTCN2019086889-appb-000006
其中,u i表示同一匹配点对中三维点云的位置坐标,c i表示匹配点对集合中的第i个元素,所述非刚性表面形变约束项中
Figure PCTCN2019086889-appb-000007
Figure PCTCN2019086889-appb-000008
分别表示经过非刚性形变驱动后的模型顶点坐标及其法向,所述刚性骨架运动约束项中
Figure PCTCN2019086889-appb-000009
Figure PCTCN2019086889-appb-000010
分别表示经过对象骨架运动驱动后的模型顶点坐标及其法向,
Figure PCTCN2019086889-appb-000011
Figure PCTCN2019086889-appb-000012
分别代表受目标刚性运动驱动后的模型顶点坐标与受三维骨架估计所得到的运动驱动后的模型顶点坐标,所述局部刚性运动约束项中,i表示模型上第i个顶点,
Figure PCTCN2019086889-appb-000013
表示模型上第i个顶点周围的邻近顶点的集合,
Figure PCTCN2019086889-appb-000014
Figure PCTCN2019086889-appb-000015
分别代表已知非刚性运动对模型表面顶点v i和v j的驱动作用,
Figure PCTCN2019086889-appb-000016
Figure PCTCN2019086889-appb-000017
代表作用在v i和v j上的非刚性运动同时作用在v j上的位置变换效果。
Where u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair, c i represents the i-th element in the set of matching point pairs, and the non-rigid surface deformation constraint item
Figure PCTCN2019086889-appb-000007
with
Figure PCTCN2019086889-appb-000008
Representing the vertex coordinates of the model and its normal direction driven by the non-rigid deformation, respectively, in the rigid skeleton motion constraint
Figure PCTCN2019086889-appb-000009
with
Figure PCTCN2019086889-appb-000010
Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton,
Figure PCTCN2019086889-appb-000011
with
Figure PCTCN2019086889-appb-000012
Representing the model vertex coordinates driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation, where i represents the i-th vertex on the model.
Figure PCTCN2019086889-appb-000013
Represents a collection of adjacent vertices around the ith vertex on the model,
Figure PCTCN2019086889-appb-000014
with
Figure PCTCN2019086889-appb-000015
Representing the driving effects of known non-rigid motion on the surface vertices v i and v j of the model, respectively.
Figure PCTCN2019086889-appb-000016
with
Figure PCTCN2019086889-appb-000017
Represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j .
进一步地,在本发明的一个实施例中,根据表面非刚性形变和对象刚性骨架运动驱动模型顶点,其中,计算公式为:Further, in one embodiment of the invention, the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
Figure PCTCN2019086889-appb-000018
Figure PCTCN2019086889-appb-000018
Figure PCTCN2019086889-appb-000019
Figure PCTCN2019086889-appb-000019
其中,
Figure PCTCN2019086889-appb-000020
为作用于顶点v i的变形矩阵,包括旋转和平移两部分;
Figure PCTCN2019086889-appb-000021
为该变形矩阵的旋转部分;
Figure PCTCN2019086889-appb-000022
为对顶点v i有驱动作用的骨骼的集合;α i,j为第j个骨骼对第i个模型顶点的驱动作用的权重,表示该骨骼对该顶点驱动作用的强弱;T bj为第j个骨骼自身的运动变形矩阵,
Figure PCTCN2019086889-appb-000023
为该变形矩阵的旋转部分。
among them,
Figure PCTCN2019086889-appb-000020
a deformation matrix acting on the vertex v i , including two parts of rotation and translation;
Figure PCTCN2019086889-appb-000021
Is the rotating portion of the deformation matrix;
Figure PCTCN2019086889-appb-000022
a set of bones that have a driving effect on the vertex v i ; α i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j bones themselves,
Figure PCTCN2019086889-appb-000023
Is the rotating part of the deformation matrix.
为达到上述目的,本发明另一方面实施例提出了一种联合刚性运动和非刚性形变的三维重建装置,包括:拍摄模块,用于对目标对象进行基于深度相机的拍摄,以得到单张深度图像;提取模块,用于通过三维骨架提取算法对深度点云进行三维骨架提取;匹配模块,将所述单张深度图像变换为三维点云,并获取所述三维点云和重建模型顶点之间的匹配点对;解算模块,用于根据所述匹配点对和三维骨架信息建立能量函数,并求解所述重建模型上每一个顶点的非刚性运动位置变换参数并优化对象骨架参数;求解模块,用于对所述能量函数进行GPU优化求解,以获得每个表面顶点的非刚性形变,并根据求解结果将前一帧的重建三维模型进行形变,使得形变模型与当前帧三维点云进行对齐;模型更新模块,用于融合当前帧三维点云与所述形变模型,以获得当前帧的更新后的模型,以进入下一帧的迭代。In order to achieve the above object, another embodiment of the present invention provides a three-dimensional reconstruction apparatus combining rigid motion and non-rigid deformation, comprising: a photographing module for performing depth camera-based photographing on a target object to obtain a single depth An image extraction module is configured to perform three-dimensional skeleton extraction on the depth point cloud by using a three-dimensional skeleton extraction algorithm; and the matching module converts the single depth image into a three-dimensional point cloud, and acquires between the three-dimensional point cloud and the reconstructed model vertex a matching point pair; a solution module, configured to establish an energy function according to the pair of matching points and the three-dimensional skeleton information, and solve a non-rigid motion position transformation parameter of each vertex on the reconstruction model and optimize an object skeleton parameter; And performing GPU optimization on the energy function to obtain a non-rigid deformation of each surface vertex, and deforming the reconstructed three-dimensional model of the previous frame according to the solution result, so that the deformation model is aligned with the current frame three-dimensional point cloud. a model update module for fusing a current frame three-dimensional point cloud and the deformation model to obtain Updated model frame to the next iteration into the frame.
本发明实施例的联合刚性运动和非刚性形变的三维重建装置,通过实时非刚性对齐的方法,逐帧地融合动态对象表面三维信息,为了实现鲁棒地跟踪,实现在无首帧关键帧三 维模板条件下的鲁棒性实时动态三维重建,从而可以有效提高重建的实时性、鲁棒性和准确性,扩展性强,简单易实现。The joint rigid motion and non-rigid deformation three-dimensional reconstruction device of the embodiment of the invention combines the three-dimensional information of the dynamic object surface frame by frame by real-time non-rigid alignment method, in order to achieve robust tracking, realizing the three-dimensional frame without the first frame Robust real-time dynamic 3D reconstruction under template conditions can effectively improve the real-time, robustness and accuracy of reconstruction, and it is scalable and easy to implement.
另外,根据本发明上述实施例的联合刚性运动和非刚性形变的三维重建装置还可以具有以下附加的技术特征:In addition, the combined rigid motion and non-rigid deformation three-dimensional reconstruction apparatus according to the above-described embodiments of the present invention may further have the following additional technical features:
进一步地,在本发明的一个实施例中,所述匹配模块进一步用于通过深度相机的内参矩阵将所述单张深度图像投影到三维空间中,以生成所述三维点云,其中,深度图投影公式为:Further, in an embodiment of the present invention, the matching module is further configured to project the single depth image into a three-dimensional space by using an internal parameter matrix of a depth camera to generate the three-dimensional point cloud, wherein the depth map The projection formula is:
Figure PCTCN2019086889-appb-000024
Figure PCTCN2019086889-appb-000024
其中,u,v为像素坐标,d(u,v)为深度图像上像素(u,v)位置处的深度值,
Figure PCTCN2019086889-appb-000025
为所述深度相机的内参矩阵。
Where u, v are pixel coordinates, and d(u, v) is a depth value at a pixel (u, v) position on the depth image,
Figure PCTCN2019086889-appb-000025
Is the internal reference matrix of the depth camera.
进一步地,在本发明的一个实施例中,所述能量函数为:Further, in an embodiment of the invention, the energy function is:
E t=λ nE nsE sjE jgE gbE bE tn E ns E sj E jg E gb E b ,
其中,E t为总能量项,E n为非刚性表面形变约束项,E s为刚性骨架运动约束项,E j为刚性骨架识别约束项,E g为局部刚性运动约束项,λ n、λ s、λ j和λ g分别为对应各个约束项的权重系数。 Where E t is the total energy term, E n is the non-rigid surface deformation constraint term, E s is the rigid skeleton motion constraint term, E j is the rigid skeleton recognition constraint term, E g is the local rigid motion constraint term, λ n , λ s , λ j and λ g are weight coefficients corresponding to respective constraint terms, respectively.
进一步地,在本发明的一个实施例中,其中,Further, in an embodiment of the present invention, wherein
Figure PCTCN2019086889-appb-000026
Figure PCTCN2019086889-appb-000026
Figure PCTCN2019086889-appb-000027
Figure PCTCN2019086889-appb-000027
Figure PCTCN2019086889-appb-000028
Figure PCTCN2019086889-appb-000028
Figure PCTCN2019086889-appb-000029
Figure PCTCN2019086889-appb-000029
其中,u i表示同一匹配点对中三维点云的位置坐标,c i表示匹配点对集合中的第i个元素,所述非刚性表面形变约束项中
Figure PCTCN2019086889-appb-000030
Figure PCTCN2019086889-appb-000031
分别表示经过非刚性形变驱动后的模型顶点坐标及其法向,所述刚性骨架运动约束项中
Figure PCTCN2019086889-appb-000032
Figure PCTCN2019086889-appb-000033
分别表示经过对象骨架运动驱动后的模型顶点坐标及其法向,
Figure PCTCN2019086889-appb-000034
Figure PCTCN2019086889-appb-000035
分别代表受目标刚性运动驱动后的模型顶点坐标与受三维骨架估计所得到的运动驱动后的模型顶点坐标,所述局部刚性运动约束项中,i表示模型上第i个顶点,
Figure PCTCN2019086889-appb-000036
表示模型上第i个顶点周围的邻近顶点的集合,
Figure PCTCN2019086889-appb-000037
Figure PCTCN2019086889-appb-000038
分别代表已知非刚性运动对模型表面顶点v i和v j的驱动作用,
Figure PCTCN2019086889-appb-000039
Figure PCTCN2019086889-appb-000040
代表作用在v i和v j上的非刚性运动同时作用在v j上的位置变换效果。
Where u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair, c i represents the i-th element in the set of matching point pairs, and the non-rigid surface deformation constraint item
Figure PCTCN2019086889-appb-000030
with
Figure PCTCN2019086889-appb-000031
Representing the vertex coordinates of the model and its normal direction driven by the non-rigid deformation, respectively, in the rigid skeleton motion constraint
Figure PCTCN2019086889-appb-000032
with
Figure PCTCN2019086889-appb-000033
Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton,
Figure PCTCN2019086889-appb-000034
with
Figure PCTCN2019086889-appb-000035
Representing the model vertex coordinates driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation, where i represents the i-th vertex on the model.
Figure PCTCN2019086889-appb-000036
Represents a collection of adjacent vertices around the ith vertex on the model,
Figure PCTCN2019086889-appb-000037
with
Figure PCTCN2019086889-appb-000038
Representing the driving effects of known non-rigid motion on the surface vertices v i and v j of the model, respectively.
Figure PCTCN2019086889-appb-000039
with
Figure PCTCN2019086889-appb-000040
Represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j .
进一步地,在本发明的一个实施例中,根据表面非刚性形变和对象刚性骨架运动驱动模型顶点,其中,计算公式为:Further, in one embodiment of the invention, the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
Figure PCTCN2019086889-appb-000041
Figure PCTCN2019086889-appb-000041
Figure PCTCN2019086889-appb-000042
Figure PCTCN2019086889-appb-000042
其中,
Figure PCTCN2019086889-appb-000043
为作用于顶点v i的变形矩阵,包括旋转和平移两部分;
Figure PCTCN2019086889-appb-000044
为该变形矩阵的旋转部分;
Figure PCTCN2019086889-appb-000045
为对顶点v i有驱动作用的骨骼的集合;α i,j为第j个骨骼对第i个模型顶点的驱动作用的权重,表示该骨骼对该顶点驱动作用的强弱;T bj为第j个骨骼自身的运动变形矩阵,rot(T bj)为该变形矩阵的旋转部分。
among them,
Figure PCTCN2019086889-appb-000043
a deformation matrix acting on the vertex v i , including two parts of rotation and translation;
Figure PCTCN2019086889-appb-000044
Is the rotating portion of the deformation matrix;
Figure PCTCN2019086889-appb-000045
a set of bones that have a driving effect on the vertex v i ; α i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j skeletons themselves, rot(T bj ) is the rotation part of the deformation matrix.
本发明附加的方面和优点将在下面的描述中部分给出,部分将从下面的描述中变得明显,或通过本发明的实践了解到。The additional aspects and advantages of the invention will be set forth in part in the description which follows.
附图说明DRAWINGS
本发明上述的和/或附加的方面和优点从下面结合附图对实施例的描述中将变得明显和容易理解,其中:The above and/or additional aspects and advantages of the present invention will become apparent and readily understood from
图1为根据本发明一个实施例的联合刚性运动和非刚性形变的三维重建方法的流程图;1 is a flow chart of a three-dimensional reconstruction method for joint rigid motion and non-rigid deformation according to an embodiment of the present invention;
图2为根据本发明一个具体实施例的联合刚性运动和非刚性形变的三维重建方法的流程图;2 is a flow chart of a three-dimensional reconstruction method combining joint rigid motion and non-rigid deformation according to an embodiment of the present invention;
图3为根据本发明一个实施例的联合刚性运动和非刚性形变的三维重建装置的结构示意图。3 is a schematic structural view of a three-dimensional reconstruction apparatus combining rigid motion and non-rigid deformation according to an embodiment of the present invention.
具体实施方式Detailed ways
下面详细描述本发明的实施例,所述实施例的示例在附图中示出,其中自始至终相同或类似的标号表示相同或类似的元件或具有相同或类似功能的元件。下面通过参考附图描述的实施例是示例性的,旨在用于解释本发明,而不能理解为对本发明的限制。The embodiments of the present invention are described in detail below, and the examples of the embodiments are illustrated in the drawings, wherein the same or similar reference numerals are used to refer to the same or similar elements or elements having the same or similar functions. The embodiments described below with reference to the drawings are intended to be illustrative of the invention and are not to be construed as limiting.
下面参照附图描述根据本发明实施例提出的联合刚性运动和非刚性形变的三维重建方法及装置,首先将参照附图描述根据本发明实施例提出的联合刚性运动和非刚性形变的三维重建方法。A three-dimensional reconstruction method and apparatus for joint rigid motion and non-rigid deformation according to an embodiment of the present invention will be described below with reference to the accompanying drawings. First, a three-dimensional reconstruction method for joint rigid motion and non-rigid deformation according to an embodiment of the present invention will be described with reference to the accompanying drawings. .
图1是本发明一个实施例的联合刚性运动和非刚性形变的三维重建方法的流程图。1 is a flow chart of a three-dimensional reconstruction method for joint rigid motion and non-rigid deformation according to an embodiment of the present invention.
如图1所示,该联合刚性运动和非刚性形变的三维重建方法包括以下步骤:As shown in FIG. 1, the three-dimensional reconstruction method of the joint rigid motion and the non-rigid deformation includes the following steps:
在步骤S101中,对目标对象进行基于深度相机的拍摄,以得到单张深度图像。In step S101, depth camera-based photographing is performed on the target object to obtain a single depth image.
可以理解的是,如图2所示,实时视频帧率深度点云获取,对动态对象进行深度图拍摄以得到逐帧深度点云。具体地,使用深度相机对动态对象进行拍摄,获得连续的单张深度图像序列。将单张深度图像变换为一组三维点云。It can be understood that, as shown in FIG. 2, the real-time video frame rate depth point cloud acquisition is performed, and the dynamic object is subjected to depth map shooting to obtain a frame-by-frame depth point cloud. Specifically, a dynamic object is photographed using a depth camera to obtain a continuous single depth image sequence. Transform a single depth image into a set of 3D point clouds.
在步骤S102中,通过三维骨架提取算法对深度点云进行三维骨架提取。In step S102, the depth point cloud is subjected to three-dimensional skeleton extraction by a three-dimensional skeleton extraction algorithm.
可以理解的是,如图2所示,通过骨架识别算法进行3D骨架提取,通过已有的骨架识别算法提取对象当前帧的三维刚性骨架信息。例如,通过KinectSDK实现对象三维骨架偷取。It can be understood that, as shown in FIG. 2, the 3D skeleton extraction is performed by the skeleton recognition algorithm, and the three-dimensional rigid skeleton information of the current frame of the object is extracted by the existing skeleton recognition algorithm. For example, the object 3D skeleton stealing is implemented by KinectSDK.
在步骤S103中,将单张深度图像变换为三维点云,并获取三维点云和重建模型顶点之间的匹配点对。In step S103, the single depth image is transformed into a three-dimensional point cloud, and a matching point pair between the three-dimensional point cloud and the reconstructed model vertex is acquired.
可以理解的是,如图2所示,建立三维模型与云匹配点对,计算当前帧三维点云与已重建模型顶点之间的匹配点对。It can be understood that, as shown in FIG. 2, a three-dimensional model and a cloud matching point pair are established, and a matching point pair between the current frame three-dimensional point cloud and the reconstructed model vertex is calculated.
进一步地,在本发明的一个实施例中,将单张深度图像变换为三维点云,进一步包括:通过深度相机的内参矩阵将单张深度图像投影到三维空间中,以生成三维点云,其中,深度图投影公式为:Further, in an embodiment of the present invention, converting the single depth image into the three-dimensional point cloud further includes: projecting the single depth image into the three-dimensional space by using an internal parameter matrix of the depth camera to generate a three-dimensional point cloud, wherein The depth map projection formula is:
Figure PCTCN2019086889-appb-000046
Figure PCTCN2019086889-appb-000046
其中,u,v为像素坐标,d(u,v)为深度图像上像素(u,v)位置处的深度值,
Figure PCTCN2019086889-appb-000047
为深度相机的内参矩阵。
Where u, v are pixel coordinates, and d(u, v) is a depth value at a pixel (u, v) position on the depth image,
Figure PCTCN2019086889-appb-000047
The internal reference matrix for the depth camera.
可以理解的是,通过深度相机对对象拍摄以得到深度图像,将深度图变换为一组三维点云,基于深度相机标定的内参矩阵,将深度图投影到三维空间中生成一组三维点云。的深度图投影公式为:It can be understood that the object is photographed by the depth camera to obtain a depth image, and the depth map is transformed into a set of three-dimensional point clouds. Based on the internal camera matrix of the depth camera calibration, the depth map is projected into the three-dimensional space to generate a set of three-dimensional point clouds. The depth map projection formula is:
Figure PCTCN2019086889-appb-000048
Figure PCTCN2019086889-appb-000048
其中,u,v为像素坐标,d(u,v)为深度图像上像素(u,v)位置处的深度值,
Figure PCTCN2019086889-appb-000049
为深度相机内参矩阵。
Where u, v are pixel coordinates, and d(u, v) is a depth value at a pixel (u, v) position on the depth image,
Figure PCTCN2019086889-appb-000049
For the depth camera internal reference matrix.
具体而言,获取深度相机的内参矩阵,根据内参矩阵将深度图投影到三维空间中变换为一组三维点云。其中,变换的公式为:
Figure PCTCN2019086889-appb-000050
其中,u,v为像素坐标,d(u,v)为深度图像上像素(u,v)位置处的深度值,
Figure PCTCN2019086889-appb-000051
为深度相机内参矩阵。在获取匹配点对方面,使用相机投影公式将三维模型的顶点投影到深度图像上以获得匹配点对。
Specifically, the internal parameter matrix of the depth camera is acquired, and the depth map is projected into the three-dimensional space according to the internal reference matrix and transformed into a set of three-dimensional point clouds. Among them, the formula of the transformation is:
Figure PCTCN2019086889-appb-000050
Where u, v are pixel coordinates, and d(u, v) is a depth value at a pixel (u, v) position on the depth image,
Figure PCTCN2019086889-appb-000051
For the depth camera internal reference matrix. In terms of obtaining matching point pairs, the vertices of the three-dimensional model are projected onto the depth image using a camera projection formula to obtain matching point pairs.
进一步地,在本发明的一个实施例中,根据表面非刚性形变和对象刚性骨架运动驱动模型顶点,其中,计算公式为:Further, in one embodiment of the invention, the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
Figure PCTCN2019086889-appb-000052
Figure PCTCN2019086889-appb-000052
Figure PCTCN2019086889-appb-000053
Figure PCTCN2019086889-appb-000053
其中,
Figure PCTCN2019086889-appb-000054
为作用于顶点v i的变形矩阵,包括旋转和平移两部分;
Figure PCTCN2019086889-appb-000055
为该变形矩阵的旋转部分;
Figure PCTCN2019086889-appb-000056
为对顶点v i有驱动作用的骨骼的集合;α i,j为第j个骨骼对第i个模型顶点的驱动作用的权重,表示该骨骼对该顶点驱动作用的强弱;T bj为第j个骨骼自身的运动变形矩阵,rot(T bj)为该变形矩阵的旋转部分。
among them,
Figure PCTCN2019086889-appb-000054
a deformation matrix acting on the vertex v i , including two parts of rotation and translation;
Figure PCTCN2019086889-appb-000055
Is the rotating portion of the deformation matrix;
Figure PCTCN2019086889-appb-000056
a set of bones that have a driving effect on the vertex v i ; α i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j skeletons themselves, rot(T bj ) is the rotation part of the deformation matrix.
在步骤S104中,根据匹配点对和三维骨架信息建立能量函数,并求解重建模型上每一个顶点的非刚性运动位置变换参数并优化对象骨架参数。In step S104, an energy function is established according to the matching point pair and the three-dimensional skeleton information, and the non-rigid motion position transformation parameters of each vertex on the reconstruction model are solved and the object skeleton parameters are optimized.
可以理解的是,建立能量函数,根据当前帧匹配点对信息,以及提取的当前帧三维刚性骨架信息,建立能量函数。It can be understood that the energy function is established, and an energy function is established according to the current frame matching point pair information and the extracted three-dimensional rigid skeleton information of the current frame.
例如,采用单个深度相机,如微软Kinect深度相机、IphoneX深度相机、奥比中光深度相机等,对动态场景进行拍摄,获得实时深度图像数据(视频帧率,20帧/秒以上)传送到计算机上,由计算机实时计算动态对象的三维几何信息,重建相同帧率下的对象三维模型并输出对象的三维骨架信息。For example, a single depth camera, such as a Microsoft Kinect depth camera, an IphoneX depth camera, an Obi immersion depth camera, etc., captures a dynamic scene, and obtains real-time depth image data (video frame rate, 20 frames/second or more) transmitted to the computer. On the computer, the three-dimensional geometric information of the dynamic object is calculated by the computer in real time, the three-dimensional model of the object at the same frame rate is reconstructed, and the three-dimensional skeleton information of the object is output.
进一步地,在本发明的一个实施例中,能量函数为:Further, in one embodiment of the invention, the energy function is:
E t=λ nE nsE sjE jgE gbE bE tn E ns E sj E jg E gb E b ,
其中,E t为总能量项,E n为非刚性表面形变约束项,E s为刚性骨架运动约束项,E j为刚性骨架识别约束项,E g为局部刚性运动约束项,λ n、λ s、λ j和λ g分别为对应各个约束项的权重系数。 Where E t is the total energy term, E n is the non-rigid surface deformation constraint term, E s is the rigid skeleton motion constraint term, E j is the rigid skeleton recognition constraint term, E g is the local rigid motion constraint term, λ n , λ s , λ j and λ g are weight coefficients corresponding to respective constraint terms, respectively.
进一步地,在本发明的一个实施例中,其中,Further, in an embodiment of the present invention, wherein
Figure PCTCN2019086889-appb-000057
Figure PCTCN2019086889-appb-000057
Figure PCTCN2019086889-appb-000058
Figure PCTCN2019086889-appb-000058
Figure PCTCN2019086889-appb-000059
Figure PCTCN2019086889-appb-000059
Figure PCTCN2019086889-appb-000060
Figure PCTCN2019086889-appb-000060
其中,u i表示同一匹配点对中三维点云的位置坐标,c i表示匹配点对集合中的第i个元素,非刚性表面形变约束项中
Figure PCTCN2019086889-appb-000061
Figure PCTCN2019086889-appb-000062
分别表示经过非刚性形变驱动后的模型顶点坐标及其法向,刚性骨架运动约束项中
Figure PCTCN2019086889-appb-000063
Figure PCTCN2019086889-appb-000064
分别表示经过对象骨架运动驱动后的模型顶点坐标及其法向,
Figure PCTCN2019086889-appb-000065
Figure PCTCN2019086889-appb-000066
分别代表受目标刚性运动驱动后的模型顶点坐标与受三维骨架估计所得到的运动驱动后的模型顶点坐标,局部刚性运动约束项中,i表示模型上第i个顶点,
Figure PCTCN2019086889-appb-000067
表 示模型上第i个顶点周围的邻近顶点的集合,
Figure PCTCN2019086889-appb-000068
Figure PCTCN2019086889-appb-000069
分别代表已知非刚性运动对模型表面顶点v i和v j的驱动作用,
Figure PCTCN2019086889-appb-000070
Figure PCTCN2019086889-appb-000071
代表作用在v i和v j上的非刚性运动同时作用在v j上的位置变换效果。
Where u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair, c i represents the i-th element in the matching point pair set, and the non-rigid surface deformation constraint item
Figure PCTCN2019086889-appb-000061
with
Figure PCTCN2019086889-appb-000062
Representing the vertex coordinates of the model and its normal direction driven by non-rigid deformation, respectively, in the rigid skeleton motion constraint
Figure PCTCN2019086889-appb-000063
with
Figure PCTCN2019086889-appb-000064
Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton,
Figure PCTCN2019086889-appb-000065
with
Figure PCTCN2019086889-appb-000066
Representing the vertex coordinates of the model driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation. In the local rigid motion constraint, i represents the i-th vertex on the model.
Figure PCTCN2019086889-appb-000067
Represents a collection of adjacent vertices around the ith vertex on the model,
Figure PCTCN2019086889-appb-000068
with
Figure PCTCN2019086889-appb-000069
Representing the driving effects of known non-rigid motion on the surface vertices v i and v j of the model, respectively.
Figure PCTCN2019086889-appb-000070
with
Figure PCTCN2019086889-appb-000071
Represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j .
具体而言,同时使用了刚性运动约束项E s和非刚性运动约束项E n进行对象运动的优化求解,同时使用了单一深度图像进行对象刚性骨架约束项E j对求解的刚性运动进行约束。 Specifically, the rigid motion constraint term E s and the non-rigid motion constraint term E n are simultaneously used to perform an optimal solution of the object motion, and a single depth image is used to perform the object rigid skeleton constraint term E j to constrain the solved rigid motion.
(1)表面非刚性约束E n保证经过非刚性形变后的模型与从深度图获得的三维点云尽可能的对齐;
Figure PCTCN2019086889-appb-000072
Figure PCTCN2019086889-appb-000073
分别表示经过非刚性形变驱动后的模型顶点坐标及其法向,
Figure PCTCN2019086889-appb-000074
Figure PCTCN2019086889-appb-000075
分别表示经过对象骨架运动驱动后的模型顶点坐标及其法向。
(1) The surface non-rigid constraint E n ensures that the model after the non-rigid deformation is aligned with the three-dimensional point cloud obtained from the depth map as much as possible;
Figure PCTCN2019086889-appb-000072
with
Figure PCTCN2019086889-appb-000073
Representing the vertex coordinates of the model and its normal direction after being driven by non-rigid deformation,
Figure PCTCN2019086889-appb-000074
with
Figure PCTCN2019086889-appb-000075
Represents the vertex coordinates of the model and its normal direction driven by the object skeleton motion.
(2)刚性骨架运动约束项E s保证经过骨架运动驱动刚性形变后的模型与从深度图获得的三维点云尽可能的对齐。 (2) The rigid skeleton motion constraint E s ensures that the rigid deformation model driven by the skeleton motion is aligned with the three-dimensional point cloud obtained from the depth map as much as possible.
(3)在刚性骨架运动和非刚性形变一致性约束项E b中,
Figure PCTCN2019086889-appb-000076
Figure PCTCN2019086889-appb-000077
分别代表受目标刚性运动驱动后的模型顶点坐标与受三维骨架估计所得到的运动驱动后的模型顶点坐标尽量一致,该约束项用于保证解算出来的刚性骨架和识别出的骨架尽可能的一致,通过单帧骨架识别,防止动态跟踪过程中出现的误差累计而无法恢复的情况,从而可以保证最终解算出来的非刚性运动即符合对象骨架动力学模型,又充分的与从深度图中获得的三维点云对齐。
(3) In the rigid skeleton motion and the non-rigid deformation consistency constraint E b ,
Figure PCTCN2019086889-appb-000076
with
Figure PCTCN2019086889-appb-000077
The vertex coordinates of the model driven by the target rigid motion are respectively consistent with the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation. The constraint is used to ensure that the calculated rigid skeleton and the identified skeleton are as much as possible. Consistently, through single-frame skeleton recognition, the error accumulated in the dynamic tracking process can be prevented from being accumulated and cannot be recovered, so that the non-rigid motion calculated by the final solution can be guaranteed to conform to the object skeleton dynamics model, and fully and from the depth map. Get the 3D point cloud alignment.
(4)在局部刚性运动约束项E g中,i表示模型上第i个顶点,
Figure PCTCN2019086889-appb-000078
表示模型上第i个顶点周围的邻近顶点的集合,
Figure PCTCN2019086889-appb-000079
Figure PCTCN2019086889-appb-000080
分别代表已知非刚性运动对模型表面顶点v i和v j的驱动作用,
Figure PCTCN2019086889-appb-000081
Figure PCTCN2019086889-appb-000082
代表作用在v i和v j上的非刚性运动同时作用在v j上的位置变换效果,即要保证模型上邻近顶点的非刚性驱动效果要尽可能的一致。
Figure PCTCN2019086889-appb-000083
是一个鲁棒惩罚函数,
Figure PCTCN2019086889-appb-000084
Figure PCTCN2019086889-appb-000085
分别代表刚性骨架运动对模型表面顶点v i和v j的驱动作用,当模型表面两个相邻顶点受刚性骨架运动驱动效果相差较大时,该鲁棒惩罚函数值较小,当两个相邻顶点受骨架运动驱动效果相差较小时,该鲁棒惩罚函数值较大,通过该鲁棒惩罚函数,可以在使模型整体受局部刚性约束运动的同时保证较大幅度的合理的非刚性运动也能被很好的解算出来,从而使模型更精确的与三维点云对齐。
(4) In the local rigid motion constraint term E g , i represents the i-th vertex on the model,
Figure PCTCN2019086889-appb-000078
Represents a collection of adjacent vertices around the ith vertex on the model,
Figure PCTCN2019086889-appb-000079
with
Figure PCTCN2019086889-appb-000080
Representing the driving effects of known non-rigid motion on the surface vertices v i and v j of the model, respectively.
Figure PCTCN2019086889-appb-000081
with
Figure PCTCN2019086889-appb-000082
It represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j , that is, to ensure that the non-rigid driving effects of adjacent vertices on the model are as uniform as possible.
Figure PCTCN2019086889-appb-000083
Is a robust penalty function,
Figure PCTCN2019086889-appb-000084
with
Figure PCTCN2019086889-appb-000085
Representing the driving effect of the rigid skeleton motion on the surface vertices v i and v j of the model respectively. When the two adjacent vertices of the model surface are driven by the rigid skeleton motion, the value of the robust penalty function is small, when the two phases are When the neighboring vertex is less affected by the skeleton motion driving effect, the robust penalty function value is larger. Through the robust penalty function, the model can be subjected to local rigid constrained motion while ensuring a large amplitude of reasonable non-rigid motion. Can be well solved, so that the model is more accurately aligned with the 3D point cloud.
在步骤S105中,对能量函数进行GPU优化求解,以获得每个表面顶点的非刚性形变,并根据求解结果将前一帧的重建三维模型进行形变,使得形变模型与当前帧三维点云进行对齐。In step S105, GPU optimization is performed on the energy function to obtain a non-rigid deformation of each surface vertex, and the reconstructed three-dimensional model of the previous frame is deformed according to the solution result, so that the deformation model is aligned with the current frame three-dimensional point cloud. .
可以理解的是,如图2所示,对能量函数进行GPU优化求解,求解重建模型上每一个顶点的非刚性运动位置变换参数,并优化对象三维刚性运动信息,根据求解结果对前一帧 重建模型进行形变,使之与当前帧三维点云进行对齐。It can be understood that, as shown in FIG. 2, the GPU optimization of the energy function is performed to solve the non-rigid motion position transformation parameters of each vertex on the reconstructed model, and the three-dimensional rigid motion information of the object is optimized, and the previous frame is reconstructed according to the solution result. The model is deformed to align with the current frame 3D point cloud.
具体而言,对能量函数进行求解,根据求解结果将重建模型与三维点云进行对齐。同求解重建模型上每一个顶点的非刚性运动位置变换参数和对象骨架运动参数。最终求解获得的信息为每一个三维模型顶点的变换矩阵和对象骨架运动参数,即每个骨骼的单独的变换矩阵。为了实现快速线性求解的要求,本发明实施例的方法对利用指数映射方法对变形方程做如下近似:Specifically, the energy function is solved, and the reconstructed model is aligned with the three-dimensional point cloud according to the solution result. The non-rigid motion position transformation parameters and the object skeleton motion parameters of each vertex on the reconstruction model are solved. The information obtained by the final solution is the transformation matrix of each 3D model vertex and the object skeleton motion parameters, that is, the individual transformation matrix of each bone. In order to achieve the requirement of fast linear solution, the method of the embodiment of the present invention approximates the deformation equation by using an exponential mapping method:
Figure PCTCN2019086889-appb-000086
Figure PCTCN2019086889-appb-000086
其中,
Figure PCTCN2019086889-appb-000087
为截至上一帧的模型顶点v i的累积变换矩阵,为已知量,
Figure PCTCN2019086889-appb-000088
为每个表面顶点的非刚性形变;I为四维单位阵;
among them,
Figure PCTCN2019086889-appb-000087
For the cumulative transformation matrix of the model vertex v i up to the previous frame, for the known amount,
Figure PCTCN2019086889-appb-000088
a non-rigid deformation for each surface apex; I is a four-dimensional unit matrix;
其中,
Figure PCTCN2019086889-appb-000089
Figure PCTCN2019086889-appb-000090
即上一帧变换后的模型顶点,则经过变换有:
among them,
Figure PCTCN2019086889-appb-000089
make
Figure PCTCN2019086889-appb-000090
That is, the model vertices after the previous frame transformation are transformed:
Figure PCTCN2019086889-appb-000091
Figure PCTCN2019086889-appb-000091
对于每个顶点,要求解的未知参数即为六维变换参数x=(v 1,v 2,v 3,w x,w y,w z) T。骨骼运动的线性化方式与非刚性运动相同。 For each vertex, the unknown parameter that requires the solution is the six-dimensional transformation parameter x = (v 1 , v 2 , v 3 , w x , w y , w z ) T . The linearization of bone movement is the same as that of non-rigid motion.
在步骤S106中,融合当前帧三维点云与形变模型,以获得当前帧的更新后的模型,以进入下一帧的迭代。In step S106, the current frame three-dimensional point cloud and the deformation model are merged to obtain an updated model of the current frame to enter an iteration of the next frame.
可以理解的是,如图2所示,对对齐后的模型和点云进行泊松融合,获得新一帧较完整的三维模型。It can be understood that, as shown in FIG. 2, Poisson fusion is performed on the aligned model and the point cloud to obtain a relatively complete three-dimensional model of the new frame.
具体而言,融合点云和三维模型,获得当前帧的更新模型。使用与深度点云对齐后的三维模型进行更新和补全,将新获得的深度信息融合到三维模型中,更新三维模型表面顶点位置或为三维模型增加新的顶点,使其更符合当前深度图像的表达。Specifically, the point cloud and the three-dimensional model are merged to obtain an updated model of the current frame. Update and complete the 3D model aligned with the depth point cloud, fuse the newly obtained depth information into the 3D model, update the surface vertex position of the 3D model or add new vertices to the 3D model to make it more consistent with the current depth image expression.
综上,本发明实施例核心功能在于实时接受深度图像码流,实时计算每帧三维模型。同时利用对象的大尺度刚性骨架运动和小尺度的表面非刚性形变信息计算动态对象的时变三维模型。本发明实施例的方法求解准确,可以实现实时对动态对象进行高精度重建,由于该方法为实时重建方法,并且仅需提供单个深度相机输入,系统具有设备简单,方便部署和可扩展等有点,所需的输入信息非常容易采集,并且可以实时的获得动态三维模型。 该方法求解准确鲁棒,简单易行,运行速度为实时,拥有广阔的应用前景,可以在PC(personal computer,个人计算机)机或工作站等硬件系统上快速实现。In summary, the core function of the embodiment of the present invention is to receive a depth image code stream in real time and calculate a three-dimensional model of each frame in real time. At the same time, the time-varying three-dimensional model of the dynamic object is calculated by using the large-scale rigid skeleton motion of the object and the small-scale surface non-rigid deformation information. The method of the embodiment of the invention is accurate and can realize high-precision reconstruction of the dynamic object in real time. Since the method is a real-time reconstruction method and only needs to provide a single depth camera input, the system has the advantages of simple equipment, convenient deployment and scalability, and the like. The required input information is very easy to acquire and a dynamic 3D model can be obtained in real time. The method is accurate, robust, simple and easy to operate, and runs at a real-time speed. It has broad application prospects and can be quickly implemented on hardware systems such as PC (personal computer) or workstations.
根据本发明实施例提出的联合刚性运动和非刚性形变的三维重建方法,通过实时非刚性对齐的方法,逐帧地融合动态对象表面三维信息,为了实现鲁棒地跟踪,实现在无首帧关键帧三维模板条件下的鲁棒性实时动态三维重建,从而可以有效提高重建的实时性、鲁棒性和准确性,扩展性强,简单易实现。According to the three-dimensional reconstruction method of joint rigid motion and non-rigid deformation according to the embodiment of the present invention, the three-dimensional information of the dynamic object surface is merged frame by frame by real-time non-rigid alignment method, in order to achieve robust tracking, realizing the key without the first frame Robust real-time dynamic 3D reconstruction under frame 3D template conditions can effectively improve the real-time, robustness and accuracy of reconstruction, and it is scalable and easy to implement.
其次参照附图描述根据本发明实施例提出的联合刚性运动和非刚性形变的三维重建装置。Next, a three-dimensional reconstruction apparatus combining joint rigid motion and non-rigid deformation according to an embodiment of the present invention will be described with reference to the accompanying drawings.
图3是本发明一个实施例的联合刚性运动和非刚性形变的三维重建装置的结构示意图。3 is a schematic view showing the structure of a three-dimensional reconstruction apparatus combining rigid motion and non-rigid deformation according to an embodiment of the present invention.
如图3所示,该联合刚性运动和非刚性形变的三维重建装置10包括:拍摄模块100、提取模块200、匹配模块300、解算模块400、求解模块500和模型更新模块600。As shown in FIG. 3, the joint rigid motion and non-rigid deformation 3D reconstruction apparatus 10 includes a photographing module 100, an extraction module 200, a matching module 300, a solution module 400, a solution module 500, and a model update module 600.
其中,拍摄模块100用于对目标对象进行基于深度相机的拍摄,以得到单张深度图像。提取模块200用于通过三维骨架提取算法对深度点云进行三维骨架提取。匹配模块300将单张深度图像变换为三维点云,并获取三维点云和重建模型顶点之间的匹配点对。解算模块400用于根据匹配点对和三维骨架信息建立能量函数,并求解重建模型上每一个顶点的非刚性运动位置变换参数并优化对象骨架参数。求解模块500用于对能量函数进行GPU优化求解,以获得每个表面顶点的非刚性形变,并根据求解结果将前一帧的重建三维模型进行形变,使得形变模型与当前帧三维点云进行对齐。模型更新模块600用于融合当前帧三维点云与形变模型,以获得当前帧的更新后的模型,以进入下一帧的迭代。本发明实施例的装置10可以有效提高重建的实时性、鲁棒性和准确性,扩展性强,简单易实现。The shooting module 100 is configured to perform depth camera-based shooting on the target object to obtain a single depth image. The extraction module 200 is configured to perform three-dimensional skeleton extraction on the depth point cloud by using a three-dimensional skeleton extraction algorithm. The matching module 300 transforms the single depth image into a three-dimensional point cloud, and acquires a matching point pair between the three-dimensional point cloud and the reconstructed model vertex. The solving module 400 is configured to establish an energy function according to the matching point pair and the three-dimensional skeleton information, and solve the non-rigid motion position transformation parameter of each vertex on the reconstruction model and optimize the object skeleton parameter. The solving module 500 is configured to perform GPU optimization on the energy function to obtain a non-rigid deformation of each surface vertex, and deform the reconstructed three-dimensional model of the previous frame according to the solution result, so that the deformation model is aligned with the current frame three-dimensional point cloud. . The model update module 600 is configured to fuse the current frame three-dimensional point cloud and the deformation model to obtain an updated model of the current frame to enter an iteration of the next frame. The device 10 of the embodiment of the invention can effectively improve the real-time, robustness and accuracy of the reconstruction, has strong scalability, and is simple and easy to implement.
进一步地,在本发明的一个实施例中,匹配模块300进一步用于通过深度相机的内参矩阵将单张深度图像投影到三维空间中,以生成三维点云,其中,深度图投影公式为:Further, in an embodiment of the present invention, the matching module 300 is further configured to project a single depth image into the three-dimensional space by using an internal parameter matrix of the depth camera to generate a three-dimensional point cloud, wherein the depth map projection formula is:
Figure PCTCN2019086889-appb-000092
Figure PCTCN2019086889-appb-000092
其中,u,v为像素坐标,d(u,v)为深度图像上像素(u,v)位置处的深度值,
Figure PCTCN2019086889-appb-000093
为深度相机的内参矩阵。
Where u, v are pixel coordinates, and d(u, v) is a depth value at a pixel (u, v) position on the depth image,
Figure PCTCN2019086889-appb-000093
The internal reference matrix for the depth camera.
进一步地,在本发明的一个实施例中,能量函数为:Further, in one embodiment of the invention, the energy function is:
E t=λ nE nsE sjE jgE gbE bE tn E ns E sj E jg E gb E b ,
其中,E t为总能量项,E n为非刚性表面形变约束项,E s为刚性骨架运动约束项,E j为刚性骨架识别约束项,E g为局部刚性运动约束项,λ n、λ s、λ j和λ g分别为对应各个约束项的权重系数。 Where E t is the total energy term, E n is the non-rigid surface deformation constraint term, E s is the rigid skeleton motion constraint term, E j is the rigid skeleton recognition constraint term, E g is the local rigid motion constraint term, λ n , λ s , λ j and λ g are weight coefficients corresponding to respective constraint terms, respectively.
进一步地,在本发明的一个实施例中,其中,Further, in an embodiment of the present invention, wherein
Figure PCTCN2019086889-appb-000094
Figure PCTCN2019086889-appb-000094
Figure PCTCN2019086889-appb-000095
Figure PCTCN2019086889-appb-000095
Figure PCTCN2019086889-appb-000096
Figure PCTCN2019086889-appb-000096
Figure PCTCN2019086889-appb-000097
Figure PCTCN2019086889-appb-000097
其中,u i表示同一匹配点对中三维点云的位置坐标,c i表示匹配点对集合中的第i个元素,非刚性表面形变约束项中
Figure PCTCN2019086889-appb-000098
Figure PCTCN2019086889-appb-000099
分别表示经过非刚性形变驱动后的模型顶点坐标及其法向,刚性骨架运动约束项中
Figure PCTCN2019086889-appb-000100
Figure PCTCN2019086889-appb-000101
分别表示经过对象骨架运动驱动后的模型顶点坐标及其法向,
Figure PCTCN2019086889-appb-000102
Figure PCTCN2019086889-appb-000103
分别代表受目标刚性运动驱动后的模型顶点坐标与受三维骨架估计所得到的运动驱动后的模型顶点坐标,局部刚性运动约束项中,i表示模型上第i个顶点,
Figure PCTCN2019086889-appb-000104
表示模型上第i个顶点周围的邻近顶点的集合,
Figure PCTCN2019086889-appb-000105
Figure PCTCN2019086889-appb-000106
分别代表已知非刚性运动对模型表面顶点v i和v j的驱动作用,
Figure PCTCN2019086889-appb-000107
Figure PCTCN2019086889-appb-000108
代表作用在v i和v j上的非刚性运动同时作用在v j上的位置变换效果。
Where u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair, c i represents the i-th element in the matching point pair set, and the non-rigid surface deformation constraint item
Figure PCTCN2019086889-appb-000098
with
Figure PCTCN2019086889-appb-000099
Representing the vertex coordinates of the model and its normal direction driven by non-rigid deformation, respectively, in the rigid skeleton motion constraint
Figure PCTCN2019086889-appb-000100
with
Figure PCTCN2019086889-appb-000101
Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton,
Figure PCTCN2019086889-appb-000102
with
Figure PCTCN2019086889-appb-000103
Representing the vertex coordinates of the model driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation. In the local rigid motion constraint, i represents the i-th vertex on the model.
Figure PCTCN2019086889-appb-000104
Represents a collection of adjacent vertices around the ith vertex on the model,
Figure PCTCN2019086889-appb-000105
with
Figure PCTCN2019086889-appb-000106
Representing the driving effects of known non-rigid motion on the surface vertices v i and v j of the model, respectively.
Figure PCTCN2019086889-appb-000107
with
Figure PCTCN2019086889-appb-000108
Represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j .
进一步地,在本发明的一个实施例中,根据表面非刚性形变和对象刚性骨架运动驱动模型顶点,其中,计算公式为:Further, in one embodiment of the invention, the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
Figure PCTCN2019086889-appb-000109
Figure PCTCN2019086889-appb-000109
Figure PCTCN2019086889-appb-000110
Figure PCTCN2019086889-appb-000110
其中,
Figure PCTCN2019086889-appb-000111
为作用于顶点v i的变形矩阵,包括旋转和平移两部分;
Figure PCTCN2019086889-appb-000112
为该变形矩阵的旋转部分;
Figure PCTCN2019086889-appb-000113
为对顶点v i有驱动作用的骨骼的集合;α i,j为第j个骨骼对第i个模型顶点的驱动作用的权重,表示该骨骼对该顶点驱动作用的强弱;T bj为第j个骨骼自身的运动变形矩阵,rot(T bj)为该变形矩阵的旋转部分。
among them,
Figure PCTCN2019086889-appb-000111
a deformation matrix acting on the vertex v i , including two parts of rotation and translation;
Figure PCTCN2019086889-appb-000112
Is the rotating portion of the deformation matrix;
Figure PCTCN2019086889-appb-000113
a set of bones that have a driving effect on the vertex v i ; α i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j skeletons themselves, rot(T bj ) is the rotation part of the deformation matrix.
需要说明的是,前述对联合刚性运动和非刚性形变的三维重建方法实施例的解释说明也适用于该实施例的联合刚性运动和非刚性形变的三维重建装置,此处不再赘述。It should be noted that the foregoing explanation of the embodiment of the three-dimensional reconstruction method for joint rigid motion and non-rigid deformation is also applicable to the joint rigid motion and non-rigid deformation three-dimensional reconstruction apparatus of the embodiment, and details are not described herein again.
根据本发明实施例提出的联合刚性运动和非刚性形变的三维重建装置,通过实时非刚性对齐的方法,逐帧地融合动态对象表面三维信息,为了实现鲁棒地跟踪,实现在无首帧关键帧三维模板条件下的鲁棒性实时动态三维重建,从而可以有效提高重建的实时性、鲁棒性和准确性,扩展性强,简单易实现。According to the three-dimensional reconstruction device of the joint rigid motion and the non-rigid deformation according to the embodiment of the present invention, the three-dimensional information of the dynamic object surface is merged frame by frame by real-time non-rigid alignment method, in order to achieve robust tracking, the key in the first frame is realized. Robust real-time dynamic 3D reconstruction under frame 3D template conditions can effectively improve the real-time, robustness and accuracy of reconstruction, and it is scalable and easy to implement.
在本发明的描述中,需要理解的是,术语“中心”、“纵向”、“横向”、“长度”、“宽度”、 “厚度”、“上”、“下”、“前”、“后”、“左”、“右”、“竖直”、“水平”、“顶”、“底”“内”、“外”、“顺时针”、“逆时针”、“轴向”、“径向”、“周向”等指示的方位或位置关系为基于附图所示的方位或位置关系,仅是为了便于描述本发明和简化描述,而不是指示或暗示所指的装置或元件必须具有特定的方位、以特定的方位构造和操作,因此不能理解为对本发明的限制。In the description of the present invention, it is to be understood that the terms "center", "longitudinal", "transverse", "length", "width", "thickness", "upper", "lower", "front", " After, "Left", "Right", "Vertical", "Horizontal", "Top", "Bottom", "Inside", "Outside", "Clockwise", "Counterclockwise", "Axial", The orientation or positional relationship of the "radial", "circumferential" and the like is based on the orientation or positional relationship shown in the drawings, and is merely for convenience of description of the present invention and simplified description, and does not indicate or imply the indicated device or component. It must be constructed and operated in a particular orientation, and is not to be construed as limiting the invention.
此外,术语“第一”、“第二”仅用于描述目的,而不能理解为指示或暗示相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括至少一个该特征。在本发明的描述中,“多个”的含义是至少两个,例如两个,三个等,除非另有明确具体的限定。Moreover, the terms "first" and "second" are used for descriptive purposes only and are not to be construed as indicating or implying a relative importance or implicitly indicating the number of technical features indicated. Thus, features defining "first" or "second" may include at least one of the features, either explicitly or implicitly. In the description of the present invention, the meaning of "a plurality" is at least two, such as two, three, etc., unless specifically defined otherwise.
在本发明中,除非另有明确的规定和限定,术语“安装”、“相连”、“连接”、“固定”等术语应做广义理解,例如,可以是固定连接,也可以是可拆卸连接,或成一体;可以是机械连接,也可以是电连接;可以是直接相连,也可以通过中间媒介间接相连,可以是两个元件内部的连通或两个元件的相互作用关系,除非另有明确的限定。对于本领域的普通技术人员而言,可以根据具体情况理解上述术语在本发明中的具体含义。In the present invention, the terms "installation", "connected", "connected", "fixed" and the like shall be understood broadly, and may be either a fixed connection or a detachable connection, unless explicitly stated and defined otherwise. , or integrated; can be mechanical or electrical connection; can be directly connected, or indirectly connected through an intermediate medium, can be the internal communication of two elements or the interaction of two elements, unless otherwise specified Limited. For those skilled in the art, the specific meanings of the above terms in the present invention can be understood on a case-by-case basis.
在本发明中,除非另有明确的规定和限定,第一特征在第二特征“上”或“下”可以是第一和第二特征直接接触,或第一和第二特征通过中间媒介间接接触。而且,第一特征在第二特征“之上”、“上方”和“上面”可是第一特征在第二特征正上方或斜上方,或仅仅表示第一特征水平高度高于第二特征。第一特征在第二特征“之下”、“下方”和“下面”可以是第一特征在第二特征正下方或斜下方,或仅仅表示第一特征水平高度小于第二特征。In the present invention, the first feature "on" or "under" the second feature may be a direct contact of the first and second features, or the first and second features may be indirectly through an intermediate medium, unless otherwise explicitly stated and defined. contact. Moreover, the first feature "above", "above" and "above" the second feature may be that the first feature is directly above or above the second feature, or merely that the first feature level is higher than the second feature. The first feature "below", "below" and "below" the second feature may be that the first feature is directly below or obliquely below the second feature, or merely that the first feature level is less than the second feature.
在本说明书的描述中,参考术语“一个实施例”、“一些实施例”、“示例”、“具体示例”、或“一些示例”等的描述意指结合该实施例或示例描述的具体特征、结构、材料或者特点包含于本发明的至少一个实施例或示例中。在本说明书中,对上述术语的示意性表述不必须针对的是相同的实施例或示例。而且,描述的具体特征、结构、材料或者特点可以在任一个或多个实施例或示例中以合适的方式结合。此外,在不相互矛盾的情况下,本领域的技术人员可以将本说明书中描述的不同实施例或示例以及不同实施例或示例的特征进行结合和组合。In the description of the present specification, the description with reference to the terms "one embodiment", "some embodiments", "example", "specific example", or "some examples" and the like means a specific feature described in connection with the embodiment or example. A structure, material or feature is included in at least one embodiment or example of the invention. In the present specification, the schematic representation of the above terms is not necessarily directed to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in a suitable manner in any one or more embodiments or examples. In addition, various embodiments or examples described in the specification, as well as features of various embodiments or examples, may be combined and combined.
尽管上面已经示出和描述了本发明的实施例,可以理解的是,上述实施例是示例性的,不能理解为对本发明的限制,本领域的普通技术人员在本发明的范围内可以对上述实施例进行变化、修改、替换和变型。Although the embodiments of the present invention have been shown and described, it is understood that the above-described embodiments are illustrative and are not to be construed as limiting the scope of the invention. The embodiments are subject to variations, modifications, substitutions and variations.

Claims (10)

  1. 一种联合刚性运动和非刚性形变的三维重建方法,其特征在于,包括以下步骤:A three-dimensional reconstruction method combining joint rigid motion and non-rigid deformation, characterized in that it comprises the following steps:
    对目标对象进行基于深度相机的拍摄,以得到单张深度图像;Performing depth camera based shooting on the target object to obtain a single depth image;
    通过三维骨架提取算法对深度点云进行三维骨架提取;Three-dimensional skeleton extraction of deep point clouds by three-dimensional skeleton extraction algorithm;
    将所述单张深度图像变换为三维点云,并获取所述三维点云和重建模型顶点之间的匹配点对;Converting the single depth image into a three-dimensional point cloud, and acquiring a matching point pair between the three-dimensional point cloud and the reconstructed model vertex;
    根据所述匹配点对和三维骨架信息建立能量函数,并求解所述重建模型上每一个顶点的非刚性运动位置变换参数并优化对象骨架参数;Generating an energy function according to the pair of matching points and the three-dimensional skeleton information, and solving non-rigid motion position transformation parameters of each vertex on the reconstruction model and optimizing the object skeleton parameters;
    对所述能量函数进行GPU优化求解,以获得每个表面顶点的非刚性形变,并根据求解结果将前一帧的重建三维模型进行形变,使得形变模型与当前帧三维点云进行对齐;以及Performing GPU optimization on the energy function to obtain a non-rigid deformation of each surface vertex, and deforming the reconstructed three-dimensional model of the previous frame according to the solution result, so that the deformation model is aligned with the current frame three-dimensional point cloud;
    融合当前帧三维点云与所述形变模型,以获得当前帧的更新后的模型,以进入下一帧的迭代。The current frame 3D point cloud is merged with the deformation model to obtain an updated model of the current frame to enter an iteration of the next frame.
  2. 根据权利要求1所述的联合刚性运动和非刚性形变的三维重建方法,其特征在于,所述将所述单张深度图像变换为三维点云,进一步包括:The three-dimensional reconstruction method of the joint rigid motion and the non-rigid deformation according to claim 1, wherein the converting the single depth image into a three-dimensional point cloud further comprises:
    通过深度相机的内参矩阵将所述单张深度图像投影到三维空间中,以生成所述三维点云,其中,深度图投影公式为:The single depth image is projected into the three-dimensional space by an internal parameter matrix of the depth camera to generate the three-dimensional point cloud, wherein the depth map projection formula is:
    Figure PCTCN2019086889-appb-100001
    Figure PCTCN2019086889-appb-100001
    其中,u,v为像素坐标,d(u,v)为深度图像上像素(u,v)位置处的深度值,
    Figure PCTCN2019086889-appb-100002
    为所述深度相机的内参矩阵。
    Where u, v are pixel coordinates, and d(u, v) is a depth value at a pixel (u, v) position on the depth image,
    Figure PCTCN2019086889-appb-100002
    Is the internal reference matrix of the depth camera.
  3. 根据权利要求1所述的联合刚性运动和非刚性形变的三维重建方法,其特征在于,所述能量函数为:The method of claim 3, wherein the energy function is:
    E t=λ nE nsE sjE jgE gbE bE tn E ns E sj E jg E gb E b ,
    其中,E t为总能量项,E n为非刚性表面形变约束项,E s为刚性骨架运动约束项,E j为刚性骨架识别约束项,E g为局部刚性运动约束项,λ n、λ s、λ j和λ g分别为对应各个约束项的权重系数。 Where E t is the total energy term, E n is the non-rigid surface deformation constraint term, E s is the rigid skeleton motion constraint term, E j is the rigid skeleton recognition constraint term, E g is the local rigid motion constraint term, λ n , λ s , λ j and λ g are weight coefficients corresponding to respective constraint terms, respectively.
  4. 根据权利要求3所述的联合刚性运动和非刚性形变的三维重建方法,其特征在于,其中,The method of three-dimensional reconstruction of joint rigid motion and non-rigid deformation according to claim 3, wherein
    Figure PCTCN2019086889-appb-100003
    Figure PCTCN2019086889-appb-100003
    Figure PCTCN2019086889-appb-100004
    Figure PCTCN2019086889-appb-100004
    Figure PCTCN2019086889-appb-100005
    Figure PCTCN2019086889-appb-100005
    Figure PCTCN2019086889-appb-100006
    Figure PCTCN2019086889-appb-100006
    其中,u i表示同一匹配点对中三维点云的位置坐标,c i表示匹配点对集合中的第i个元素,所述非刚性表面形变约束项中
    Figure PCTCN2019086889-appb-100007
    Figure PCTCN2019086889-appb-100008
    分别表示经过非刚性形变驱动后的模型顶点坐标及其法向,所述刚性骨架运动约束项中
    Figure PCTCN2019086889-appb-100009
    Figure PCTCN2019086889-appb-100010
    分别表示经过对象骨架运动驱动后的模型顶点坐标及其法向,
    Figure PCTCN2019086889-appb-100011
    Figure PCTCN2019086889-appb-100012
    分别代表受目标刚性运动驱动后的模型顶点坐标与受三维骨架估计所得到的运动驱动后的模型顶点坐标,所述局部刚性运动约束项中,i表示模型上第i个顶点,
    Figure PCTCN2019086889-appb-100013
    表示模型上第i个顶点周围的邻近顶点的集合,
    Figure PCTCN2019086889-appb-100014
    Figure PCTCN2019086889-appb-100015
    分别代表已知非刚性运动对模型表面顶点v i和v j的驱动作用,
    Figure PCTCN2019086889-appb-100016
    Figure PCTCN2019086889-appb-100017
    代表作用在v i和v j上的非刚性运动同时作用在v j上的位置变换效果。
    Where u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair, c i represents the i-th element in the set of matching point pairs, and the non-rigid surface deformation constraint item
    Figure PCTCN2019086889-appb-100007
    with
    Figure PCTCN2019086889-appb-100008
    Representing the vertex coordinates of the model and its normal direction driven by the non-rigid deformation, respectively, in the rigid skeleton motion constraint
    Figure PCTCN2019086889-appb-100009
    with
    Figure PCTCN2019086889-appb-100010
    Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton,
    Figure PCTCN2019086889-appb-100011
    with
    Figure PCTCN2019086889-appb-100012
    Representing the model vertex coordinates driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation, where i represents the i-th vertex on the model.
    Figure PCTCN2019086889-appb-100013
    Represents a collection of adjacent vertices around the ith vertex on the model,
    Figure PCTCN2019086889-appb-100014
    with
    Figure PCTCN2019086889-appb-100015
    Representing the driving effects of known non-rigid motion on the surface vertices v i and v j of the model, respectively.
    Figure PCTCN2019086889-appb-100016
    with
    Figure PCTCN2019086889-appb-100017
    Represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j .
  5. 根据权利要求1-4任一项所述的联合刚性运动和非刚性形变的三维重建方法,其特征在于,根据表面非刚性形变和对象刚性骨架运动驱动模型顶点,其中,计算公式为:A three-dimensional reconstruction method for joint rigid motion and non-rigid deformation according to any one of claims 1 to 4, characterized in that the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
    Figure PCTCN2019086889-appb-100018
    Figure PCTCN2019086889-appb-100018
    Figure PCTCN2019086889-appb-100019
    Figure PCTCN2019086889-appb-100019
    其中,
    Figure PCTCN2019086889-appb-100020
    为作用于顶点v i的变形矩阵,包括旋转和平移两部分;
    Figure PCTCN2019086889-appb-100021
    为该变形矩阵的旋转部分;
    Figure PCTCN2019086889-appb-100022
    为对顶点v i有驱动作用的骨骼的集合;α i,j为第j个骨骼对第i个模型顶点的驱动作用的权重,表示该骨骼对该顶点驱动作用的强弱;T bj为第j个骨骼自身的运动变形矩阵,rot(T bj)为该变形矩阵的旋转部分。
    among them,
    Figure PCTCN2019086889-appb-100020
    a deformation matrix acting on the vertex v i , including two parts of rotation and translation;
    Figure PCTCN2019086889-appb-100021
    Is the rotating portion of the deformation matrix;
    Figure PCTCN2019086889-appb-100022
    a set of bones that have a driving effect on the vertex v i ; α i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j skeletons themselves, rot(T bj ) is the rotation part of the deformation matrix.
  6. 一种联合刚性运动和非刚性形变的三维重建装置,其特征在于,包括:A three-dimensional reconstruction device combining rigid motion and non-rigid deformation, comprising:
    拍摄模块,用于对目标对象进行基于深度相机的拍摄,以得到单张深度图像;a shooting module for performing depth camera based shooting on the target object to obtain a single depth image;
    提取模块,用于通过三维骨架提取算法对深度点云进行三维骨架提取;An extraction module for performing three-dimensional skeleton extraction on the depth point cloud by using a three-dimensional skeleton extraction algorithm;
    匹配模块,将所述单张深度图像变换为三维点云,并获取所述三维点云和重建模型顶点之间的匹配点对;a matching module, transforming the single depth image into a three-dimensional point cloud, and acquiring a matching point pair between the three-dimensional point cloud and the reconstructed model vertex;
    解算模块,用于根据所述匹配点对和三维骨架信息建立能量函数,并求解所述重建模型上每一个顶点的非刚性运动位置变换参数并优化对象骨架参数;a solution module, configured to establish an energy function according to the pair of matching points and the three-dimensional skeleton information, and solve a non-rigid motion position transformation parameter of each vertex on the reconstruction model and optimize an object skeleton parameter;
    求解模块,用于对所述能量函数进行GPU优化求解,以获得每个表面顶点的非刚性形变,并根据求解结果将前一帧的重建三维模型进行形变,使得形变模型与当前帧三维点云 进行对齐;以及a solution module for performing GPU optimization on the energy function to obtain a non-rigid deformation of each surface vertex, and deforming the reconstructed three-dimensional model of the previous frame according to the solution result, so that the deformation model and the current frame three-dimensional point cloud Align; and
    模型更新模块,用于融合当前帧三维点云与所述形变模型,以获得当前帧的更新后的模型,以进入下一帧的迭代。And a model updating module, configured to fuse the current frame three-dimensional point cloud and the deformation model to obtain an updated model of the current frame to enter an iteration of the next frame.
  7. 根据权利要求6所述的联合刚性运动和非刚性形变的三维重建装置,其特征在于,所述匹配模块进一步用于通过深度相机的内参矩阵将所述单张深度图像投影到三维空间中,以生成所述三维点云,其中,深度图投影公式为:The combined rigid motion and non-rigid deformation three-dimensional reconstruction apparatus according to claim 6, wherein the matching module is further configured to project the single depth image into the three-dimensional space by using an internal parameter matrix of the depth camera, Generating the three-dimensional point cloud, wherein the depth map projection formula is:
    Figure PCTCN2019086889-appb-100023
    Figure PCTCN2019086889-appb-100023
    其中,u,v为像素坐标,d(u,v)为深度图像上像素(u,v)位置处的深度值,
    Figure PCTCN2019086889-appb-100024
    为所述深度相机的内参矩阵。
    Where u, v are pixel coordinates, and d(u, v) is a depth value at a pixel (u, v) position on the depth image,
    Figure PCTCN2019086889-appb-100024
    Is the internal reference matrix of the depth camera.
  8. 根据权利要求6所述的联合刚性运动和非刚性形变的三维重建装置,其特征在于,所述能量函数为:The combined rigid motion and non-rigid deformation three-dimensional reconstruction apparatus according to claim 6, wherein the energy function is:
    E t=λ nE nsE sjE jgE gbE bE tn E ns E sj E jg E gb E b ,
    其中,E t为总能量项,E n为非刚性表面形变约束项,E s为刚性骨架运动约束项,E j为刚性骨架识别约束项,E g为局部刚性运动约束项,λ n、λ s、λ j和λ g分别为对应各个约束项的权重系数。 Where E t is the total energy term, E n is the non-rigid surface deformation constraint term, E s is the rigid skeleton motion constraint term, E j is the rigid skeleton recognition constraint term, E g is the local rigid motion constraint term, λ n , λ s , λ j and λ g are weight coefficients corresponding to respective constraint terms, respectively.
  9. 根据权利要求8所述的联合刚性运动和非刚性形变的三维重建装置,其特征在于,其中,The combined rigid motion and non-rigid deformation three-dimensional reconstruction apparatus according to claim 8, wherein
    Figure PCTCN2019086889-appb-100025
    Figure PCTCN2019086889-appb-100025
    Figure PCTCN2019086889-appb-100026
    Figure PCTCN2019086889-appb-100026
    Figure PCTCN2019086889-appb-100027
    Figure PCTCN2019086889-appb-100027
    Figure PCTCN2019086889-appb-100028
    Figure PCTCN2019086889-appb-100028
    其中,u i表示同一匹配点对中三维点云的位置坐标,c i表示匹配点对集合中的第i个元素,所述非刚性表面形变约束项中
    Figure PCTCN2019086889-appb-100029
    Figure PCTCN2019086889-appb-100030
    分别表示经过非刚性形变驱动后的模型顶点坐标及其法向,所述刚性骨架运动约束项中
    Figure PCTCN2019086889-appb-100031
    Figure PCTCN2019086889-appb-100032
    分别表示经过对象骨架运动驱动后的模型顶点坐标及其法向,
    Figure PCTCN2019086889-appb-100033
    Figure PCTCN2019086889-appb-100034
    分别代表受目标刚性运动驱动后的模型顶点坐标与受三维骨架估计所得到的运动驱动后的模型顶点坐标,所述局部刚性运动约束项中,i表示模型上第i个顶点,
    Figure PCTCN2019086889-appb-100035
    表示模型上第i个顶点周围的邻近顶点的集合,
    Figure PCTCN2019086889-appb-100036
    Figure PCTCN2019086889-appb-100037
    分别代表已知非刚性运动对模型表面顶点v i和v j的驱动作用,
    Figure PCTCN2019086889-appb-100038
    Figure PCTCN2019086889-appb-100039
    代表作用在v i和v j上的非刚性运动同时作用 在v j上的位置变换效果。
    Where u i represents the position coordinate of the three-dimensional point cloud in the same matching point pair, c i represents the i-th element in the set of matching point pairs, and the non-rigid surface deformation constraint item
    Figure PCTCN2019086889-appb-100029
    with
    Figure PCTCN2019086889-appb-100030
    Representing the vertex coordinates of the model and its normal direction driven by the non-rigid deformation, respectively, in the rigid skeleton motion constraint
    Figure PCTCN2019086889-appb-100031
    with
    Figure PCTCN2019086889-appb-100032
    Representing the coordinates of the model vertex and its normal direction driven by the motion of the object skeleton,
    Figure PCTCN2019086889-appb-100033
    with
    Figure PCTCN2019086889-appb-100034
    Representing the model vertex coordinates driven by the target rigid motion and the motion-driven model vertex coordinates obtained by the three-dimensional skeleton estimation, where i represents the i-th vertex on the model.
    Figure PCTCN2019086889-appb-100035
    Represents a collection of adjacent vertices around the ith vertex on the model,
    Figure PCTCN2019086889-appb-100036
    with
    Figure PCTCN2019086889-appb-100037
    Representing the driving effects of known non-rigid motion on the surface vertices v i and v j of the model, respectively.
    Figure PCTCN2019086889-appb-100038
    with
    Figure PCTCN2019086889-appb-100039
    Represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j .
  10. 根据权利要求6-9任一项所述的联合刚性运动和非刚性形变的三维重建装置,其特征在于,根据表面非刚性形变和对象刚性骨架运动驱动模型顶点,其中,计算公式为:The combined rigid motion and non-rigid deformation three-dimensional reconstruction apparatus according to any one of claims 6-9, wherein the model vertices are driven according to the surface non-rigid deformation and the object rigid skeleton motion, wherein the calculation formula is:
    Figure PCTCN2019086889-appb-100040
    Figure PCTCN2019086889-appb-100040
    Figure PCTCN2019086889-appb-100041
    Figure PCTCN2019086889-appb-100041
    其中,
    Figure PCTCN2019086889-appb-100042
    为作用于顶点v i的变形矩阵,包括旋转和平移两部分;
    Figure PCTCN2019086889-appb-100043
    为该变形矩阵的旋转部分;
    Figure PCTCN2019086889-appb-100044
    为对顶点v i有驱动作用的骨骼的集合;α i,j为第j个骨骼对第i个模型顶点的驱动作用的权重,表示该骨骼对该顶点驱动作用的强弱;T bj为第j个骨骼自身的运动变形矩阵,rot(T bj)为该变形矩阵的旋转部分。
    among them,
    Figure PCTCN2019086889-appb-100042
    a deformation matrix acting on the vertex v i , including two parts of rotation and translation;
    Figure PCTCN2019086889-appb-100043
    Is the rotating portion of the deformation matrix;
    Figure PCTCN2019086889-appb-100044
    a set of bones that have a driving effect on the vertex v i ; α i, j is the weight of the driving action of the jth bone on the i-th model vertex, indicating the strength of the bone driving the vertex; T bj is the first The motion deformation matrix of j skeletons themselves, rot(T bj ) is the rotation part of the deformation matrix.
PCT/CN2019/086889 2018-05-15 2019-05-14 Three-dimensional reconstruction method and device uniting rigid motion and non-rigid deformation WO2019219012A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810460091.5A CN108711185B (en) 2018-05-15 2018-05-15 Three-dimensional reconstruction method and device combining rigid motion and non-rigid deformation
CN201810460091.5 2018-05-15

Publications (1)

Publication Number Publication Date
WO2019219012A1 true WO2019219012A1 (en) 2019-11-21

Family

ID=63869046

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/086889 WO2019219012A1 (en) 2018-05-15 2019-05-14 Three-dimensional reconstruction method and device uniting rigid motion and non-rigid deformation

Country Status (2)

Country Link
CN (1) CN108711185B (en)
WO (1) WO2019219012A1 (en)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108711185B (en) * 2018-05-15 2021-05-28 清华大学 Three-dimensional reconstruction method and device combining rigid motion and non-rigid deformation
CN109785440A (en) * 2018-12-18 2019-05-21 合肥阿巴赛信息科技有限公司 A kind of curved surface distorted pattern
CN109829972B (en) * 2019-01-19 2023-06-02 北京工业大学 Three-dimensional human standard skeleton extraction method for continuous frame point cloud
CN109840940B (en) * 2019-02-11 2023-06-27 清华-伯克利深圳学院筹备办公室 Dynamic three-dimensional reconstruction method, device, equipment, medium and system
CN111768504B (en) * 2019-03-30 2023-07-14 华为技术有限公司 Model processing method, deformation control method and related equipment
CN110070595B (en) * 2019-04-04 2020-11-24 东南大学深圳研究院 Single image 3D object reconstruction method based on deep learning
CN110006408B (en) * 2019-04-17 2020-04-24 武汉大学 LiDAR data cloud control aerial image photogrammetry method
CN111862139B (en) * 2019-08-16 2023-08-18 中山大学 Dynamic object parametric modeling method based on color-depth camera
CN110689625B (en) * 2019-09-06 2021-07-16 清华大学 Automatic generation method and device for customized face mixed expression model
CN111968169B (en) * 2020-08-19 2024-01-19 北京拙河科技有限公司 Dynamic human body three-dimensional reconstruction method, device, equipment and medium
CN113096249B (en) * 2021-03-30 2023-02-17 Oppo广东移动通信有限公司 Method for training vertex reconstruction model, image reconstruction method and electronic equipment
CN112991524B (en) * 2021-04-20 2022-03-25 北京的卢深视科技有限公司 Three-dimensional reconstruction method, electronic device and storage medium
CN114373018A (en) * 2021-12-06 2022-04-19 聚好看科技股份有限公司 Real-time driving method, device and equipment
CN114648613B (en) * 2022-05-18 2022-08-23 杭州像衍科技有限公司 Three-dimensional head model reconstruction method and device based on deformable nerve radiation field
CN115082512A (en) * 2022-07-08 2022-09-20 北京大学深圳研究生院 Point cloud motion estimation method and device, electronic equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106469465A (en) * 2016-08-31 2017-03-01 深圳市唯特视科技有限公司 A kind of three-dimensional facial reconstruction method based on gray scale and depth information
CN107358645A (en) * 2017-06-08 2017-11-17 上海交通大学 Product method for reconstructing three-dimensional model and its system
US20180061018A1 (en) * 2016-08-24 2018-03-01 Korea Institute Of Science And Technology Method of multi-view deblurring for 3d shape reconstruction, recording medium and device for performing the method
CN108711185A (en) * 2018-05-15 2018-10-26 清华大学 Joint rigid moves and the three-dimensional rebuilding method and device of non-rigid shape deformations

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8285499B2 (en) * 2009-03-16 2012-10-09 Apple Inc. Event recognition
CN102800103B (en) * 2012-06-18 2015-02-18 清华大学 Unmarked motion capturing method and device based on multi-visual angle depth camera
CN103198523B (en) * 2013-04-26 2016-09-21 清华大学 A kind of three-dimensional non-rigid body reconstruction method based on many depth maps and system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180061018A1 (en) * 2016-08-24 2018-03-01 Korea Institute Of Science And Technology Method of multi-view deblurring for 3d shape reconstruction, recording medium and device for performing the method
CN106469465A (en) * 2016-08-31 2017-03-01 深圳市唯特视科技有限公司 A kind of three-dimensional facial reconstruction method based on gray scale and depth information
CN107358645A (en) * 2017-06-08 2017-11-17 上海交通大学 Product method for reconstructing three-dimensional model and its system
CN108711185A (en) * 2018-05-15 2018-10-26 清华大学 Joint rigid moves and the three-dimensional rebuilding method and device of non-rigid shape deformations

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
FEI TINGTING: "Virtual human reconstruction and skinning animation algorithm based on depth data", INFORMATION SCIENCE & TECHNOLOGY, CHINA MASTER'S THESES, no. 10, 15 October 2016 (2016-10-15), ISSN: 1674-0246 *

Also Published As

Publication number Publication date
CN108711185A (en) 2018-10-26
CN108711185B (en) 2021-05-28

Similar Documents

Publication Publication Date Title
WO2019219012A1 (en) Three-dimensional reconstruction method and device uniting rigid motion and non-rigid deformation
WO2019219013A1 (en) Three-dimensional reconstruction method and system for joint optimization of human body posture model and appearance model
CN110335343B (en) Human body three-dimensional reconstruction method and device based on RGBD single-view-angle image
CN108154550B (en) RGBD camera-based real-time three-dimensional face reconstruction method
CN110648398B (en) Real-time ortho image generation method and system based on unmanned aerial vehicle aerial data
WO2019219014A1 (en) Three-dimensional geometry and eigencomponent reconstruction method and device based on light and shadow optimization
Huang et al. Leveraging motion capture and 3d scanning for high-fidelity facial performance acquisition
CN103810685B (en) A kind of super-resolution processing method of depth map
KR102097016B1 (en) Apparatus and methdo for analayzing motion
CN108629831B (en) Three-dimensional human body reconstruction method and system based on parameterized human body template and inertial measurement
CN110189399B (en) Indoor three-dimensional layout reconstruction method and system
US20170330375A1 (en) Data Processing Method and Apparatus
CN108053437B (en) Three-dimensional model obtaining method and device based on posture
CN110599540A (en) Real-time three-dimensional human body shape and posture reconstruction method and device under multi-viewpoint camera
KR102152436B1 (en) A skeleton processing system for dynamic 3D model based on 3D point cloud and the method thereof
JP2012185772A (en) Method and program for enhancing accuracy of composited picture quality of free viewpoint picture using non-fixed zoom camera
CN112927362A (en) Map reconstruction method and device, computer readable medium and electronic device
CN113077519B (en) Multi-phase external parameter automatic calibration method based on human skeleton extraction
CN112734890A (en) Human face replacement method and device based on three-dimensional reconstruction
US11640687B2 (en) Volumetric capture and mesh-tracking based machine learning 4D face/body deformation training
WO2021063271A1 (en) Human body model reconstruction method and reconstruction system, and storage medium
KR20150068895A (en) Apparatus and method for generating three dimensional output data
Lan et al. Development of a virtual reality teleconference system using distributed depth sensors
EP2800055A1 (en) Method and system for generating a 3D model
CN111047678B (en) Three-dimensional face acquisition device and method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19803388

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19803388

Country of ref document: EP

Kind code of ref document: A1