WO2022205760A1 - Three-dimensional human body reconstruction method and apparatus, and device and storage medium - Google Patents

Three-dimensional human body reconstruction method and apparatus, and device and storage medium Download PDF

Info

Publication number
WO2022205760A1
WO2022205760A1 PCT/CN2021/115122 CN2021115122W WO2022205760A1 WO 2022205760 A1 WO2022205760 A1 WO 2022205760A1 CN 2021115122 W CN2021115122 W CN 2021115122W WO 2022205760 A1 WO2022205760 A1 WO 2022205760A1
Authority
WO
WIPO (PCT)
Prior art keywords
human body
model
dimensional
image
texture
Prior art date
Application number
PCT/CN2021/115122
Other languages
French (fr)
Chinese (zh)
Inventor
宋勃宇
邓又铭
刘文韬
钱晨
Original Assignee
深圳市慧鲤科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市慧鲤科技有限公司 filed Critical 深圳市慧鲤科技有限公司
Publication of WO2022205760A1 publication Critical patent/WO2022205760A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques

Definitions

  • the present disclosure relates to image processing technology, and in particular, to a three-dimensional human body reconstruction method, device, equipment and storage medium.
  • 3D human body reconstruction is an important problem in the field of computer vision and computer graphics.
  • the reconstructed human digital model has important applications in many fields, such as body measurement, virtual fitting, virtual anchor, game character custom design, virtual reality social networking and other fields.
  • how to project the human body in the real world into the virtual world to obtain a three-dimensional human body digital model is an important issue.
  • the digital reconstruction of the 3D human body is very complicated, requiring the scanner to perform continuous scanning at multiple angles without dead ends around the scanning target; and the reconstruction results have the problem that the local reconstruction effect is not fine enough.
  • the embodiments of the present disclosure provide at least a three-dimensional human body reconstruction method, apparatus, device, and storage medium.
  • a three-dimensional human body reconstruction method comprising:
  • reconstruction of the human body texture of the target human body is performed to obtain a three-dimensional human body model of the target human body.
  • the performing geometric reconstruction of the human body based on the human body image of the target human body to obtain the three-dimensional mesh model of the target human body includes: performing three-dimensional reconstruction of the human body image of the target human body through a first deep neural network branch, obtaining a first human body model; performing three-dimensional reconstruction on a partial image in the human body image through a second deep neural network branch to obtain a second human body model; wherein, the partial image includes a partial area of the target human body; The first human body model and the second human body model are fused to obtain a fused human body model; the fused human body model is meshed to obtain a three-dimensional mesh model of the target human body.
  • the first deep neural network branch includes: a global feature sub-network and a first fitting sub-network;
  • the second deep neural network branch includes: a local feature sub-network and a second fitting sub-network;
  • the three-dimensional reconstruction of the human body image of the target human body through the first deep neural network branch to obtain a first human body model includes: performing feature extraction on the human body image through the global feature sub-network to obtain first image features;
  • the first human body model is obtained based on the first image features through the first fitting sub-network;
  • the second human body is obtained by performing three-dimensional reconstruction of the partial image in the human body image through the second deep neural network branch
  • the model includes: extracting features from the local image through the local feature sub-network to obtain second image features; using the second fitting sub-network based on the second image features and the first fitter
  • the intermediate features output by the network are obtained to obtain the second human body model.
  • performing local geometric reconstruction on a local part of the target human body based on the human body image of the target human body to obtain a three-dimensional mesh model of the local part including: performing a human body image on the target human body Perform feature extraction to obtain a third image feature; and determine a three-dimensional mesh model of the partial portion according to the third image feature and the three-dimensional topology template of the partial portion.
  • the obtaining the initial 3D model by fusing the 3D mesh model of the local part with the 3D mesh model of the target body includes: obtaining the local body image according to the body image of the target body multiple key points of the part; determine the information of the first model key point corresponding to the multiple key points on the 3D mesh model of the target body, and determine the multiple key points in the local part The information of the corresponding second model key points on the three-dimensional mesh model; based on the information of the first model key points and the information of the second model key points, the three-dimensional mesh model of the local part is fused to the The three-dimensional mesh model of the target human body is obtained to obtain the initial three-dimensional model.
  • the 3D mesh model of the local part is fused to the 3D mesh model of the target human body based on the information of the key points of the first model and the information of the key points of the second model
  • Obtaining the initial three-dimensional model includes: determining the three-dimensional mesh model of the target body and the three-dimensional mesh model of the local part based on the information of the key points of the first model and the information of the key points of the second model
  • the three-dimensional grid model of the local part is transformed into the coordinate system of the three-dimensional grid model of the target body; in the transformed coordinate system, the local The three-dimensional mesh model of the part is fused to the three-dimensional mesh model of the target body to obtain the initial three-dimensional model.
  • the human body image includes: a frontal texture and a background image of the target human body; the reconstruction of the human body texture of the target human body is performed according to the initial three-dimensional model and the human body image to obtain the
  • the three-dimensional human body model of the target human body includes: performing human body segmentation on the human body image to obtain a first segmentation mask, a second segmentation mask and a frontal texture of the target human body; wherein, the first segmentation mask corresponds to the The mask area of the front texture, the second segmentation mask corresponds to the mask area of the back texture of the target body; the front texture, the first segmentation mask and the second segmentation mask , input the texture generation network to obtain the back texture of the target body; based on the back texture and the front texture, obtain a textured three-dimensional body model corresponding to the target body.
  • the training of the texture generation network includes the following processing: performing human body segmentation on images of human body samples in the training sample image set to obtain a first sample segmentation mask, a second sample segmentation mask and the human body sample , wherein the first sample segmentation mask corresponds to the mask area of the front texture of the human sample, and the second sample segmentation mask corresponds to the mask area of the back texture of the human sample; according to The frontal texture of the human body, the third sample segmentation mask and the fourth sample segmentation mask in the auxiliary human body image, and the auxiliary texture generation network is trained, wherein the auxiliary human body image is obtained by reducing the resolution of the image of the human body sample.
  • the third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human image
  • the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human image
  • the partial part of the target human body is the face of the target human body; and/or the human body image is an RGB image.
  • the method further includes: when performing the human body geometry reconstruction based on the human body image of the target human body, also obtaining the human skeleton structure of the target human body; after the obtaining the three-dimensional human body model of the target human body, based on the The three-dimensional human body model and the human skeleton structure determine skin weights for driving the three-dimensional human body model.
  • a three-dimensional human body reconstruction device comprising:
  • an overall reconstruction module used for performing geometric reconstruction of the human body based on the human body image of the target human body to obtain a three-dimensional mesh model of the target human body;
  • a local reconstruction module configured to perform local geometric reconstruction on the local part of the target human body based on the human body image of the target human body to obtain a three-dimensional mesh model of the local part
  • a fusion processing module configured to fuse the 3D mesh model of the local part with the 3D mesh model of the target human body to obtain an initial 3D model
  • the texture reconstruction module is used for reconstructing the human body texture of the target human body according to the initial three-dimensional model and the human body image, so as to obtain the three-dimensional human body model of the target human body.
  • the method when the overall reconstruction module is used to obtain the three-dimensional mesh model of the target human body, the method includes: performing three-dimensional reconstruction on the human body image of the target human body through a first deep neural network branch to obtain the first deep neural network branch. a human body model; three-dimensional reconstruction is performed on a partial image in the human body image through a second deep neural network branch to obtain a second human body model; wherein, the partial image includes a partial area of the target human body; the first human body The model and the second human body model are fused to obtain a fused human body model; the fused human body model is meshed to obtain a three-dimensional mesh model of the target human body.
  • the local reconstruction module is specifically configured to: perform feature extraction on the human body image of the target human body to obtain third image features; according to the third image features and the three-dimensional topology template of the local part , and determine the three-dimensional mesh model of the local part.
  • the fusion processing module is specifically configured to: obtain multiple key points of the local part according to the human body image of the target human body; determine that the multiple key points are in a three-dimensional network of the target human body information on the key points of the first model corresponding to the grid model, and determining the information on the key points of the second model corresponding to the plurality of key points on the three-dimensional mesh model of the local part; based on the key points of the first model point information and the information of the key points of the second model, and fuse the three-dimensional mesh model of the local part into the three-dimensional mesh model of the target body to obtain the initial three-dimensional model.
  • the fusion processing module is configured to fuse the three-dimensional mesh model of the local part to the target based on the information of the key points of the first model and the information of the key points of the second model
  • the three-dimensional mesh model of the human body when the initial three-dimensional model is obtained, including: based on the information of the key points of the first model and the information of the key points of the second model, determining the three-dimensional mesh model of the target human body and the The coordinate transformation relationship between the three-dimensional mesh models of the local parts; according to the coordinate transformation relationship, the three-dimensional mesh model of the local parts is transformed into the coordinate system of the three-dimensional mesh model of the target body; after the transformation The 3D mesh model of the local part is fused to the 3D mesh model of the target human body under the coordinate system of 1 to obtain the initial 3D model.
  • the texture reconstruction module is specifically configured to: perform human body segmentation on the human body image to obtain a first segmentation mask, a second segmentation mask and a frontal texture of the target human body; wherein the first segmentation mask A segmentation mask corresponds to the mask area of the front texture, and the second segmentation mask corresponds to the mask area of the back texture of the target human body; the front texture, the first segmentation mask and all the The second segmentation mask is input into a texture generation network to obtain the back texture of the target body; based on the back texture and the front texture, a textured 3D body model corresponding to the target body is obtained.
  • the apparatus further includes: a model training module for training the texture generation network, including: performing human body segmentation on images of human body samples in the training sample image set to obtain a first sample segmentation mask, The second sample segmentation mask and the frontal texture of the human sample, wherein the first sample segmentation mask corresponds to the mask area of the frontal texture of the human sample, and the second sample segmentation mask corresponds to the The mask area of the back texture of the human sample; according to the front texture of the human body in the auxiliary human image, the third sample segmentation mask and the fourth sample segmentation mask, the auxiliary texture generation network is trained, wherein, by reducing the image of the human sample The resolution of the auxiliary human body image is obtained, the third sample segmentation mask corresponds to the mask area of the frontal texture of the human body in the auxiliary human body image, and the fourth sample segmentation mask corresponds to the The mask area of the back texture of the human body; after the auxiliary texture generation network training is completed, based on the front texture of the human sample, the first sample segmentation
  • an electronic device comprising: a memory and a processor, where the memory is used for storing computer-readable instructions, and the processor is used for invoking the computer instructions to implement any of the embodiments of the present disclosure.
  • a computer-readable storage medium on which a computer program is stored, and when the computer program is executed by a processor, implements the method described in any of the embodiments of the present disclosure.
  • a computer program product including a computer program, which implements the method described in any embodiment of the present disclosure when the computer program is executed by a processor.
  • the three-dimensional human body reconstruction method, device, device, and storage medium provided by the embodiments of the present disclosure perform local geometric reconstruction on a local part of the target human body, and combine the three-dimensional mesh model of the local part obtained by the local geometric reconstruction with the three-dimensional mesh model of the target human body.
  • the mesh model is fused, so that the local parts in the 3D mesh model of the target body are more clear, fine and accurate, and the reconstruction effect of the local parts is improved; It simplifies the user's cooperation process and makes the three-dimensional human body reconstruction easier.
  • FIG. 1 shows a flowchart of a three-dimensional human body reconstruction method provided by at least one embodiment of the present disclosure
  • FIG. 2 shows a schematic diagram of a manner for obtaining a 3D mesh model based on a single human body image provided by at least one embodiment of the present disclosure
  • FIG. 3 shows a schematic diagram of an acquisition process of an initial three-dimensional model provided by at least one embodiment of the present disclosure
  • FIG. 4 shows a schematic diagram of a texture reconstruction process provided by at least one embodiment of the present disclosure
  • FIG. 5 shows a schematic diagram of a skin weight determination process provided by at least one embodiment of the present disclosure
  • FIG. 6 shows a schematic diagram of a manner for obtaining a three-dimensional mesh model based on a single human body image provided by at least one embodiment of the present disclosure
  • FIG. 7 shows a schematic diagram of the principle of texture generation provided by at least one embodiment of the present disclosure
  • FIG. 8 shows a schematic diagram of a training process of a texture generation network provided by at least one embodiment of the present disclosure
  • FIG. 9 shows a schematic diagram of a human body image provided by at least one embodiment of the present disclosure.
  • FIG. 10 shows a structural diagram of a three-dimensional human body reconstruction apparatus provided by at least one embodiment of the present disclosure
  • FIG. 11 shows a structural diagram of a three-dimensional human body reconstruction apparatus provided by at least one embodiment of the present disclosure.
  • 3D human body reconstruction has important applications in many fields, including but not limited to the following application scenarios:
  • the realism of some virtual reality application scenarios can be enhanced through 3D human reconstruction.
  • 3D human reconstruction For example, virtual fitting, virtual cloud meeting, virtual classroom, etc.
  • the 3D human body model obtained by 3D human body reconstruction can be imported into the game data to complete the generation of the personalized character.
  • 3D human body reconstruction has the following requirements: On the one hand, the user's cooperation process should be simplified as much as possible. Bad experience. On the other hand, try to obtain a 3D human body model with higher accuracy. For example, in scenarios such as virtual cloud conferences or AR virtual interaction scenarios, the 3D human body model obtained from 3D human body reconstruction has a higher sense of realism and immersion. need.
  • an embodiment of the present disclosure provides a three-dimensional human body reconstruction method, which aims to perform three-dimensional human body reconstruction of the user based on a photo of the user, simplify the user's cooperation process, and achieve high-precision reconstruction Effect.
  • FIG. 1 shows a flowchart of a three-dimensional human body reconstruction method provided by at least one embodiment of the present disclosure.
  • the method may include steps 100 to 106 .
  • step 100 the geometric reconstruction of the human body is performed based on the single human body image of the target human body to obtain a three-dimensional mesh model of the target human body.
  • the target human body is the basic user of the 3D human body reconstruction.
  • 3D human body reconstruction is performed on user Xiao Zhang
  • Xiao Zhang can be called the target human body
  • the reconstructed 3D human body model is also obtained based on Xiao Zhang's body, which is similar to Xiao Zhang's posture, appearance, clothing and hairstyle. have high similarity.
  • the single human body image is a human body image of the target human body.
  • the embodiment of the present disclosure has no special requirements on the collection method and format of the human body image.
  • the single human body image may be a frontal photograph of the target human body.
  • the single human body image may be an RGB color image.
  • the acquisition cost of this RGB format image is low. For example, it is not necessary to use high-cost equipment such as a depth-of-field camera during image acquisition, and it can be acquired by ordinary shooting equipment.
  • the human body geometry can be reconstructed based on the single human body image of the target human body to obtain a three-dimensional mesh model.
  • the three-dimensional mesh model is a three-dimensional mesh Mesh representing the human body geometry, and the mesh includes several vertices and faces.
  • the three-dimensional mesh Mesh obtained by the above reconstruction and a pre-stored parameterized human body model can also be aligned and fitted with respect to the posture and body shape.
  • the parameterized human body model includes a mesh of the human body surface and a set of skeletal structures, which are controlled by a set of pose and body shape parameters, and the skeleton position and surface shape of the human body will change as the parameter values change.
  • Fig. 2 illustrates a method of obtaining a 3D mesh model based on a single human image reconstruction.
  • a single human body image 21 of the target human body can be input into the first deep neural network branch 22 for three-dimensional reconstruction.
  • the first deep neural network branch 22 may include a global feature sub-network 221 and a first fitting sub-network 222 .
  • the features of the single human body image 21 can be extracted through the global feature sub-network 221 to obtain high-level image features of the single human body image 21, and the high-level image features can be referred to as first image features.
  • the global feature sub-network 221 may be a HourGlass convolutional network.
  • the first image feature is input to the first fitting sub-network 222, and the first fitting sub-network 222 can predict whether each voxel block in the three-dimensional space belongs to the interior of the target human body according to the first image feature.
  • the first fitting sub-network 222 may be a multilayer perceptron structure.
  • the first fitting sub-network 222 outputs a first human body model including each three-dimensional voxel block located inside the target human body.
  • the meshing process may continue to be performed on the first human body model.
  • the meshing process may be to apply the MarchingCubes algorithm in the voxel space to the first human body model to obtain a three-dimensional mesh model of the target human body.
  • step 102 based on the single human body image of the target human body, a local high-definition geometric reconstruction is performed on a local part of the target human body to obtain a three-dimensional mesh model of the local part.
  • the three-dimensional mesh model of the target human body reconstructed in step 100 may be blurred in local parts of the target human body.
  • the local part may be a human face, or may be other local parts, such as a hand and other parts that need to reflect detailed features.
  • the above-mentioned 3D mesh model is relatively vague in the face details of the target human body, but the face is usually the area that the user pays more attention to. Therefore, in this step, the partial parts of the target human body can be individually geometrically reconstructed.
  • the reconstruction of the human face can use a fine reconstruction of a fixed topology, that is, the three-dimensional topology of the human face can be reconstructed based on the image features obtained by feature extraction from a single human image of the target human body.
  • the position of each vertex in the template is fitted to obtain a three-dimensional mesh model of the face.
  • the semantic structure of the human face is consistent, so a 3D human face with a fixed topology structure can be used as a template, and the template can be called a 3D topology template of the human face.
  • the template includes a plurality of vertices, each vertex is fixedly corresponding to a face semantics, for example, one vertex represents the tip of the nose, and the other vertex represents the corner of the eye.
  • each vertex position of the above-mentioned three-dimensional topology template of the face can be obtained by regression through a deep neural network.
  • the deep neural network may include a deep convolutional network and a graph convolutional network, and a single human body image of the target human body may be input into the deep convolutional network to extract image features, and the extracted features may be referred to as the third image feature. Then the third image feature and the 3D topology template of the face are used as the input of the graph convolution network, and finally a 3D mesh model of the face output by the graph convolution network is obtained, and the 3D network model is closer to the target human face.
  • the input of the deep convolutional network may also be a partial image area containing a face captured from a single human body image of the target human body.
  • step 104 the 3D mesh model of the local part is fused with the 3D mesh model of the target human body to obtain an initial 3D model.
  • the 3D mesh model of the target human body reconstructed in step 100 may be somewhat blurred in the local part of the human body, and the local part is taken as an example of a human face, and in step 102, the 3D mesh model of the human face is obtained through the separate geometric reconstruction of the face Grid model, in this step, the corresponding part in the 3D grid model of the target body in step 100 can be replaced by the 3D grid model of the face, so that the 3D grid model of the target body can be retained.
  • Information such as head shape, body shape, and posture can also make the facial features more refined and accurate, and achieve better reconstruction effects.
  • the partial part is a human face as an example here, and other partial parts can also be independently reconstructed to make it clearer in actual implementation.
  • a single human body image of the target human body may be input into a pre-trained key point detection model, and a plurality of key points of local parts of the target human body in the image may be determined by the key point detection model.
  • FIG. 3 still taking the local part of the human face as an example, after acquiring multiple key points 31 of the human face, it can be determined according to the coordinates of these key points 31 on the human face that the key points are located in the target human body.
  • the 3D mesh model of the face, and the corresponding model key points on the 3D mesh model of the face can be determined.
  • the information may include the key point identifiers of each first model key point and the corresponding key points. key point location. It is also possible to determine the information of the key points of the second model corresponding to the multiple key points of the face on the three-dimensional mesh model of the face, for example, the information may include the key point identification of each key point of the second model and the corresponding key point. point location.
  • the three-dimensional mesh model of the face can be fused to the information of the key points of the first model and the information of the key points of the second model.
  • the 3D mesh model of the target human body is obtained, and the initial 3D model is obtained.
  • fusing the 3D mesh model of the face into the 3D mesh model of the target body includes: based on the information of the key points of the first model and the information of the key points of the second model, and combining the cameras of the two models
  • the external parameter determines the coordinate transformation relationship between the 3D mesh model of the target body and the 3D mesh model of the face; based on the coordinate transformation relationship, the 3D mesh model of the face can be transformed into the 3D mesh model of the target body In the transformed coordinate system, the 3D mesh model of the face is fused to the 3D mesh model of the target body.
  • the facial geometry on the 3D mesh model of the target body can be removed.
  • the 3D mesh model of the face and the 3D mesh model of the target body are integrated into a whole by means of Poisson reconstruction, and the obtained model can be called the initial 3D model.
  • the initial 3D model already has relatively clear facial features, similar head shape, body shape and other information, and the accuracy is high.
  • step 106 reconstruction of the human body texture of the target human body is performed according to the initial three-dimensional model and the single human body image to obtain a three-dimensional human body model with colored textures of the target human body.
  • this embodiment performs three-dimensional human body reconstruction based on a single human body image of the target human body, part of the human body area is invisible. For example, if the frontal human body image of the target human body is used for reconstruction, the back of the target human body is invisible. , which will cause missing textures. Therefore, in this step, the prediction and completion of the human body texture in the invisible area of the target human body can be performed according to the initial three-dimensional model and the single human body image of the target human body, and the human body texture in the single human body image can be fused. Then a textured 3D human body model is generated.
  • the deep learning network can be used to predict the human body back texture 41, and combine the human body back texture 41 with the front of the human body in the single human body image.
  • Texture 42 performing texture mapping on the initial 3D model, that is, performing texture reconstruction on the initial 3D model.
  • the three-dimensional model 43 in FIG. 4 has mapped the above-mentioned back and front textures of the human body on the initial three-dimensional model.
  • the initial three-dimensional model obtained in step 104 is a mesh Mesh of human body geometry, and this step is to add human body texture to the model on the basis of the mesh model.
  • the interpolation technology can be used to fill some gaps in the model with textures, so as to complete the texture of the initial 3D model, and obtain the 3D human body model 44 of the target human body.
  • the three-dimensional human body reconstruction method of this embodiment performs local geometric reconstruction on a local part of the target human body, and fuses the three-dimensional mesh model of the local part obtained by the local geometric reconstruction with the three-dimensional mesh model of the target human body, so that the target human body can be reconstructed.
  • the local parts in the initial 3D model are more clear, fine and accurate, which improves the reconstruction effect of the local parts; moreover, this method is based on the single human body image of the target human body for reconstruction, which also simplifies the user's cooperation process, so that the three-dimensional human body can be reconstructed. Rebuilding is easier.
  • the skin weight for driving the three-dimensional human body model can be determined based on the three-dimensional human body model and the human skeleton structure of the target human body.
  • the skin weight is used to drive the built 3D human model. For example, if you want to drive the 3D human model to do various actions, you need to bind the model to the human skeleton structure. Binding the model to the bone is called a mask. Skin. Then, the model can be driven by the movement of the bones, and the skin weight is used to represent the influence of the nodes of the bones on the model vertices. According to the skin weight, the size of each vertex in the 3D human model can be controlled by the influence of each bone joint point. , so as to better control the movement of the model.
  • calculating the skin weight of the three-dimensional human body model may include the following processing: in step 100, the human skeleton structure has been obtained according to the single human body image of the target human body, and in this step, the human skeleton structure and the obtained three-dimensional human body can be obtained.
  • the model is input to the deep learning network, and the skin weight of the model is automatically obtained through the deep learning network.
  • the attribute features corresponding to the vertices in the three-dimensional human body model 51 may be generated first according to the three-dimensional human body model 51 and the human skeleton structure 52 .
  • the attribute feature can be constructed by using the spatial positional relationship between each vertex and the human skeleton structure.
  • the attribute features of the vertex can include the following four features:
  • K is a positive integer.
  • the attribute features of each vertex can be used as the input of the spatial graph convolutional attention network in the deep learning network.
  • the above features can be transformed into hidden layer features through a multilayer perceptron.
  • the spatial graph convolutional attention network can predict the weight of each vertex affected by each of the above K skeletal joint points according to the above hidden layer features, and the latter multi-layer perceptron in the deep learning network can be used for this.
  • the weights are normalized so that for a certain vertex, the sum of the influence weights of each bone joint point on the vertex is 1.
  • the weight corresponding to each vertex in the finally obtained 3D human model and affected by each skeleton joint point is the skin weight of the vertex.
  • the three-dimensional human body reconstruction method of this embodiment can obtain the human skeleton structure according to a single human body image of the target human body, and automatically calculate the skin weight according to the human skeleton structure and the reconstructed three-dimensional human body model, which not only ensures that different input images are
  • the semantic structure of the lower bones is consistent, and appropriate skin weights can be quickly generated in combination with different clothing and apparel shapes.
  • the semantic consistency of the skeleton can facilitate the registration of the model and the ready-made action library.
  • the advantage of the semantic consistency is that it is conducive to the application (registration) of the generated model and the skeleton and the action library.
  • the action library can store some human action sequences in advance, such as dancing, boxing, etc.
  • the action library stores a series of motion bones.
  • the present disclosure provides a method for three-dimensional human body reconstruction in another embodiment.
  • the reconstruction process of this embodiment is different in that the human body image is performed on the single human body image based on the target human body in step 100 .
  • the process of geometric reconstruction has been improved to improve the geometric reconstruction accuracy of the reconstructed 3D mesh model of the target body.
  • the same processing steps as the embodiment in FIG. 1 will not be described in detail, and only the differences will be mainly described.
  • a second deep neural network branch 61 is added.
  • the second deep neural network branch 61 may include: a local feature sub-network 611 and a second fitting sub-network 612 .
  • An image of a local area can be extracted from the single human body image 21 of the target human body to obtain a local image 62 , and the second deep neural network is used for three-dimensional reconstruction of the local image 62 .
  • the body region of the target human body included in the partial image here may not be exactly the same as the partial part corresponding to the local geometric reconstruction in step 102.
  • the partial image here may include the area above the shoulder of the target human body , and the local part reconstructed in step 102 may be the face of the target human body.
  • the reconstruction above the shoulder of the target human body in FIG. 6 is just an example, and refined geometric reconstruction can also be performed on other human body regions of the target human body.
  • the first human body model is reconstructed through the first deep neural network branch 22, and the partial image 62 is input into the second deep neural network branch 61, and the partial image is processed by the partial feature sub-network 611. Feature extraction to obtain second image features. Then, a second human body model is obtained through the second fitting sub-network 612 based on the second image feature and the intermediate feature output by the first fitting sub-network 222 .
  • the intermediate features may be the features output by part of the network structure in the first fitting sub-network 222.
  • the first fitting sub-network 222 includes a certain number of fully connected layers, then the The outputs of the partial number of fully connected layers are input to the second fitting sub-network 612 as the intermediate features.
  • the structure of the second deep neural network branch 61 may be basically the same as that of the first deep neural network branch 22, for example, the global feature sub-network 221 in the first deep neural network branch 22 may include four Blocks, Each block may include a certain number of feature extraction layers such as convolution layers and pooling layers, and the local feature sub-network 611 in the second deep neural network branch 61 may include one of the above-mentioned blocks.
  • the first human body model and the second human body model may be fused to obtain a fused human body model. And continue to mesh the fused human body model to obtain a three-dimensional mesh model of the target human body.
  • the three-dimensional human body reconstruction method of this embodiment not only improves the reconstruction effect of local parts by performing local geometric reconstruction on the local parts of the target human body, but also performs reconstruction based on a single human body image of the target human body, which simplifies the cooperation process of users; , and also reconstruct the local image through the second deep neural network, which improves the reconstruction effect of the local human body area of the target human body.
  • the present disclosure provides a three-dimensional human body reconstruction method in yet another embodiment.
  • the reconstruction process of the further embodiment provides a specific method for predicting the back texture of the human body through a deep learning network.
  • the same processing steps as the embodiment in FIG. 1 will not be described in detail, and only the differences will be mainly described.
  • a single human body image of the target human body sometimes includes a background image and a frontal texture of the human body.
  • image segmentation can be performed first to segment the frontal texture of the human body, and then predict the human body based on the frontal texture. back texture.
  • the frontal image 71 of the target human body may be segmented to obtain a first segmentation mask 72 and the segmented frontal texture 73 of the target human body.
  • the first segmentation mask 72 is horizontally flipped to obtain a second segmentation mask 74, and then the front texture 73, the first segmentation mask 72 and the second segmentation mask 74 are input into the texture generation network 75, and finally the The texture generation network 75 outputs the back texture of the target body.
  • FIG. 7 is an example of obtaining the second segmentation mask 74 by horizontally flipping the first segmentation mask 72.
  • the actual implementation is not limited to this.
  • the frontal image of the target human body can be input into a pre-training
  • the neural network directly outputs the first segmentation mask and the second segmentation mask.
  • the front and back textures of the target human body can be mapped to the initial three-dimensional model of the human body, and the three-dimensional human body model of the target human body can be obtained.
  • the above-mentioned training process of the texture generation network 75 may include the following processing: please refer to FIG. 8 in combination, an auxiliary texture generation network 76 may be used.
  • the auxiliary texture generation network 76 may include a part of the network structure of the texture generation network 75 .
  • the texture generation network 75 may add a certain number of convolution layers to the auxiliary texture generation network 76 .
  • the auxiliary texture generation network can be trained according to the auxiliary human body image, the third sample segmentation mask and the fourth sample segmentation mask in the training sample image set, and after the auxiliary texture generation network is trained, the auxiliary texture generation network can be generated. At least part of the network parameters of the network are initialized as part of the texture generation network parameters, and the texture generation network is continued to be trained based on the frontal texture of the human sample, the first sample segmentation mask and the second sample segmentation mask.
  • the auxiliary human image is obtained by reducing the resolution of a single image of the human sample
  • the first sample segmentation mask corresponds to the mask area of the front texture of the human sample
  • the second sample segmentation mask corresponds to the mask of the back texture of the human sample
  • the third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human image
  • the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human image.
  • the frontal texture 82 of the human body in the auxiliary human body image 81, the third sample segmentation mask 83 and the fourth sample segmentation mask 84 can be obtained by performing image segmentation on the auxiliary human body image 81, and input auxiliary human body image 81.
  • the texture generation network 76 obtains the first predicted value of the back texture of the human body in the auxiliary human body image 81; and then adjusts the auxiliary texture based on the first predicted value and the first real value of the back texture of the human body in the auxiliary human body image 81 Network parameters for the network 76 are generated. After several iterations, the trained auxiliary texture generation network 76 can be obtained.
  • the training supervision of the auxiliary texture generation network in addition to the loss calculated based on the first predicted value and the first real value, may also include other losses based on the first predicted value, for example, based on the auxiliary body image and the first predicted value.
  • the auxiliary human body image can be obtained by reducing the resolution of the frontal human body image 71 in FIG. 7 .
  • the resolution of the frontal texture 82 of the human body in the auxiliary human body image 81 is also higher than the resolution of the frontal texture 73 in FIG. 7 .
  • the third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human body image
  • the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human body image.
  • the network parameters of the auxiliary texture generation network can be used as the initialization of part of the network parameters of the texture generation network, that is, the network parameters of the texture generation network include: at least some of the network parameters. That is, the auxiliary texture generation network and the texture generation network share some network weights. Then, the frontal texture of the human body, the first sample segmentation mask and the second sample segmentation mask in the training sample image set used to train the texture generation network are input into the texture generation network to obtain the second prediction of the back texture of the human body sample value. Based on the second predicted value and the second real value of the back texture, network parameters of the texture generation network are adjusted. The resolution of the second real value is higher than the resolution of the first real value, that is, the resolution of the back texture output by the texture generation network is higher than the resolution of the back texture output by the auxiliary texture generation network.
  • the three-dimensional human body reconstruction method of this embodiment not only improves the reconstruction effect of local parts by performing local geometric reconstruction on the local parts of the target human body, but also performs reconstruction based on a single human body image of the target human body, which simplifies the cooperation process of users; It also automatically predicts the texture through the neural network, so that the generated texture effect is better, for example, the texture around the human body is more uniform and the color is more realistic; and, by training the auxiliary texture generation network first and then training the texture generation network, the texture is The training process of the generative network is more stable and easier to converge.
  • a plurality of images of the target body from different angles may also be acquired to comprehensively perform the three-dimensional reconstruction of the target body.
  • the three images may be acquired from different angles. Referring to FIG. 2 , the three images can be used as the input of the global feature sub-network 221 respectively, and a first image feature output by the global feature sub-network 221 corresponding to the three images can be obtained. Then, the three first image features are fused, and the image features obtained after fusion are used as the input of the first fitting sub-network 222 to continue processing.
  • the local images can also be obtained by extracting local regions from the three images, and the three Each of the local images is used as the input of the local feature sub-network 611, respectively, and the second image features output by the local feature sub-network 611 corresponding to the three local images are obtained, and then the three second image features are fused, and the result obtained after fusion is obtained.
  • the image features continue to be processed as input to the second fitting sub-network 612 .
  • the neural network models involved can be trained separately.
  • the first deep neural network branch and the texture generation network may each perform their own training.
  • the three-dimensional human body model of user U1 is to be constructed based on a single human body image of user U1.
  • the single human body image may be a frontal image of user U1, including the frontal texture and background image of user U1.
  • the single human body image 91 of the user U1 includes a front texture 92 and a background image 93 of the user.
  • One aspect of reconstruction is to perform geometric reconstruction of the human body based on the single human body image 91 to obtain the three-dimensional mesh model of U1 and the human skeleton structure.
  • the single human body image 91 can be processed through the network shown in FIG. 6 , and the single human body image 91 can be processed through the global feature sub-network and the first fitting sub-network in the first deep neural network branch to obtain:
  • the first human body model; the local feature sub-network and the second fitting sub-network in the second deep neural network branch process the image of the area above the human body shoulder in the single human body image 91 to obtain the second human body model.
  • a fused human body model is obtained.
  • the fused human body model is then meshed to obtain a three-dimensional mesh model (mesh) of the user U1.
  • Another aspect of reconstruction is to perform local geometric reconstruction on the face of the user U1 based on the single human body image 91 to obtain a three-dimensional mesh model of the face.
  • feature extraction can be performed on a single human body image 91, and the extracted image features and the three-dimensional face topology template are input into a graph convolutional neural network to obtain the face mesh of the user U1.
  • the face mesh (three-dimensional mesh model of the human face) obtained by the above reconstruction and the human body mesh of the user U1 (the three-dimensional mesh model of the human body of U1) can be combined to obtain the initial three-dimensional model of U1.
  • the identification and position of the key points of each model corresponding to the key points on the face mesh and the human mesh respectively can be determined, and based on the identification of these model key points and position, camera external parameters of the model and other parameters to determine the coordinate transformation relationship between the models.
  • the face mesh is transformed into the coordinate system of the human mesh, the face in the human mesh is replaced with the face mesh, and the face mesh and the human mesh are fused together through Poisson reconstruction to obtain the user U1 the initial 3D model.
  • the human body can be segmented on the single human body image 91 to obtain the human body frontal texture with the background image removed, and the first segmentation mask used to represent the frontal texture area of the human body.
  • a second segmentation mask representing the textured regions of the back of the human body.
  • texture mapping is performed on the initial 3D model based on the front texture and back texture, and the texture is filled and completed in the gap area of the model, and finally the 3D human model with texture U1 is obtained.
  • the skin weight of the 3D human model can also be calculated by combining the reconstructed 3D human model of U1 and the human skeleton structure obtained when reconstructing the 3D mesh model of U1. You can then drive the model to perform actions through this skin weight.
  • FIG. 10 illustrates a schematic structural diagram of a three-dimensional human body reconstruction apparatus.
  • the apparatus may include: an overall reconstruction module 1001 , a local reconstruction module 1002 , a fusion processing module 1003 and a texture reconstruction module 1004 .
  • the overall reconstruction module 1001 is configured to perform geometric reconstruction of the human body based on a single human body image of the target human body to obtain a three-dimensional mesh model of the target human body.
  • the local reconstruction module 1002 is configured to perform local geometric reconstruction on the local part of the target human body based on the single human body image of the target human body to obtain a three-dimensional mesh model of the local part.
  • the fusion processing module 1003 is configured to fuse the 3D mesh model of the local part with the 3D mesh model of the target human body to obtain an initial 3D model.
  • the texture reconstruction module 1004 is configured to reconstruct the human body texture of the target human body according to the initial three-dimensional model and the single human body image, so as to obtain a three-dimensional human body model of the target human body.
  • the overall reconstruction module 1001 when used to obtain the 3D mesh model of the target human body, it includes: performing 3D reconstruction on the single human body image of the target human body through the first deep neural network branch to obtain the first deep neural network branch. a human body model; three-dimensional reconstruction is performed on the partial image in the single human body image through the second deep neural network branch to obtain a second human body model; wherein, the partial image includes a partial area of the target human body; The first human body model and the second human body model are fused to obtain a fused human body model; the fused human body model is meshed to obtain a three-dimensional mesh model of the target human body.
  • the local reconstruction module 1002 is specifically configured to: perform feature extraction on a single human body image of the target human body to obtain a third image feature; according to the third image feature and the three-dimensional topology of the local part A template is used to determine the three-dimensional mesh model of the local part.
  • the fusion processing module 1003 is specifically configured to: obtain multiple key points of the local part according to a single human body image of the target human body; information on the key points of the first model corresponding to the grid model, and determining the information on the key points of the second model corresponding to the plurality of key points on the three-dimensional grid model of the local part; based on the first model The information of the key points and the information of the key points of the second model are fused with the three-dimensional mesh model of the local part into the three-dimensional mesh model of the target body to obtain the initial three-dimensional model.
  • the fusion processing module 1003 is configured to fuse the three-dimensional mesh model of the local part to the target human body based on the information of the key points of the first model and the information of the key points of the second model
  • it includes: based on the information of the key points of the first model and the information of the key points of the second model, determining the three-dimensional mesh model of the target human body and the The coordinate transformation relationship between the three-dimensional mesh models of the local parts; according to the coordinate transformation relationship, the three-dimensional mesh model of the local part is transformed into the coordinate system of the three-dimensional mesh model of the target body; The three-dimensional mesh model of the local part is fused to the three-dimensional mesh model of the target body under the coordinate system to obtain the initial three-dimensional model.
  • the texture reconstruction module 1004 is specifically configured to: perform human body segmentation on the single human body image to obtain a first segmentation mask, a second segmentation mask and a frontal texture of the target human body; wherein the first segmentation mask The segmentation mask corresponds to the mask area of the front texture, and the second segmentation mask corresponds to the mask area of the back texture of the target human body; the front texture, the first segmentation mask and the second segmentation mask are code, input the texture generation network to obtain the back texture of the target body; based on the back texture and the front texture, obtain a textured three-dimensional body model corresponding to the target body.
  • the apparatus may further include: a model training module 1005 .
  • the model training module 1005 is used to perform the training of the texture generation network, including: performing human body segmentation on a single image of a human body sample in the training sample image set to obtain a first sample segmentation mask, a second sample segmentation mask and all the front texture of the human sample, wherein the first sample segmentation mask corresponds to the mask area of the front texture of the human sample, and the second sample segmentation mask corresponds to the mask of the back texture of the human sample area; according to the frontal texture of the human body, the third sample segmentation mask and the fourth sample segmentation mask in the auxiliary human body image, the auxiliary texture generation network is trained, wherein the said human body sample is obtained by reducing the resolution of a single image of the human body sample.
  • the auxiliary human body image, the third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human body image, and the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human body image.
  • the foregoing apparatus may be configured to execute any corresponding method described above, which is not repeated here for brevity.
  • An embodiment of the present disclosure further provides an electronic device, where the device includes a memory and a processor, where the memory is used to store computer-readable instructions, and the processor is used to invoke the computer instructions to implement any embodiment of this specification Methods.
  • An embodiment of the present disclosure further provides a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed by a processor, implements the method of any embodiment of the present specification.
  • one or more embodiments of the present disclosure may be provided as a method, a system or a computer program product, the computer program product comprising a computer program that, when executed by a processor, is capable of implementing any of the embodiments of the present specification Methods. Accordingly, one or more embodiments of the present disclosure may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, one or more embodiments of the present disclosure may employ a computer program implemented on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein form of the product.
  • computer-usable storage media including, but not limited to, disk storage, CD-ROM, optical storage, etc.
  • Embodiments of the subject matter and functional operations described in this disclosure can be implemented in digital electronic circuitry, in tangible embodied computer software or firmware, in computer hardware including the structures disclosed in this disclosure and their structural equivalents, or in a combination of one or more.
  • Embodiments of the subject matter described in this disclosure may be implemented as one or more computer programs, ie, one or more of computer program instructions encoded on a tangible, non-transitory program carrier for execution by, or to control the operation of, data processing apparatus. multiple modules.
  • the program instructions may be encoded on an artificially generated propagated signal, such as a machine-generated electrical, optical or electromagnetic signal, which is generated to encode and transmit information to a suitable receiver device for interpretation by the data.
  • the processing device executes.
  • the computer storage medium may be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of one or more of these.
  • the processes and logic flows described in this disclosure can be performed by one or more programmable computers executing one or more computer programs to perform corresponding functions by operating on input data and generating output.
  • the processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, eg, an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
  • Computers suitable for the execution of a computer program include, for example, general and/or special purpose microprocessors, or any other type of central processing unit.
  • the central processing unit will receive instructions and data from read only memory and/or random access memory.
  • the basic components of a computer include a central processing unit for implementing or executing instructions and one or more memory devices for storing instructions and data.
  • a computer will also include, or be operatively coupled to, one or more mass storage devices for storing data, such as magnetic, magneto-optical or optical disks, to receive data therefrom or to It transmits data, or both.
  • the computer does not have to have such a device.
  • the computer may be embedded in another device, such as a mobile phone, personal digital assistant (PDA), mobile audio or video player, game console, global positioning system (GPS) receiver, or a universal serial bus (USB) ) flash drives for portable storage devices, to name a few.
  • PDA personal digital assistant
  • GPS global positioning system
  • USB universal serial bus
  • Computer-readable media suitable for storage of computer program instructions and data include all forms of non-volatile memory, media, and memory devices including, for example, semiconductor memory devices (eg, EPROM, EEPROM, and flash memory devices), magnetic disks (eg, internal hard disks or memory devices). removable disks), magneto-optical disks, and CD-ROM and DVD-ROM disks.
  • semiconductor memory devices eg, EPROM, EEPROM, and flash memory devices
  • magnetic disks eg, internal hard disks or memory devices. removable disks
  • magneto-optical disks e.g., CD-ROM and DVD-ROM disks.
  • the processor and memory may be supplemented by or incorporated in special purpose logic circuitry.

Abstract

Provided in the embodiments of the present disclosure are a three-dimensional human body reconstruction method and apparatus, and a device and a storage medium. The method may comprise: performing human body geometric reconstruction on the basis of a human body image of a target human body to obtain a three-dimensional mesh model of the target human body; on the basis of the human body image, performing local geometric reconstruction on a local part of the target human body to obtain a three-dimensional mesh model of the local part; fusing the three-dimensional mesh model of the local part and the three-dimensional mesh model of the target human body to obtain an initial three-dimensional model; and performing human body texture reconstruction according to the initial three-dimensional model and the human body image, so as to obtain a three-dimensional human body model of the target human body. According to the embodiments of the present disclosure, a local part in a three-dimensional mesh model of a target human body is clear and accurate, thereby improving the reconstruction effect of the local part.

Description

三维人体重建方法、装置、设备及存储介质Three-dimensional human body reconstruction method, device, equipment and storage medium
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本专利申请要求于2021年3月31日提交的、申请号为202110352199.4、发明名称为“三维人体重建方法、装置、设备及存储介质”的中国专利申请的优先权,该申请以引用的方式并入文本中。This patent application claims the priority of the Chinese patent application filed on March 31, 2021, with the application number of 202110352199.4 and the invention titled "Three-dimensional Human Body Reconstruction Method, Apparatus, Equipment and Storage Medium", which is incorporated by reference. into the text.
技术领域technical field
本公开涉及图像处理技术,具体涉及一种三维人体重建方法、装置、设备及存储介质。The present disclosure relates to image processing technology, and in particular, to a three-dimensional human body reconstruction method, device, equipment and storage medium.
背景技术Background technique
三维人体重建是计算机视觉与计算机图形学领域的重要问题。重建出来的人体数字模型在很多领域有着重要应用,如人体测量、虚拟试衣、虚拟主播、游戏角色自定义设计、虚拟现实社交等领域。其中,如何将真实世界中的人体投射到虚拟世界中得到三维人体数字模型是一个重要问题。然而,三维人体的数字化重建是很复杂的,需要扫描者围绕扫描目标进行多角度无死角的连续扫描;并且,重建结果存在着局部重建效果不够精细的问题。3D human body reconstruction is an important problem in the field of computer vision and computer graphics. The reconstructed human digital model has important applications in many fields, such as body measurement, virtual fitting, virtual anchor, game character custom design, virtual reality social networking and other fields. Among them, how to project the human body in the real world into the virtual world to obtain a three-dimensional human body digital model is an important issue. However, the digital reconstruction of the 3D human body is very complicated, requiring the scanner to perform continuous scanning at multiple angles without dead ends around the scanning target; and the reconstruction results have the problem that the local reconstruction effect is not fine enough.
发明内容SUMMARY OF THE INVENTION
有鉴于此,本公开实施例至少提供一种三维人体重建方法、装置、设备及存储介质。In view of this, the embodiments of the present disclosure provide at least a three-dimensional human body reconstruction method, apparatus, device, and storage medium.
第一方面,提供一种三维人体重建方法,所述方法包括:In a first aspect, a three-dimensional human body reconstruction method is provided, the method comprising:
基于目标人体的人体图像进行人体几何重建,得到所述目标人体的三维网格模型;Performing human body geometry reconstruction based on the human body image of the target human body to obtain a three-dimensional mesh model of the target human body;
基于所述目标人体的人体图像,对所述目标人体的局部部位进行局部几何重建,得到所述局部部位的三维网格模型;Based on the human body image of the target human body, perform local geometric reconstruction on the local part of the target human body to obtain a three-dimensional mesh model of the local part;
将所述局部部位的三维网格模型与所述目标人体的三维网格模型进行融合,得到初始三维模型;fusing the three-dimensional mesh model of the local part with the three-dimensional mesh model of the target body to obtain an initial three-dimensional model;
根据所述初始三维模型和所述人体图像,进行所述目标人体的人体纹理的重建,得到所述目标人体的三维人体模型。According to the initial three-dimensional model and the human body image, reconstruction of the human body texture of the target human body is performed to obtain a three-dimensional human body model of the target human body.
在一个例子中,所述基于目标人体的人体图像进行人体几何重建,得到所述目标人体的三维网格模型,包括:通过第一深度神经网络分支对所述目标人体的人体图像进行三维重建,得到第一人体模型;通过第二深度神经网络分支对所述人体图像中的局部图像进行三维重建,得到第二人体模型;其中,所述局部图像包括所述目标人体的局部区域;将所述第一人体模型和所述第二人体模型进行融合,得到融合人体模型;对所述融合人体模型进行网格化处理,得到所述目标人体的三维网格模型。In one example, the performing geometric reconstruction of the human body based on the human body image of the target human body to obtain the three-dimensional mesh model of the target human body includes: performing three-dimensional reconstruction of the human body image of the target human body through a first deep neural network branch, obtaining a first human body model; performing three-dimensional reconstruction on a partial image in the human body image through a second deep neural network branch to obtain a second human body model; wherein, the partial image includes a partial area of the target human body; The first human body model and the second human body model are fused to obtain a fused human body model; the fused human body model is meshed to obtain a three-dimensional mesh model of the target human body.
在一个例子中,所述第一深度神经网络分支包括:全局特征子网络和第一拟合子网络;所述第二深度神经网络分支包括:局部特征子网络和第二拟合子网络;所述通过第一深度神经网络分支对所述目标人体的人体图像进行三维重建,得到第一人体模型,包括:通过所述全局特征子网络对所述人体图像进行特征提取,得到第一图像特征;通过 所述第一拟合子网络基于所述第一图像特征得到所述第一人体模型;所述通过第二深度神经网络分支对所述人体图像中的局部图像进行三维重建,得到第二人体模型,包括:通过所述局部特征子网络对所述局部图像进行特征提取,得到第二图像特征;通过所述第二拟合子网络基于所述第二图像特征以及所述第一拟合子网络输出的中间特征,得到所述第二人体模型。In one example, the first deep neural network branch includes: a global feature sub-network and a first fitting sub-network; the second deep neural network branch includes: a local feature sub-network and a second fitting sub-network; the The three-dimensional reconstruction of the human body image of the target human body through the first deep neural network branch to obtain a first human body model includes: performing feature extraction on the human body image through the global feature sub-network to obtain first image features; The first human body model is obtained based on the first image features through the first fitting sub-network; the second human body is obtained by performing three-dimensional reconstruction of the partial image in the human body image through the second deep neural network branch The model includes: extracting features from the local image through the local feature sub-network to obtain second image features; using the second fitting sub-network based on the second image features and the first fitter The intermediate features output by the network are obtained to obtain the second human body model.
在一个例子中,所述基于所述目标人体的人体图像,对所述目标人体的局部部位进行局部几何重建,得到所述局部部位的三维网格模型,包括:对所述目标人体的人体图像进行特征提取,得到第三图像特征;根据所述第三图像特征、以及所述局部部位的三维拓扑模板,确定所述局部部位的三维网格模型。In one example, performing local geometric reconstruction on a local part of the target human body based on the human body image of the target human body to obtain a three-dimensional mesh model of the local part, including: performing a human body image on the target human body Perform feature extraction to obtain a third image feature; and determine a three-dimensional mesh model of the partial portion according to the third image feature and the three-dimensional topology template of the partial portion.
在一个例子中,所述将所述局部部位的三维网格模型与所述目标人体的三维网格模型进行融合,得到初始三维模型,包括:根据所述目标人体的人体图像,获得所述局部部位的多个关键点;确定所述多个关键点在所述目标人体的三维网格模型上对应的第一模型关键点的信息,以及,确定所述多个关键点在所述局部部位的三维网格模型上对应的第二模型关键点的信息;基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型。In one example, the obtaining the initial 3D model by fusing the 3D mesh model of the local part with the 3D mesh model of the target body includes: obtaining the local body image according to the body image of the target body multiple key points of the part; determine the information of the first model key point corresponding to the multiple key points on the 3D mesh model of the target body, and determine the multiple key points in the local part The information of the corresponding second model key points on the three-dimensional mesh model; based on the information of the first model key points and the information of the second model key points, the three-dimensional mesh model of the local part is fused to the The three-dimensional mesh model of the target human body is obtained to obtain the initial three-dimensional model.
在一个例子中,所述基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型,包括:基于所述第一模型关键点的信息和所述第二模型关键点的信息,确定所述目标人体的三维网格模型与所述局部部位的三维网格模型间的坐标变换关系;根据所述坐标变换关系,将所述局部部位的三维网格模型变换到所述目标人体的三维网格模型的坐标系下;在变换后的坐标系下将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型。In an example, the 3D mesh model of the local part is fused to the 3D mesh model of the target human body based on the information of the key points of the first model and the information of the key points of the second model, Obtaining the initial three-dimensional model includes: determining the three-dimensional mesh model of the target body and the three-dimensional mesh model of the local part based on the information of the key points of the first model and the information of the key points of the second model According to the coordinate transformation relationship, the three-dimensional grid model of the local part is transformed into the coordinate system of the three-dimensional grid model of the target body; in the transformed coordinate system, the local The three-dimensional mesh model of the part is fused to the three-dimensional mesh model of the target body to obtain the initial three-dimensional model.
在一个例子中,所述人体图像包括:所述目标人体的正面纹理和背景图像;所述根据所述初始三维模型和所述人体图像,进行所述目标人体的人体纹理的重建,得到所述目标人体的三维人体模型,包括:对所述人体图像进行人体分割,得到第一分割掩码、第二分割掩码和所述目标人体的正面纹理;其中,所述第一分割掩码对应所述正面纹理的掩码区域,所述第二分割掩码对应于所述目标人体的背面纹理的掩码区域;将所述正面纹理、所述第一分割掩码和所述第二分割掩码,输入纹理生成网络,得到所述目标人体的所述背面纹理;基于所述背面纹理和所述正面纹理,得到所述目标人体对应的带有纹理的三维人体模型。In one example, the human body image includes: a frontal texture and a background image of the target human body; the reconstruction of the human body texture of the target human body is performed according to the initial three-dimensional model and the human body image to obtain the The three-dimensional human body model of the target human body includes: performing human body segmentation on the human body image to obtain a first segmentation mask, a second segmentation mask and a frontal texture of the target human body; wherein, the first segmentation mask corresponds to the The mask area of the front texture, the second segmentation mask corresponds to the mask area of the back texture of the target body; the front texture, the first segmentation mask and the second segmentation mask , input the texture generation network to obtain the back texture of the target body; based on the back texture and the front texture, obtain a textured three-dimensional body model corresponding to the target body.
在一个例子中,所述纹理生成网络的训练,包括如下处理:对训练样本图像集中人体样本的图像进行人体分割,得到第一样本分割掩码、第二样本分割掩码和所述人体样本的正面纹理,其中,所述第一样本分割掩码对应所述人体样本的正面纹理的掩码区域,所述第二样本分割掩码对应所述人体样本的背面纹理的掩码区域;根据辅助人体图像中人体的正面纹理、第三样本分割掩码和第四样本分割掩码,训练辅助纹理生成网络,其中,通过降低所述人体样本的图像的分辨率得到所述辅助人体图像,所述第三样本分割掩码对应所述辅助人体图像中人体得正面纹理的掩码区域,所述第四样本分割掩码对应所述辅助人体图像中人体的背面纹理的掩码区域;在所述辅助纹理生成网络训练完成之 后,基于所述人体样本的正面纹理、所述第一样本分割掩码和所述第二样本分割掩码,训练所述纹理生成网络,其中,所述纹理生成网络的网络参数包括:训练完成的所述辅助纹理生成网络的至少部分网络参数。In one example, the training of the texture generation network includes the following processing: performing human body segmentation on images of human body samples in the training sample image set to obtain a first sample segmentation mask, a second sample segmentation mask and the human body sample , wherein the first sample segmentation mask corresponds to the mask area of the front texture of the human sample, and the second sample segmentation mask corresponds to the mask area of the back texture of the human sample; according to The frontal texture of the human body, the third sample segmentation mask and the fourth sample segmentation mask in the auxiliary human body image, and the auxiliary texture generation network is trained, wherein the auxiliary human body image is obtained by reducing the resolution of the image of the human body sample. The third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human image, and the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human image; After the auxiliary texture generation network training is completed, the texture generation network is trained based on the frontal texture of the human sample, the first sample segmentation mask and the second sample segmentation mask, wherein the texture generation network The network parameters include: at least part of the network parameters of the auxiliary texture generation network that has been trained.
在一个例子中,所述目标人体的局部部位是所述目标人体的人脸;和/或,所述人体图像是RGB图像。In one example, the partial part of the target human body is the face of the target human body; and/or the human body image is an RGB image.
在一个例子中,所述方法还包括:在所述基于目标人体的人体图像进行人体几何重建时,还得到所述目标人体的人体骨骼结构;在所述得到目标人体的三维人体模型之后,基于所述三维人体模型和所述人体骨骼结构,确定用于驱动所述三维人体模型的蒙皮权重。In one example, the method further includes: when performing the human body geometry reconstruction based on the human body image of the target human body, also obtaining the human skeleton structure of the target human body; after the obtaining the three-dimensional human body model of the target human body, based on the The three-dimensional human body model and the human skeleton structure determine skin weights for driving the three-dimensional human body model.
第二方面,提供一种三维人体重建装置,所述装置包括:In a second aspect, a three-dimensional human body reconstruction device is provided, the device comprising:
整体重建模块,用于基于目标人体的人体图像进行人体几何重建,得到所述目标人体的三维网格模型;an overall reconstruction module, used for performing geometric reconstruction of the human body based on the human body image of the target human body to obtain a three-dimensional mesh model of the target human body;
局部重建模块,用于基于所述目标人体的人体图像,对所述目标人体的局部部位进行局部几何重建,得到所述局部部位的三维网格模型;a local reconstruction module, configured to perform local geometric reconstruction on the local part of the target human body based on the human body image of the target human body to obtain a three-dimensional mesh model of the local part;
融合处理模块,用于将所述局部部位的三维网格模型与所述目标人体的三维网格模型进行融合,得到初始三维模型;a fusion processing module, configured to fuse the 3D mesh model of the local part with the 3D mesh model of the target human body to obtain an initial 3D model;
纹理重建模块,用于根据所述初始三维模型和所述人体图像,进行所述目标人体的人体纹理的重建,得到所述目标人体的三维人体模型。The texture reconstruction module is used for reconstructing the human body texture of the target human body according to the initial three-dimensional model and the human body image, so as to obtain the three-dimensional human body model of the target human body.
在一个例子中,所述整体重建模块,在用于得到所述目标人体的三维网格模型时,包括:通过第一深度神经网络分支对所述目标人体的人体图像进行三维重建,得到第一人体模型;通过第二深度神经网络分支对所述人体图像中的局部图像进行三维重建,得到第二人体模型;其中,所述局部图像包括所述目标人体的局部区域;将所述第一人体模型和第二人体模型进行融合,得到融合人体模型;对所述融合人体模型进行网格化处理,得到所述目标人体的三维网格模型。In an example, when the overall reconstruction module is used to obtain the three-dimensional mesh model of the target human body, the method includes: performing three-dimensional reconstruction on the human body image of the target human body through a first deep neural network branch to obtain the first deep neural network branch. a human body model; three-dimensional reconstruction is performed on a partial image in the human body image through a second deep neural network branch to obtain a second human body model; wherein, the partial image includes a partial area of the target human body; the first human body The model and the second human body model are fused to obtain a fused human body model; the fused human body model is meshed to obtain a three-dimensional mesh model of the target human body.
在一个例子中,所述局部重建模块,具体用于:对所述目标人体的人体图像进行特征提取,得到第三图像特征;根据所述第三图像特征、以及所述局部部位的三维拓扑模板,确定所述局部部位的三维网格模型。In one example, the local reconstruction module is specifically configured to: perform feature extraction on the human body image of the target human body to obtain third image features; according to the third image features and the three-dimensional topology template of the local part , and determine the three-dimensional mesh model of the local part.
在一个例子中,所述融合处理模块,具体用于:根据所述目标人体的人体图像,获得所述局部部位的多个关键点;确定所述多个关键点在所述目标人体的三维网格模型上对应的第一模型关键点的信息,以及,确定所述多个关键点在所述局部部位的三维网格模型上对应的第二模型关键点的信息;基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型。In an example, the fusion processing module is specifically configured to: obtain multiple key points of the local part according to the human body image of the target human body; determine that the multiple key points are in a three-dimensional network of the target human body information on the key points of the first model corresponding to the grid model, and determining the information on the key points of the second model corresponding to the plurality of key points on the three-dimensional mesh model of the local part; based on the key points of the first model point information and the information of the key points of the second model, and fuse the three-dimensional mesh model of the local part into the three-dimensional mesh model of the target body to obtain the initial three-dimensional model.
在一个例子中,所述融合处理模块,在用于基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型时,包括:基于所述第一模型关键点的信息和所述第二模型关键点的信息,确定所述目标人体的三维网格模型与所述局部部位的三维网格模型 间的坐标变换关系;根据所述坐标变换关系,将所述局部部位的三维网格模型变换到所述目标人体的三维网格模型的坐标系下;在变换后的坐标系下将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型。In one example, the fusion processing module is configured to fuse the three-dimensional mesh model of the local part to the target based on the information of the key points of the first model and the information of the key points of the second model The three-dimensional mesh model of the human body, when the initial three-dimensional model is obtained, including: based on the information of the key points of the first model and the information of the key points of the second model, determining the three-dimensional mesh model of the target human body and the The coordinate transformation relationship between the three-dimensional mesh models of the local parts; according to the coordinate transformation relationship, the three-dimensional mesh model of the local parts is transformed into the coordinate system of the three-dimensional mesh model of the target body; after the transformation The 3D mesh model of the local part is fused to the 3D mesh model of the target human body under the coordinate system of 1 to obtain the initial 3D model.
在一个例子中,所述纹理重建模块,具体用于:对所述人体图像进行人体分割,得到第一分割掩码、第二分割掩码和所述目标人体的正面纹理;其中,所述第一分割掩码对应所述正面纹理的掩码区域,所述第二分割掩码对应于所述目标人体的背面纹理的掩码区域;将所述正面纹理、所述第一分割掩码和所述第二分割掩码,输入纹理生成网络,得到所述目标人体的所述背面纹理;基于所述背面纹理和所述正面纹理,得到所述目标人体对应的带有纹理的三维人体模型。In one example, the texture reconstruction module is specifically configured to: perform human body segmentation on the human body image to obtain a first segmentation mask, a second segmentation mask and a frontal texture of the target human body; wherein the first segmentation mask A segmentation mask corresponds to the mask area of the front texture, and the second segmentation mask corresponds to the mask area of the back texture of the target human body; the front texture, the first segmentation mask and all the The second segmentation mask is input into a texture generation network to obtain the back texture of the target body; based on the back texture and the front texture, a textured 3D body model corresponding to the target body is obtained.
在一个例子中,所述装置还包括:模型训练模块,用于进行所述纹理生成网络的训练,包括:对训练样本图像集中人体样本的图像进行人体分割,得到第一样本分割掩码、第二样本分割掩码和所述人体样本的正面纹理,其中,所述第一样本分割掩码对应所述人体样本的正面纹理的掩码区域,所述第二样本分割掩码对应所述人体样本的背面纹理的掩码区域;根据辅助人体图像中人体的正面纹理、第三样本分割掩码和第四样本分割掩码,训练辅助纹理生成网络,其中,通过降低所述人体样本的图像的分辨率得到所述辅助人体图像,所述第三样本分割掩码对应所述辅助人体图像中人体的正面纹理的掩码区域,所述第四样本分割掩码对应所述辅助人体图像中所述人体的背面纹理的掩码区域;在所述辅助纹理生成网络训练完成之后,基于所述人体样本的正面纹理、所述第一样本分割掩码和所述第二样本分割掩码,训练所述纹理生成网络,其中,所述纹理生成网络的网络参数包括:训练完成的所述辅助纹理生成网络的至少部分网络参数。In one example, the apparatus further includes: a model training module for training the texture generation network, including: performing human body segmentation on images of human body samples in the training sample image set to obtain a first sample segmentation mask, The second sample segmentation mask and the frontal texture of the human sample, wherein the first sample segmentation mask corresponds to the mask area of the frontal texture of the human sample, and the second sample segmentation mask corresponds to the The mask area of the back texture of the human sample; according to the front texture of the human body in the auxiliary human image, the third sample segmentation mask and the fourth sample segmentation mask, the auxiliary texture generation network is trained, wherein, by reducing the image of the human sample The resolution of the auxiliary human body image is obtained, the third sample segmentation mask corresponds to the mask area of the frontal texture of the human body in the auxiliary human body image, and the fourth sample segmentation mask corresponds to the The mask area of the back texture of the human body; after the auxiliary texture generation network training is completed, based on the front texture of the human sample, the first sample segmentation mask and the second sample segmentation mask, training The texture generation network, wherein the network parameters of the texture generation network include: at least part of the network parameters of the auxiliary texture generation network that has been trained.
第三方面,提供一种电子设备,该设备包括:存储器、处理器,所述存储器用于存储计算机可读指令,所述处理器用于调用所述计算机指令,实现本公开任一实施例所述的方法。In a third aspect, an electronic device is provided, the device comprising: a memory and a processor, where the memory is used for storing computer-readable instructions, and the processor is used for invoking the computer instructions to implement any of the embodiments of the present disclosure. Methods.
第四方面,提供一种计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现本公开任一实施例所述的方法。In a fourth aspect, a computer-readable storage medium is provided, on which a computer program is stored, and when the computer program is executed by a processor, implements the method described in any of the embodiments of the present disclosure.
第五方面,提供一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时实现本公开任一实施例所述的方法。In a fifth aspect, a computer program product is provided, including a computer program, which implements the method described in any embodiment of the present disclosure when the computer program is executed by a processor.
本公开实施例提供的三维人体重建方法、装置、设备及存储介质,通过将目标人体的局部部位进行局部几何重建,并将该局部几何重建得到的局部部位的三维网格模型与目标人体的三维网格模型进行融合,使得目标人体的三维网格模型中的局部部位更加清晰、精细和准确,提高了局部部位的重建效果;并且,该方法可以依据目标人体的单张人体图像进行重建,也简化了用户的配合过程,使得三维人体重建更加简便。The three-dimensional human body reconstruction method, device, device, and storage medium provided by the embodiments of the present disclosure perform local geometric reconstruction on a local part of the target human body, and combine the three-dimensional mesh model of the local part obtained by the local geometric reconstruction with the three-dimensional mesh model of the target human body. The mesh model is fused, so that the local parts in the 3D mesh model of the target body are more clear, fine and accurate, and the reconstruction effect of the local parts is improved; It simplifies the user's cooperation process and makes the three-dimensional human body reconstruction easier.
附图说明Description of drawings
为了更清楚地说明本公开一个或多个实施例或相关技术中的技术方案,下面将对实施例或相关技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本公开一个或多个实施例中记载的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in one or more embodiments of the present disclosure or related technologies, the accompanying drawings required in the description of the embodiments or related technologies will be briefly introduced below. Obviously, the accompanying drawings in the following description The drawings are only some of the embodiments described in one or more embodiments of the present disclosure. For those of ordinary skill in the art, other drawings can also be obtained from these drawings without creative effort.
图1示出了本公开至少一个实施例提供的一种三维人体重建方法的流程图;FIG. 1 shows a flowchart of a three-dimensional human body reconstruction method provided by at least one embodiment of the present disclosure;
图2示出了本公开至少一个实施例提供的基于单张人体图像获取三维网格模型的方式示意图;FIG. 2 shows a schematic diagram of a manner for obtaining a 3D mesh model based on a single human body image provided by at least one embodiment of the present disclosure;
图3示出了本公开至少一个实施例提供的初始三维模型的获取过程示意图;FIG. 3 shows a schematic diagram of an acquisition process of an initial three-dimensional model provided by at least one embodiment of the present disclosure;
图4示出了本公开至少一个实施例提供的纹理重建过程的示意图;FIG. 4 shows a schematic diagram of a texture reconstruction process provided by at least one embodiment of the present disclosure;
图5示出了本公开至少一个实施例提供的蒙皮权重的确定过程示意图;FIG. 5 shows a schematic diagram of a skin weight determination process provided by at least one embodiment of the present disclosure;
图6示出了本公开至少一个实施例提供的基于单张人体图像获取三维网格模型的方式示意图;FIG. 6 shows a schematic diagram of a manner for obtaining a three-dimensional mesh model based on a single human body image provided by at least one embodiment of the present disclosure;
图7示出了本公开至少一个实施例提供的纹理生成的原理示意图;FIG. 7 shows a schematic diagram of the principle of texture generation provided by at least one embodiment of the present disclosure;
图8示出了本公开至少一个实施例提供的纹理生成网络的训练过程示意图;FIG. 8 shows a schematic diagram of a training process of a texture generation network provided by at least one embodiment of the present disclosure;
图9示出了本公开至少一个实施例提供的一种人体图像的示意图;FIG. 9 shows a schematic diagram of a human body image provided by at least one embodiment of the present disclosure;
图10示出了本公开至少一个实施例提供的一种三维人体重建装置的结构图;FIG. 10 shows a structural diagram of a three-dimensional human body reconstruction apparatus provided by at least one embodiment of the present disclosure;
图11示出了本公开至少一个实施例提供的一种三维人体重建装置的结构图。FIG. 11 shows a structural diagram of a three-dimensional human body reconstruction apparatus provided by at least one embodiment of the present disclosure.
具体实施方式Detailed ways
为了使本技术领域的人员更好地理解本公开一个或多个实施例中的技术方案,下面将结合本公开一个或多个实施例中的附图,对本公开一个或多个实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。基于本公开一个或多个实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都应当属于本公开保护的范围。In order to make those skilled in the art better understand the technical solutions in one or more embodiments of the present disclosure, the following will describe the technical solutions in one or more embodiments of the present disclosure with reference to the accompanying drawings in one or more embodiments of the present disclosure. The technical solutions are clearly and completely described, and obviously, the described embodiments are only a part of the embodiments of the present disclosure, but not all of the embodiments. Based on one or more embodiments of the present disclosure, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present disclosure.
三维人体重建在很多领域有着重要应用,包括但不限于如下的应用场景:3D human body reconstruction has important applications in many fields, including but not limited to the following application scenarios:
例如,可以通过三维人体重建,增强一些虚拟现实应用场景的真实感。比如,虚拟试衣、虚拟云会议、虚拟课堂等。For example, the realism of some virtual reality application scenarios can be enhanced through 3D human reconstruction. For example, virtual fitting, virtual cloud meeting, virtual classroom, etc.
又例如,可以将通过三维人体重建得到的三维人体模型,导入到游戏数据里,完成个性化人物角色的生成。For another example, the 3D human body model obtained by 3D human body reconstruction can be imported into the game data to complete the generation of the personalized character.
再例如,目前制作科幻电影需要用到绿幕、动作捕捉等多种科技技术,硬件设备昂贵、整体流程耗时繁杂。通过三维人体重建得到虚拟的三维人体模型,可以简化流程,节省资源。For another example, the production of science fiction movies currently requires the use of various technologies such as green screens and motion capture. The hardware equipment is expensive and the overall process is time-consuming and complicated. Obtaining a virtual three-dimensional human body model through three-dimensional human body reconstruction can simplify the process and save resources.
不论何种应用场景,三维人体重建都存在着如下的需求:一方面,尽可能的简化用户的配合过程,比如,需要用户配合进行多角度的扫描,将使得用户提供较多的配合,对用户来说体验不好。另一方面,尽可能的得到精度更高的三维人体模型,比如,在诸如虚拟云会议、或者AR虚拟交互场景下,对三维人体重建得到的三维人体模型有着更高的真实感和沉浸感的需求。Regardless of the application scenario, 3D human body reconstruction has the following requirements: On the one hand, the user's cooperation process should be simplified as much as possible. Bad experience. On the other hand, try to obtain a 3D human body model with higher accuracy. For example, in scenarios such as virtual cloud conferences or AR virtual interaction scenarios, the 3D human body model obtained from 3D human body reconstruction has a higher sense of realism and immersion. need.
为了解决上述问题,本公开实施例提供了一种三维人体重建方法,该方法旨在基于用户的一张照片来进行该用户的三维人体重建,简化用户的配合流程,并且达到较高精 度的重建效果。In order to solve the above problem, an embodiment of the present disclosure provides a three-dimensional human body reconstruction method, which aims to perform three-dimensional human body reconstruction of the user based on a photo of the user, simplify the user's cooperation process, and achieve high-precision reconstruction Effect.
请参见图1所示,图1示出了本公开至少一个实施例提供的一种三维人体重建方法的流程图。该方法可以包括步骤100至步骤106。Referring to FIG. 1 , FIG. 1 shows a flowchart of a three-dimensional human body reconstruction method provided by at least one embodiment of the present disclosure. The method may include steps 100 to 106 .
在步骤100中,基于目标人体的单张人体图像进行人体几何重建,得到所述目标人体的三维网格模型。In step 100, the geometric reconstruction of the human body is performed based on the single human body image of the target human body to obtain a three-dimensional mesh model of the target human body.
其中,目标人体即三维人体重建的基础用户。例如,对用户小张进行三维人体重建,小张可以称为目标人体,而重建得到的三维人体模型也是以小张的身体为基础得到,与小张的体态、样貌、服装和发型等都具有较高的相似性。Among them, the target human body is the basic user of the 3D human body reconstruction. For example, when 3D human body reconstruction is performed on user Xiao Zhang, Xiao Zhang can be called the target human body, and the reconstructed 3D human body model is also obtained based on Xiao Zhang's body, which is similar to Xiao Zhang's posture, appearance, clothing and hairstyle. have high similarity.
所述单张人体图像是该目标人体的一张人体图像。本公开实施例对该人体图像的采集方式、格式没有特殊要求,在一个示例性的方式中,该单张人体图像可以是目标人体的一张全身人体正面照片。又例如,该单张人体图像可以是RGB彩色图像。这种RGB格式的图像的获得成本较低,比如,在图像采集时不需要使用景深摄像头等成本较高的设备,普通的拍摄设备就可以采集得到。The single human body image is a human body image of the target human body. The embodiment of the present disclosure has no special requirements on the collection method and format of the human body image. In an exemplary manner, the single human body image may be a frontal photograph of the target human body. For another example, the single human body image may be an RGB color image. The acquisition cost of this RGB format image is low. For example, it is not necessary to use high-cost equipment such as a depth-of-field camera during image acquisition, and it can be acquired by ordinary shooting equipment.
本步骤中,可以基于目标人体的单张人体图像进行人体几何重建,得到三维网格模型,该三维网格模型即一个表示人体几何形状的三维网格Mesh,该网格包括若干顶点和面。In this step, the human body geometry can be reconstructed based on the single human body image of the target human body to obtain a three-dimensional mesh model. The three-dimensional mesh model is a three-dimensional mesh Mesh representing the human body geometry, and the mesh includes several vertices and faces.
在一个示例中,本实施例还可以将上述重建得到的三维网格Mesh与预先存储的一个参数化人体模型进行姿态和体型的对齐拟合。具体的,该参数化人体模型包括一个人体表面的mesh以及一组骨骼结构,它们是由一组姿态、体型参数控制的,人体的骨骼位置和表面形状会随着参数值的改变而变化。通过将本步骤100重建得到的三维网格Mesh与该参数化人体模型进行几何对齐后,就得到了本步骤100重建得到的三维网格Mesh对应的骨骼结构。该骨骼结构将用于后续步骤中蒙皮权重的计算。In an example, in this embodiment, the three-dimensional mesh Mesh obtained by the above reconstruction and a pre-stored parameterized human body model can also be aligned and fitted with respect to the posture and body shape. Specifically, the parameterized human body model includes a mesh of the human body surface and a set of skeletal structures, which are controlled by a set of pose and body shape parameters, and the skeleton position and surface shape of the human body will change as the parameter values change. By geometrically aligning the three-dimensional mesh Mesh reconstructed in this step 100 with the parameterized human body model, the bone structure corresponding to the three-dimensional mesh Mesh reconstructed in this step 100 is obtained. This bone structure will be used in the calculation of skinning weights in subsequent steps.
请结合参见图2,示例了一种基于单张人体图像重建获取三维网格模型的方式。如图2所示,可以将目标人体的单张人体图像21输入第一深度神经网络分支22进行三维重建。在一个示例性的实施方式中,该第一深度神经网络分支22可以包括全局特征子网络221和第一拟合子网络222。Please refer to Fig. 2 in combination, which illustrates a method of obtaining a 3D mesh model based on a single human image reconstruction. As shown in FIG. 2 , a single human body image 21 of the target human body can be input into the first deep neural network branch 22 for three-dimensional reconstruction. In an exemplary embodiment, the first deep neural network branch 22 may include a global feature sub-network 221 and a first fitting sub-network 222 .
其中,可以通过全局特征子网络221对单张人体图像21进行特征提取,得到该单张人体图像21的高层图像特征,可以将该高层图像特征称为第一图像特征。例如,该全局特征子网络221可以是一个HourGlass卷积网络。该第一图像特征输入到第一拟合子网络222,该第一拟合子网络222可以依据第一图像特征对三维空间的每一个体素块是否属于目标人体的内部进行预测。例如,该第一拟合子网络222可以是一个多层感知机结构。该第一拟合子网络222输出第一人体模型,该第一人体模型包括位于目标人体内部的各三维体素块。The features of the single human body image 21 can be extracted through the global feature sub-network 221 to obtain high-level image features of the single human body image 21, and the high-level image features can be referred to as first image features. For example, the global feature sub-network 221 may be a HourGlass convolutional network. The first image feature is input to the first fitting sub-network 222, and the first fitting sub-network 222 can predict whether each voxel block in the three-dimensional space belongs to the interior of the target human body according to the first image feature. For example, the first fitting sub-network 222 may be a multilayer perceptron structure. The first fitting sub-network 222 outputs a first human body model including each three-dimensional voxel block located inside the target human body.
接着,可以继续对该第一人体模型进行网格化处理,例如,该网格化处理可以是对该第一人体模型在体素空间应用MarchingCubes算法,得到目标人体的三维网格模型。Next, the meshing process may continue to be performed on the first human body model. For example, the meshing process may be to apply the MarchingCubes algorithm in the voxel space to the first human body model to obtain a three-dimensional mesh model of the target human body.
在步骤102中,基于所述目标人体的单张人体图像,对所述目标人体的局部部位进行局部高清几何重建,得到所述局部部位的三维网格模型。In step 102, based on the single human body image of the target human body, a local high-definition geometric reconstruction is performed on a local part of the target human body to obtain a three-dimensional mesh model of the local part.
步骤100中重建得到的目标人体的三维网格模型,有可能在目标人体的局部部位是模糊的。例如,该局部部位可以是人脸,也可以是其他局部部位,如手部等需体现细节特征的部位。上述三维网格模型在目标人体的脸部细节上较为模糊,而脸部却通常是用户较为关注的区域,因此,本步骤可以对目标人体的局部部位单独进行几何重建。The three-dimensional mesh model of the target human body reconstructed in step 100 may be blurred in local parts of the target human body. For example, the local part may be a human face, or may be other local parts, such as a hand and other parts that need to reflect detailed features. The above-mentioned 3D mesh model is relatively vague in the face details of the target human body, but the face is usually the area that the user pays more attention to. Therefore, in this step, the partial parts of the target human body can be individually geometrically reconstructed.
以所述局部部位是人脸为例:对人体脸部的重建可以采用固定拓扑的精细重建,即可以基于由目标人体的单张人体图像进行特征提取得到的图像特征,对人脸的三维拓扑模板中的各个顶点的位置进行拟合,得到人脸的三维网格模型。具体的,人体脸部的语义结构具有一致性,因此可以采用一个固定拓扑结构的三维人脸作为模板,该模板可以称为人脸的三维拓扑模板。该模板上包括多个顶点,每个顶点固定对应一个脸部语义,例如,一个顶点表示鼻子尖部,另一个顶点表示眼角。在脸部重建时,可以通过一个深度神经网络来回归得到上述人脸的三维拓扑模板的各顶点位置。Taking the local part being a human face as an example: the reconstruction of the human face can use a fine reconstruction of a fixed topology, that is, the three-dimensional topology of the human face can be reconstructed based on the image features obtained by feature extraction from a single human image of the target human body. The position of each vertex in the template is fitted to obtain a three-dimensional mesh model of the face. Specifically, the semantic structure of the human face is consistent, so a 3D human face with a fixed topology structure can be used as a template, and the template can be called a 3D topology template of the human face. The template includes a plurality of vertices, each vertex is fixedly corresponding to a face semantics, for example, one vertex represents the tip of the nose, and the other vertex represents the corner of the eye. During face reconstruction, each vertex position of the above-mentioned three-dimensional topology template of the face can be obtained by regression through a deep neural network.
例如,该深度神经网络可以包括一个深度卷积网络和一个图卷积网络,可以将目标人体的单张人体图像输入所述深度卷积网络提取图像特征,提取得到的特征可以称为第三图像特征。再将该第三图像特征和人脸的三维拓扑模板作为图卷积网络的输入,最终得到图卷积网络输出的一个脸部的三维网格模型,该三维网络模型与目标人体脸部较接近。可选的,深度卷积网络的输入也可以是由目标人体的单张人体图像中截取的包含脸部的部分图像区域。For example, the deep neural network may include a deep convolutional network and a graph convolutional network, and a single human body image of the target human body may be input into the deep convolutional network to extract image features, and the extracted features may be referred to as the third image feature. Then the third image feature and the 3D topology template of the face are used as the input of the graph convolution network, and finally a 3D mesh model of the face output by the graph convolution network is obtained, and the 3D network model is closer to the target human face. . Optionally, the input of the deep convolutional network may also be a partial image area containing a face captured from a single human body image of the target human body.
在步骤104中,将所述局部部位的三维网格模型与所述目标人体的三维网格模型进行融合,得到初始三维模型。In step 104, the 3D mesh model of the local part is fused with the 3D mesh model of the target human body to obtain an initial 3D model.
在步骤100中重建得到的目标人体的三维网格模型在人体的局部部位可能有些模糊,该局部部位以人脸部为例,而步骤102中通过脸部的单独几何重建得到了人脸的三维网格模型,本步骤中,可以将人脸的三维网格模型替换掉步骤100中的目标人体的三维网格模型中的对应部分,这样就可以既保留了目标人体的三维网格模型中的头型、体型、体态等信息,又能使得脸部的五官结构更加的精细和准确,达到更好的重建效果。当然可以理解的是,这里只是以局部部位为人脸为例,实际实施中也可以对其他局部部位进行单独重建以更清晰。The 3D mesh model of the target human body reconstructed in step 100 may be somewhat blurred in the local part of the human body, and the local part is taken as an example of a human face, and in step 102, the 3D mesh model of the human face is obtained through the separate geometric reconstruction of the face Grid model, in this step, the corresponding part in the 3D grid model of the target body in step 100 can be replaced by the 3D grid model of the face, so that the 3D grid model of the target body can be retained. Information such as head shape, body shape, and posture can also make the facial features more refined and accurate, and achieve better reconstruction effects. Of course, it can be understood that the partial part is a human face as an example here, and other partial parts can also be independently reconstructed to make it clearer in actual implementation.
具体的,可以先将目标人体的单张人体图像输入预先训练好的关键点检测模型,通过该关键点检测模型确定图像中目标人体的局部部位的多个关键点。请结合参见图3,仍以局部部位是人脸为例,在获取到人脸的多个关键点31之后,可以根据这些关键点31在人脸的坐标,分别确定关键点在所述目标人体的三维网格模型、以及人脸的三维网格模型上对应的模型关键点。具体可以确定人脸的多个关键点在目标人体的三维网格模型上对应的多个第一模型关键点的信息,比如,该信息可以包括各个第一模型关键点的关键点标识以及对应的关键点位置。还可以确定该人脸的多个关键点在人脸的三维网格模型上对应的第二模型关键点的信息,比如,该信息可以包括各个第二模型关键点的关键点标识和对应的关键点位置。Specifically, a single human body image of the target human body may be input into a pre-trained key point detection model, and a plurality of key points of local parts of the target human body in the image may be determined by the key point detection model. Please refer to FIG. 3 , still taking the local part of the human face as an example, after acquiring multiple key points 31 of the human face, it can be determined according to the coordinates of these key points 31 on the human face that the key points are located in the target human body. The 3D mesh model of the face, and the corresponding model key points on the 3D mesh model of the face. Specifically, the information of multiple first model key points corresponding to multiple key points of the face on the 3D mesh model of the target body can be determined. For example, the information may include the key point identifiers of each first model key point and the corresponding key points. key point location. It is also possible to determine the information of the key points of the second model corresponding to the multiple key points of the face on the three-dimensional mesh model of the face, for example, the information may include the key point identification of each key point of the second model and the corresponding key point. point location.
在获取到上述第一模型关键点的信息和第二模型关键点的信息后,可以基于该第一模型关键点的信息和第二模型关键点的信息,将人脸的三维网格模型融合至目标人体的三维网格模型,得到初始三维模型。After obtaining the above-mentioned information of the key points of the first model and the information of the key points of the second model, the three-dimensional mesh model of the face can be fused to the information of the key points of the first model and the information of the key points of the second model. The 3D mesh model of the target human body is obtained, and the initial 3D model is obtained.
本公开实施例中,将人脸的三维网格模型融合至目标人体的三维网格模型包括:基于第一模型关键点的信息和第二模型关键点的信息,并结合这两个模型的相机外参,确定目标人体的三维网格模型与人脸的三维网格模型之间的坐标变换关系;基于该坐标变换关系,可以将人脸的三维网格模型变换到目标人体的三维网格模型的坐标系下;在变换后的坐标系下,将人脸的三维网格模型融合至目标人体的三维网格模型,例如,可以将目标人体的三维网格模型上的脸部几何结构去掉,使用人脸的三维网格模型补充上,通过泊松重建的方式将人脸的三维网格模型与目标人体的三维网格模型融合为一个整体,得到的模型可以称为初始三维模型。该初始三维模型已经具有了较为清晰的五官结构和相似的头型、体态等信息,精度较高。In the embodiment of the present disclosure, fusing the 3D mesh model of the face into the 3D mesh model of the target body includes: based on the information of the key points of the first model and the information of the key points of the second model, and combining the cameras of the two models The external parameter determines the coordinate transformation relationship between the 3D mesh model of the target body and the 3D mesh model of the face; based on the coordinate transformation relationship, the 3D mesh model of the face can be transformed into the 3D mesh model of the target body In the transformed coordinate system, the 3D mesh model of the face is fused to the 3D mesh model of the target body. For example, the facial geometry on the 3D mesh model of the target body can be removed, In addition to using the 3D mesh model of the face, the 3D mesh model of the face and the 3D mesh model of the target body are integrated into a whole by means of Poisson reconstruction, and the obtained model can be called the initial 3D model. The initial 3D model already has relatively clear facial features, similar head shape, body shape and other information, and the accuracy is high.
在步骤106中,根据所述初始三维模型和所述单张人体图像,进行所述目标人体的人体纹理的重建,得到所述目标人体的带有彩色纹理的三维人体模型。In step 106, reconstruction of the human body texture of the target human body is performed according to the initial three-dimensional model and the single human body image to obtain a three-dimensional human body model with colored textures of the target human body.
由于本实施例是基于目标人体的单张人体图像进行三维人体重建,所以部分的人体区域是不可见的,比如,如果使用目标人体的正面人体图像进行重建,那么该目标人体的背面是不可见的,这将导致纹理缺失的问题。因此,本步骤中可以依据初始三维模型和目标人体的单张人体图像,进行目标人体的不可见区域的人体纹理的预测和补全,并与所述单张人体图像中的人体纹理进行融合,进而生成纹理完整的三维人体模型。Since this embodiment performs three-dimensional human body reconstruction based on a single human body image of the target human body, part of the human body area is invisible. For example, if the frontal human body image of the target human body is used for reconstruction, the back of the target human body is invisible. , which will cause missing textures. Therefore, in this step, the prediction and completion of the human body texture in the invisible area of the target human body can be performed according to the initial three-dimensional model and the single human body image of the target human body, and the human body texture in the single human body image can be fused. Then a textured 3D human body model is generated.
请结合图4所示,以目标人体的单张人体图像是正面图像为例,可以使用深度学习网络进行人体背面纹理41的预测,并结合该人体背面纹理41和单张人体图像中的人体正面纹理42,对初始三维模型进行纹理贴图,即对初始三维模型进行纹理重建。图4中的三维模型43已将上述的人体背面和正面纹理映射在初始三维模型上。在步骤104中得到的初始三维模型是一种人体几何结构的网格Mesh,本步骤是在该网格模型的基础上给模型加上人体纹理。此外,还有一些不可见的人体部位区域,可以使用插值技术在模型的一些缝隙进行纹理的填充,从而将初始三维模型的纹理补全,得到目标人体的三维人体模型44。As shown in Figure 4, taking the single human body image of the target human body as an example of a frontal image, the deep learning network can be used to predict the human body back texture 41, and combine the human body back texture 41 with the front of the human body in the single human body image. Texture 42, performing texture mapping on the initial 3D model, that is, performing texture reconstruction on the initial 3D model. The three-dimensional model 43 in FIG. 4 has mapped the above-mentioned back and front textures of the human body on the initial three-dimensional model. The initial three-dimensional model obtained in step 104 is a mesh Mesh of human body geometry, and this step is to add human body texture to the model on the basis of the mesh model. In addition, there are some invisible human body regions, and the interpolation technology can be used to fill some gaps in the model with textures, so as to complete the texture of the initial 3D model, and obtain the 3D human body model 44 of the target human body.
本实施例的三维人体重建方法,通过将目标人体的局部部位进行局部几何重建,并将该局部几何重建得到的局部部位的三维网格模型与目标人体的三维网格模型进行融合,使得目标人体的初始三维模型中的局部部位更加清晰、精细和准确,提高了局部部位的重建效果;并且,该方法是依据目标人体的单张人体图像进行重建,也简化了用户的配合过程,使得三维人体重建更加简便。The three-dimensional human body reconstruction method of this embodiment performs local geometric reconstruction on a local part of the target human body, and fuses the three-dimensional mesh model of the local part obtained by the local geometric reconstruction with the three-dimensional mesh model of the target human body, so that the target human body can be reconstructed. The local parts in the initial 3D model are more clear, fine and accurate, which improves the reconstruction effect of the local parts; moreover, this method is based on the single human body image of the target human body for reconstruction, which also simplifies the user's cooperation process, so that the three-dimensional human body can be reconstructed. Rebuilding is easier.
此外,在得到人体的三维人体模型后,可以基于所述三维人体模型和目标人体的人体骨骼结构,确定用于驱动所述三维人体模型的蒙皮权重。该蒙皮权重用于驱动建好的三维人体模型,比如,如果要驱动三维人体模型做各种动作,需要将模型绑定到人体骨骼结构上,这种将模型绑定到骨骼上即为蒙皮。继而可以通过骨骼的运动带动模型做动作,而蒙皮权重是用于表示骨骼的节点对模型顶点的影响大小,根据该蒙皮权重可以控制三维人体模型中各个顶点受各个骨骼关节点影响的大小,从而更好的控制模型运动。In addition, after the three-dimensional human body model of the human body is obtained, the skin weight for driving the three-dimensional human body model can be determined based on the three-dimensional human body model and the human skeleton structure of the target human body. The skin weight is used to drive the built 3D human model. For example, if you want to drive the 3D human model to do various actions, you need to bind the model to the human skeleton structure. Binding the model to the bone is called a mask. Skin. Then, the model can be driven by the movement of the bones, and the skin weight is used to represent the influence of the nodes of the bones on the model vertices. According to the skin weight, the size of each vertex in the 3D human model can be controlled by the influence of each bone joint point. , so as to better control the movement of the model.
具体的,计算该三维人体模型的蒙皮权重可以包括如下处理:在步骤100中已经根据目标人体的单张人体图像得到了人体骨骼结构,本步骤可以将该人体骨骼结构和上述得到的三维人体模型输入深度学习网络,通过深度学习网络自动得到模型的蒙皮权重。Specifically, calculating the skin weight of the three-dimensional human body model may include the following processing: in step 100, the human skeleton structure has been obtained according to the single human body image of the target human body, and in this step, the human skeleton structure and the obtained three-dimensional human body can be obtained. The model is input to the deep learning network, and the skin weight of the model is automatically obtained through the deep learning network.
请结合参见图5的示例,可以先根据三维人体模型51和人体骨骼结构52来生成所述三维人体模型51中的各顶点对应的属性特征。该属性特征可以是利用各顶点与人体骨骼结构的空间位置关系来构造得到。例如,对于其中一个顶点来说,该顶点的属性特征可以包括如下四个特征:Referring to the example in FIG. 5 , the attribute features corresponding to the vertices in the three-dimensional human body model 51 may be generated first according to the three-dimensional human body model 51 and the human skeleton structure 52 . The attribute feature can be constructed by using the spatial positional relationship between each vertex and the human skeleton structure. For example, for one of the vertices, the attribute features of the vertex can include the following four features:
1)该顶点的位置坐标;1) The position coordinates of the vertex;
2)离该顶点最近的K个骨骼关节点的位置坐标;2) The position coordinates of the K bone joint points closest to the vertex;
3)由该顶点的位置分别到上述的K个骨骼关节点中各个骨骼关节点之间的测地线距离;3) from the position of the vertex to the geodesic distance between each skeleton joint point in the above-mentioned K skeleton joint points;
4)以上述K个骨骼关节点中的每个骨骼关节点为起点,由该起点指向所述顶点的向量与所述骨骼关节点所在的骨骼之间的夹角;4) Taking each bone joint point in the above-mentioned K bone joint points as a starting point, the angle between the vector of the vertex and the bone where the bone joint point is located by the starting point;
其中,K为正整数。Among them, K is a positive integer.
请继续参见图5,在获得各顶点的属性特征后,可以将该各顶点的属性特征、以及各顶点之间的邻接关系特征作为深度学习网络中的空间图卷积注意力网络的输入。在将这些特征输入空间图卷积注意力网络之前,可以通过一个多层感知机将上述特征转换为隐层特征。空间图卷积注意力网络可以依据上述隐层特征预测每个顶点受上述K个骨骼关节点中的各个骨骼关节点影响的权重,深度学习网络中的后一个多层感知机可以用于将该权重进行归一化处理,使得对于某一个顶点来说,各个骨骼关节点对该顶点的影响权重和为1。最后得到的三维人体模型中各个顶点对应的受各个骨骼关节点影响的权重即为该顶点的蒙皮权重。Please continue to refer to Figure 5. After the attribute features of each vertex are obtained, the attribute features of each vertex and the adjacency relationship feature between the vertices can be used as the input of the spatial graph convolutional attention network in the deep learning network. Before feeding these features into the spatial graph convolutional attention network, the above features can be transformed into hidden layer features through a multilayer perceptron. The spatial graph convolutional attention network can predict the weight of each vertex affected by each of the above K skeletal joint points according to the above hidden layer features, and the latter multi-layer perceptron in the deep learning network can be used for this. The weights are normalized so that for a certain vertex, the sum of the influence weights of each bone joint point on the vertex is 1. The weight corresponding to each vertex in the finally obtained 3D human model and affected by each skeleton joint point is the skin weight of the vertex.
本实施例的三维人体重建方法,能够依据目标人体的单张人体图像得到人体骨骼结构,并依据该人体骨骼结构和重建得到的三维人体模型自动计算得到蒙皮权重,这样既保证了不同输入图像下骨骼的语义结构一致性,又能结合不同的衣物服饰形状快速生成合适的蒙皮权重。其中,骨骼的语义一致性可以方便模型与现成动作库的注册,语义一致的好处是利于生成的模型和骨骼与动作库的套用(注册)。动作库中可以提前存储一些人的动作序列,比如跳舞、拳击之类,动作库存储的是一系列的运动的骨骼。动作库中的这些骨骼的语义和结构都是一致的。如果生成的骨骼具有随机性(关节语义不确定),那么就不利于生成的模型去套用动作库中的动作。因此,本实施例通过保证生成的骨骼的语义结构的一致性,使得更方便动作库的注册。而依具体形状计算生成的蒙皮权重可以使得不同人体模型运动的视觉效果更加自然。The three-dimensional human body reconstruction method of this embodiment can obtain the human skeleton structure according to a single human body image of the target human body, and automatically calculate the skin weight according to the human skeleton structure and the reconstructed three-dimensional human body model, which not only ensures that different input images are The semantic structure of the lower bones is consistent, and appropriate skin weights can be quickly generated in combination with different clothing and apparel shapes. Among them, the semantic consistency of the skeleton can facilitate the registration of the model and the ready-made action library. The advantage of the semantic consistency is that it is conducive to the application (registration) of the generated model and the skeleton and the action library. The action library can store some human action sequences in advance, such as dancing, boxing, etc. The action library stores a series of motion bones. The semantics and structure of these bones in the action library are consistent. If the generated bones are random (the joint semantics are uncertain), it is not conducive to the generated model to apply the actions in the action library. Therefore, this embodiment makes the registration of the action library more convenient by ensuring the consistency of the semantic structure of the generated bones. The skin weights calculated according to the specific shape can make the visual effect of the movement of different human models more natural.
本公开在另一个实施例提供了一种三维人体重建的方法,本实施例的重建流程与图1的实施例相比,区别在于,对步骤100中的基于目标人体的单张人体图像进行人体几何重建的流程进行了改进,以提高重建得到的目标人体的三维网格模型的几何重建精度。其中,本实施例对于图1的实施例相同的处理步骤将不再详述,仅重点描述区别的地方所在。The present disclosure provides a method for three-dimensional human body reconstruction in another embodiment. Compared with the embodiment in FIG. 1 , the reconstruction process of this embodiment is different in that the human body image is performed on the single human body image based on the target human body in step 100 . The process of geometric reconstruction has been improved to improve the geometric reconstruction accuracy of the reconstructed 3D mesh model of the target body. Wherein, in this embodiment, the same processing steps as the embodiment in FIG. 1 will not be described in detail, and only the differences will be mainly described.
如图6所示,在图2所示的网络结构的基础上,增加了第二深度神经网络分支61。该第二深度神经网络分支61可以包括:局部特征子网络611和第二拟合子网络612。可以由目标人体的单张人体图像21中提取出局部区域的图像,得到局部图像62,第二深度神经网络是用于对该局部图像62进行三维重建。As shown in FIG. 6 , on the basis of the network structure shown in FIG. 2 , a second deep neural network branch 61 is added. The second deep neural network branch 61 may include: a local feature sub-network 611 and a second fitting sub-network 612 . An image of a local area can be extracted from the single human body image 21 of the target human body to obtain a local image 62 , and the second deep neural network is used for three-dimensional reconstruction of the local image 62 .
需要说明的是,这里的局部图像中包括的目标人体的人体区域可以与步骤102中局部几何重建对应的局部部位不完全相同,例如,这里的局部图像可以包括目标人体的肩部以上的区域范围,而步骤102中重建的局部部位可以是目标人体的人脸。当然,图6中对目标人体的肩部以上进行重建只是示例,也可以对目标人体的其他人体区域进行精细化几何重建。It should be noted that the body region of the target human body included in the partial image here may not be exactly the same as the partial part corresponding to the local geometric reconstruction in step 102. For example, the partial image here may include the area above the shoulder of the target human body , and the local part reconstructed in step 102 may be the face of the target human body. Of course, the reconstruction above the shoulder of the target human body in FIG. 6 is just an example, and refined geometric reconstruction can also be performed on other human body regions of the target human body.
具体的,请继续参见图6,通过第一深度神经网络分支22重建得到第一人体模型,并将局部图像62输入第二深度神经网络分支61,由局部特征子网络611对所述局部图像进行特征提取,得到第二图像特征。再通过第二拟合子网络612基于所述第二图像特征以及第一拟合子网络222输出的中间特征,得到第二人体模型。其中,所述的中间特征可以是第一拟合子网络222中的部分网络结构输出的特征,示例性的,假设第一拟合子网络222中包括一定数量的全连接层,那么可以将其中部分数量的全连接层的输出作为所述中间特征输入至第二拟合子网络612。Specifically, please continue to refer to FIG. 6, the first human body model is reconstructed through the first deep neural network branch 22, and the partial image 62 is input into the second deep neural network branch 61, and the partial image is processed by the partial feature sub-network 611. Feature extraction to obtain second image features. Then, a second human body model is obtained through the second fitting sub-network 612 based on the second image feature and the intermediate feature output by the first fitting sub-network 222 . Wherein, the intermediate features may be the features output by part of the network structure in the first fitting sub-network 222. Exemplarily, if the first fitting sub-network 222 includes a certain number of fully connected layers, then the The outputs of the partial number of fully connected layers are input to the second fitting sub-network 612 as the intermediate features.
示例性的,第二深度神经网络分支61的结构可以与第一深度神经网络分支22的结构基本相同,例如,第一深度神经网络分支22中的全局特征子网络221中可以包括四个Block,每一个Block中可以包括一定数量的卷积层、池化层等特征提取层,而第二深度神经网络分支61中的局部特征子网络611可以包括一个上述的Block。在得到第一人体模型和第二人体模型之后,接着,可以将第一人体模型和第二人体模型进行融合,得到融合人体模型。并继续对该融合人体模型进行网格化处理,得到目标人体的三维网格模型。Exemplarily, the structure of the second deep neural network branch 61 may be basically the same as that of the first deep neural network branch 22, for example, the global feature sub-network 221 in the first deep neural network branch 22 may include four Blocks, Each block may include a certain number of feature extraction layers such as convolution layers and pooling layers, and the local feature sub-network 611 in the second deep neural network branch 61 may include one of the above-mentioned blocks. After the first human body model and the second human body model are obtained, then the first human body model and the second human body model may be fused to obtain a fused human body model. And continue to mesh the fused human body model to obtain a three-dimensional mesh model of the target human body.
本实施例的三维人体重建方法,不仅通过对目标人体的局部部位进行局部几何重建,提高了局部部位的重建效果,并且依据目标人体的单张人体图像进行重建,简化了用户的配合过程;此外,还通过第二深度神经网络对局部图像进行重建,提高了对目标人体的局部人体区域的重建效果。The three-dimensional human body reconstruction method of this embodiment not only improves the reconstruction effect of local parts by performing local geometric reconstruction on the local parts of the target human body, but also performs reconstruction based on a single human body image of the target human body, which simplifies the cooperation process of users; , and also reconstruct the local image through the second deep neural network, which improves the reconstruction effect of the local human body area of the target human body.
本公开在又一个实施例提供了一种三维人体重建的方法,该又一个实施例的重建流程与图1的实施例相比,提供了一种具体的通过深度学习网络进行人体背面纹理的预测的方式。其中,本实施例对于图1的实施例相同的处理步骤将不再详述,仅重点描述区别的地方所在。The present disclosure provides a three-dimensional human body reconstruction method in yet another embodiment. Compared with the embodiment in FIG. 1 , the reconstruction process of the further embodiment provides a specific method for predicting the back texture of the human body through a deep learning network. The way. Wherein, in this embodiment, the same processing steps as the embodiment in FIG. 1 will not be described in detail, and only the differences will be mainly described.
如图7所示,目标人体的单张人体图像有时会包括背景图像和人体的正面纹理,这种情况下,可以先进行图像分割,将人体的正面纹理分割出来,在基于该正面纹理预测人体的背面纹理。例如,可以将目标人体的正面图像71进行人体分割,得到第一分割掩码72、和分割后的目标人体的正面纹理73。并且,还将该第一分割掩码72水平翻转后得到第二分割掩码74,再将正面纹理73、第一分割掩码72和第二分割掩码74输入纹理生成网络75,最终得到该纹理生成网络75输出的目标人体的背面纹理。As shown in Figure 7, a single human body image of the target human body sometimes includes a background image and a frontal texture of the human body. In this case, image segmentation can be performed first to segment the frontal texture of the human body, and then predict the human body based on the frontal texture. back texture. For example, the frontal image 71 of the target human body may be segmented to obtain a first segmentation mask 72 and the segmented frontal texture 73 of the target human body. In addition, the first segmentation mask 72 is horizontally flipped to obtain a second segmentation mask 74, and then the front texture 73, the first segmentation mask 72 and the second segmentation mask 74 are input into the texture generation network 75, and finally the The texture generation network 75 outputs the back texture of the target body.
此外,图7是以将第一分割掩码72水平翻转得到第二分割掩码74为例,实际实施中并不局限于此,比如,还可以是将目标人体的正面图像输入一个预先训练得到的神经网络后,该神经网络直接输出第一分割掩码和第二分割掩码。当获得了目标人体的正面纹理和背面纹理后,可以将该正面纹理和背面纹理映射至人体的初始三维模型,即可得到目标人体的三维人体模型。In addition, FIG. 7 is an example of obtaining the second segmentation mask 74 by horizontally flipping the first segmentation mask 72. The actual implementation is not limited to this. For example, the frontal image of the target human body can be input into a pre-training After the neural network is created, the neural network directly outputs the first segmentation mask and the second segmentation mask. After the front and back textures of the target human body are obtained, the front and back textures can be mapped to the initial three-dimensional model of the human body, and the three-dimensional human body model of the target human body can be obtained.
其中,上述的纹理生成网络75的训练过程可以包括如下处理:请结合参见图8,可以使用一个辅助纹理生成网络76。其中,该辅助纹理生成网络76可以包括一部分的纹理生成网络75的网络结构,例如,纹理生成网络75可以是在辅助纹理生成网络76的基础上增加了一定数量的卷积层。Wherein, the above-mentioned training process of the texture generation network 75 may include the following processing: please refer to FIG. 8 in combination, an auxiliary texture generation network 76 may be used. The auxiliary texture generation network 76 may include a part of the network structure of the texture generation network 75 . For example, the texture generation network 75 may add a certain number of convolution layers to the auxiliary texture generation network 76 .
在训练时,可以根据训练样本图像集中的辅助人体图像、第三样本分割掩码和第四样本分割掩码,训练辅助纹理生成网络,并在该辅助纹理生成网络训练完成之后,将辅助纹理生成网络的至少部分网络参数作为纹理生成网络的部分初始化网络参数,继续基于人体样本的正面纹理、第一样本分割掩码和第二样本分割掩码,训练所述纹理生成网络。其中,辅助人体图像是人体样本的单张图像降低分辨率得到,第一样本分割掩码对应人体样本的正面纹理的掩码区域,第二样本分割掩码对应人体样本的背面纹理的掩码区域,第三样本分割掩码对应所述辅助人体图像中人体的正面纹理的掩码区域,所述第四样本分割掩码对应所述辅助人体图像中人体的背面纹理的掩码区域。During training, the auxiliary texture generation network can be trained according to the auxiliary human body image, the third sample segmentation mask and the fourth sample segmentation mask in the training sample image set, and after the auxiliary texture generation network is trained, the auxiliary texture generation network can be generated. At least part of the network parameters of the network are initialized as part of the texture generation network parameters, and the texture generation network is continued to be trained based on the frontal texture of the human sample, the first sample segmentation mask and the second sample segmentation mask. The auxiliary human image is obtained by reducing the resolution of a single image of the human sample, the first sample segmentation mask corresponds to the mask area of the front texture of the human sample, and the second sample segmentation mask corresponds to the mask of the back texture of the human sample The third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human image, and the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human image.
请继续结合参见图8所示:可以通过对辅助人体图像81进行图像分割,得到辅助人体图像81中人体的正面纹理82、第三样本分割掩码83和第四样本分割掩码84,输入辅助纹理生成网络76,得到辅助人体图像81中人体的背面纹理的第一预测值;再基于第一预测值和所述辅助人体图像81中人体的背面纹理的第一真实值,调整所述辅助纹理生成网络76的网络参数。经过多次迭代,可以得到训练完成的辅助纹理生成网络76。其中,对辅助纹理生成网络的训练监督,除了基于第一预测值和第一真实值计算的损失Loss,还可以包括基于第一预测值的其他损失,例如,基于辅助人体图像和第一预测值的纹理特征计算的特征损失,等。其中,所述辅助人体图像可以是图7中的人体正面图像71降低分辨率得到,相应的,辅助人体图像81中人体的正面纹理82的分辨率也比图7中的正面纹理73的分辨率低。所述第三样本分割掩码对应所述辅助人体图像中的人体的正面纹理的掩码区域,所述第四样本分割掩码对应辅助人体图像中的人体的背面纹理的掩码区域。Please continue to refer to FIG. 8 in combination: the frontal texture 82 of the human body in the auxiliary human body image 81, the third sample segmentation mask 83 and the fourth sample segmentation mask 84 can be obtained by performing image segmentation on the auxiliary human body image 81, and input auxiliary human body image 81. The texture generation network 76 obtains the first predicted value of the back texture of the human body in the auxiliary human body image 81; and then adjusts the auxiliary texture based on the first predicted value and the first real value of the back texture of the human body in the auxiliary human body image 81 Network parameters for the network 76 are generated. After several iterations, the trained auxiliary texture generation network 76 can be obtained. Among them, the training supervision of the auxiliary texture generation network, in addition to the loss calculated based on the first predicted value and the first real value, may also include other losses based on the first predicted value, for example, based on the auxiliary body image and the first predicted value. Feature loss for texture feature computation, etc. The auxiliary human body image can be obtained by reducing the resolution of the frontal human body image 71 in FIG. 7 . Correspondingly, the resolution of the frontal texture 82 of the human body in the auxiliary human body image 81 is also higher than the resolution of the frontal texture 73 in FIG. 7 . Low. The third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human body image, and the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human body image.
在所述辅助纹理生成网络训练完成之后,可以将辅助纹理生成网络的网络参数作为纹理生成网络的部分网络参数的初始化,即纹理生成网络的网络参数包括:训练完成的所述辅助纹理生成网络的至少部分网络参数。即辅助纹理生成网络和纹理生成网络共享部分网络权重。然后将用于训练纹理生成网络的训练样本图像集中的人体正面纹理、第一样本分割掩码和第二样本分割掩码,输入所述纹理生成网络,得到人体样本的背面纹理的第二预测值。基于所述第二预测值和所述背面纹理的第二真实值,调整纹理生成网络的网络参数。其中,所述第二真实值的分辨率高于第一真实值的分辨率,即纹理生成网络输出的背面纹理的分辨率会比辅助纹理生成网络输出的背面纹理的分辨率高一些。After the training of the auxiliary texture generation network is completed, the network parameters of the auxiliary texture generation network can be used as the initialization of part of the network parameters of the texture generation network, that is, the network parameters of the texture generation network include: at least some of the network parameters. That is, the auxiliary texture generation network and the texture generation network share some network weights. Then, the frontal texture of the human body, the first sample segmentation mask and the second sample segmentation mask in the training sample image set used to train the texture generation network are input into the texture generation network to obtain the second prediction of the back texture of the human body sample value. Based on the second predicted value and the second real value of the back texture, network parameters of the texture generation network are adjusted. The resolution of the second real value is higher than the resolution of the first real value, that is, the resolution of the back texture output by the texture generation network is higher than the resolution of the back texture output by the auxiliary texture generation network.
本实施例的三维人体重建方法,不仅通过对目标人体的局部部位进行局部几何重建,提高了局部部位的重建效果,并且依据目标人体的单张人体图像进行重建,简化了用户的配合过程;此外,还通过神经网络自动进行纹理的预测,使得生成的纹理效果较好,比如人体周身的纹理更均匀,颜色更真实;并且,通过先训练辅助纹理生成网络再训练纹理生成网络的方式,使得纹理生成网络的训练过程更稳定、更容易收敛。The three-dimensional human body reconstruction method of this embodiment not only improves the reconstruction effect of local parts by performing local geometric reconstruction on the local parts of the target human body, but also performs reconstruction based on a single human body image of the target human body, which simplifies the cooperation process of users; It also automatically predicts the texture through the neural network, so that the generated texture effect is better, for example, the texture around the human body is more uniform and the color is more realistic; and, by training the auxiliary texture generation network first and then training the texture generation network, the texture is The training process of the generative network is more stable and easier to converge.
在其他的实施例中,为了提高重建的效果,也可以获取目标人体的多张不同角度的图像来综合进行该目标人体的三维重建。例如,以获取了该目标人体的三张图像为例, 这三张图像可以是从不同角度采集得到。请结合图2来看,可以将这三张图像分别作为全局特征子网络221的输入,得到全局特征子网络221输出的分别对应这三张图像的一个第一图像特征。然后将三个第一图像特征进行融合,将融合后得到的图像特征作为第一拟合子网络222的输入继续处理。In other embodiments, in order to improve the reconstruction effect, a plurality of images of the target body from different angles may also be acquired to comprehensively perform the three-dimensional reconstruction of the target body. For example, taking three images of the target human body as an example, the three images may be acquired from different angles. Referring to FIG. 2 , the three images can be used as the input of the global feature sub-network 221 respectively, and a first image feature output by the global feature sub-network 221 corresponding to the three images can be obtained. Then, the three first image features are fused, and the image features obtained after fusion are used as the input of the first fitting sub-network 222 to continue processing.
当三维人体重建采用图6所示的网络结构时,除了将上述三张图像分别作为全局特征子网络221的输入之外,还可以由该三张图像中提取局部区域得到局部图像,并将三个局部图像分别作为局部特征子网络611的输入,得到局部特征子网络611输出的分别对应这三张局部图像的第二图像特征,然后将三个第二图像特征进行融合,将融合后得到的图像特征作为第二拟合子网络612的输入继续处理。When the three-dimensional human body reconstruction adopts the network structure shown in FIG. 6 , in addition to using the above three images as the input of the global feature sub-network 221 respectively, the local images can also be obtained by extracting local regions from the three images, and the three Each of the local images is used as the input of the local feature sub-network 611, respectively, and the second image features output by the local feature sub-network 611 corresponding to the three local images are obtained, and then the three second image features are fused, and the result obtained after fusion is obtained. The image features continue to be processed as input to the second fitting sub-network 612 .
如上,通过获取目标人体的多张不同角度的图像来综合进行该目标人体的三维人体重建,能够得到该目标人体对应的更精细的三维人体模型。As above, by acquiring a plurality of images of the target human body from different angles to comprehensively perform the three-dimensional human body reconstruction of the target human body, a more refined three-dimensional human body model corresponding to the target human body can be obtained.
此外,还需要说明的是,本公开任一实施例描述的三维人体重建方法的各个流程步骤中,涉及到的神经网络模型,都可以分别进行训练。例如,第一深度神经网络分支和纹理生成网络可以是各自进行自身的训练。In addition, it should also be noted that, in each process step of the three-dimensional human body reconstruction method described in any embodiment of the present disclosure, the neural network models involved can be trained separately. For example, the first deep neural network branch and the texture generation network may each perform their own training.
如下描述一个三维人体重建流程的示例,其中,与前述任一方法实施例中描述的过程相同的处理,在此简单说明,详细过程可以结合参见前述实施例。An example of a three-dimensional human body reconstruction process is described as follows, wherein the process is the same as the process described in any of the foregoing method embodiments, which is briefly described here, and the detailed process may be combined with reference to the foregoing embodiments.
在该例子中,假设要基于用户U1的单张人体图像构建该U1的三维人体模型,所述的单张人体图像可以是用户U1的正面图像,其中包括用户U1的正面纹理以及背景图像。可以参见图9的示意,用户U1的单张人体图像91中包括该用户的正面纹理92和背景图像93。In this example, it is assumed that the three-dimensional human body model of user U1 is to be constructed based on a single human body image of user U1. The single human body image may be a frontal image of user U1, including the frontal texture and background image of user U1. Referring to the illustration in FIG. 9 , the single human body image 91 of the user U1 includes a front texture 92 and a background image 93 of the user.
首先,可以基于用户U1的单张人体图像91分别进行两方面的重建。First, two aspects of reconstruction can be performed based on the single human body image 91 of the user U1.
一个方面的重建是,基于单张人体图像91进行人体几何重建得到U1的三维网格模型和人体骨骼结构。示例的,可以通过图6所示的网络对单张人体图像91进行处理,通过第一深度神经网络分支中的全局特征子网络和第一拟合子网络对单张人体图像91进行处理,得到第一人体模型;通过第二深度神经网络分支中的局部特征子网络和第二拟合子网络对单张人体图像91中人体肩部以上区域的图像进行处理,得到第二人体模型。并将第一人体模型和第二人体模型进行融合后,得到融合人体模型。再将融合人体模型进行网格化处理,得到用户U1的三维网格模型(mesh)。One aspect of reconstruction is to perform geometric reconstruction of the human body based on the single human body image 91 to obtain the three-dimensional mesh model of U1 and the human skeleton structure. Illustratively, the single human body image 91 can be processed through the network shown in FIG. 6 , and the single human body image 91 can be processed through the global feature sub-network and the first fitting sub-network in the first deep neural network branch to obtain: The first human body model; the local feature sub-network and the second fitting sub-network in the second deep neural network branch process the image of the area above the human body shoulder in the single human body image 91 to obtain the second human body model. After the first human body model and the second human body model are fused, a fused human body model is obtained. The fused human body model is then meshed to obtain a three-dimensional mesh model (mesh) of the user U1.
另一个方面的重建是,基于单张人体图像91对用户U1的人脸进行局部几何重建,得到人脸的三维网格模型。具体的,可以对单张人体图像91进行特征提取,将提取得到的图像特征以及人脸三维拓扑模板输入图卷积神经网络,得到该用户U1的人脸mesh。Another aspect of reconstruction is to perform local geometric reconstruction on the face of the user U1 based on the single human body image 91 to obtain a three-dimensional mesh model of the face. Specifically, feature extraction can be performed on a single human body image 91, and the extracted image features and the three-dimensional face topology template are input into a graph convolutional neural network to obtain the face mesh of the user U1.
接着,可以结合上述重建得到的人脸mesh(人脸的三维网格模型)和用户U1的人体mesh(U1人体的三维网格模型),进行两者的融合,得到U1的初始三维模型。Next, the face mesh (three-dimensional mesh model of the human face) obtained by the above reconstruction and the human body mesh of the user U1 (the three-dimensional mesh model of the human body of U1) can be combined to obtain the initial three-dimensional model of U1.
具体的,可以根据图3的示意流程,结合人脸部的关键点,确定关键点分别在人脸mesh和人体mesh上对应的各模型关键点的标识和位置,并基于这些模型关键点的标识和位置、模型的相机外参等参数,确定模型之间的坐标变换关系。基于该坐标变换 关系,将人脸mesh变换到人体mesh的坐标系下,用人脸mesh替换掉人体mesh中的脸部,并通过泊松重建将人脸mesh和人体mesh融合在一起,得到用户U1的初始三维模型。Specifically, according to the schematic process of FIG. 3, combined with the key points of the human face, the identification and position of the key points of each model corresponding to the key points on the face mesh and the human mesh respectively can be determined, and based on the identification of these model key points and position, camera external parameters of the model and other parameters to determine the coordinate transformation relationship between the models. Based on the coordinate transformation relationship, the face mesh is transformed into the coordinate system of the human mesh, the face in the human mesh is replaced with the face mesh, and the face mesh and the human mesh are fused together through Poisson reconstruction to obtain the user U1 the initial 3D model.
然后,基于上述的初始三维模型和用户U1的单张人体图像91,进行U1的人体纹理的重建。其中,单张人体图像91由于是用户U1的正面纹理,可以基于该正面纹理去预测U1的背面纹理。Then, based on the above-mentioned initial three-dimensional model and the single human body image 91 of the user U1, reconstruction of the human body texture of U1 is performed. Among them, since the single human body image 91 is the front texture of the user U1, the back texture of U1 can be predicted based on the front texture.
具体的,可以对单张人体图像91进行人体分割,得到去除了背景图像的人体正面纹理、用于表示人体正面纹理区域的第一分割掩码,并将第一分割掩码翻转后得到用于表示人体背面纹理区域的第二分割掩码。再将该人体正面纹理、第一分割掩码和第二分割掩码输入预先训练好的纹理生成网络,得到用户U1的背面纹理。最后基于该正面纹理和背面纹理对初始三维模型进行纹理贴图,并将模型缝隙区域进行纹理的填充和补全,最终得到带有纹理的U1的三维人体模型。Specifically, the human body can be segmented on the single human body image 91 to obtain the human body frontal texture with the background image removed, and the first segmentation mask used to represent the frontal texture area of the human body. A second segmentation mask representing the textured regions of the back of the human body. Then input the human body front texture, the first segmentation mask and the second segmentation mask into the pre-trained texture generation network to obtain the back texture of the user U1. Finally, texture mapping is performed on the initial 3D model based on the front texture and back texture, and the texture is filled and completed in the gap area of the model, and finally the 3D human model with texture U1 is obtained.
为了方便对建好的三维人体模型进行模型驱动,还可以结合重建得到的U1的三维人体模型以及在重建U1的三维网格模型时得到的人体骨骼结构,计算三维人体模型的蒙皮权重。后续可以通过该蒙皮权重驱动模型执行动作。In order to facilitate the model driving of the built 3D human model, the skin weight of the 3D human model can also be calculated by combining the reconstructed 3D human model of U1 and the human skeleton structure obtained when reconstructing the 3D mesh model of U1. You can then drive the model to perform actions through this skin weight.
图10示例了一种三维人体重建装置的结构示意图,如图10所示,该装置可以包括:整体重建模块1001、局部重建模块1002、融合处理模块1003和纹理重建模块1004。FIG. 10 illustrates a schematic structural diagram of a three-dimensional human body reconstruction apparatus. As shown in FIG. 10 , the apparatus may include: an overall reconstruction module 1001 , a local reconstruction module 1002 , a fusion processing module 1003 and a texture reconstruction module 1004 .
整体重建模块1001,用于基于目标人体的单张人体图像进行人体几何重建,得到所述目标人体的三维网格模型。The overall reconstruction module 1001 is configured to perform geometric reconstruction of the human body based on a single human body image of the target human body to obtain a three-dimensional mesh model of the target human body.
局部重建模块1002,用于基于所述目标人体的单张人体图像,对所述目标人体的局部部位进行局部几何重建,得到所述局部部位的三维网格模型。The local reconstruction module 1002 is configured to perform local geometric reconstruction on the local part of the target human body based on the single human body image of the target human body to obtain a three-dimensional mesh model of the local part.
融合处理模块1003,用于将所述局部部位的三维网格模型与所述目标人体的三维网格模型进行融合,得到初始三维模型。The fusion processing module 1003 is configured to fuse the 3D mesh model of the local part with the 3D mesh model of the target human body to obtain an initial 3D model.
纹理重建模块1004,用于根据所述初始三维模型和所述单张人体图像,进行所述目标人体的人体纹理的重建,得到所述目标人体的三维人体模型。The texture reconstruction module 1004 is configured to reconstruct the human body texture of the target human body according to the initial three-dimensional model and the single human body image, so as to obtain a three-dimensional human body model of the target human body.
在一个例子中,整体重建模块1001,在用于得到所述目标人体的三维网格模型时,包括:通过第一深度神经网络分支对所述目标人体的单张人体图像进行三维重建,得到第一人体模型;通过第二深度神经网络分支对所述单张人体图像中的局部图像进行三维重建,得到第二人体模型;其中,所述局部图像包括所述目标人体的局部区域;将所述第一人体模型和第二人体模型进行融合,得到融合人体模型;对所述融合人体模型进行网格化处理,得到所述目标人体的三维网格模型。In one example, when the overall reconstruction module 1001 is used to obtain the 3D mesh model of the target human body, it includes: performing 3D reconstruction on the single human body image of the target human body through the first deep neural network branch to obtain the first deep neural network branch. a human body model; three-dimensional reconstruction is performed on the partial image in the single human body image through the second deep neural network branch to obtain a second human body model; wherein, the partial image includes a partial area of the target human body; The first human body model and the second human body model are fused to obtain a fused human body model; the fused human body model is meshed to obtain a three-dimensional mesh model of the target human body.
在一个例子中,局部重建模块1002,具体用于:对所述目标人体的单张人体图像进行特征提取,得到第三图像特征;根据所述第三图像特征、以及所述局部部位的三维拓扑模板,确定所述局部部位的三维网格模型。In one example, the local reconstruction module 1002 is specifically configured to: perform feature extraction on a single human body image of the target human body to obtain a third image feature; according to the third image feature and the three-dimensional topology of the local part A template is used to determine the three-dimensional mesh model of the local part.
在一个例子中,融合处理模块1003,具体用于:根据所述目标人体的单张人体图像,获得所述局部部位的多个关键点;确定所述多个关键点在所述目标人体的三维网格模型上对应的第一模型关键点的信息,以及,确定所述多个关键点在所述局部部位的 三维网格模型上对应的第二模型关键点的信息;基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型。In one example, the fusion processing module 1003 is specifically configured to: obtain multiple key points of the local part according to a single human body image of the target human body; information on the key points of the first model corresponding to the grid model, and determining the information on the key points of the second model corresponding to the plurality of key points on the three-dimensional grid model of the local part; based on the first model The information of the key points and the information of the key points of the second model are fused with the three-dimensional mesh model of the local part into the three-dimensional mesh model of the target body to obtain the initial three-dimensional model.
在一个例子中,融合处理模块1003,在用于基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型时,包括:基于所述第一模型关键点的信息和所述第二模型关键点的信息,确定所述目标人体的三维网格模型与所述局部部位的三维网格模型间的坐标变换关系;根据所述坐标变换关系,将所述局部部位的三维网格模型变换到所述目标人体的三维网格模型的坐标系下;在变换后的坐标系下将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型。In one example, the fusion processing module 1003 is configured to fuse the three-dimensional mesh model of the local part to the target human body based on the information of the key points of the first model and the information of the key points of the second model When obtaining the initial three-dimensional model, it includes: based on the information of the key points of the first model and the information of the key points of the second model, determining the three-dimensional mesh model of the target human body and the The coordinate transformation relationship between the three-dimensional mesh models of the local parts; according to the coordinate transformation relationship, the three-dimensional mesh model of the local part is transformed into the coordinate system of the three-dimensional mesh model of the target body; The three-dimensional mesh model of the local part is fused to the three-dimensional mesh model of the target body under the coordinate system to obtain the initial three-dimensional model.
在一个例子中,纹理重建模块1004,具体用于:对所述单张人体图像进行人体分割,得到第一分割掩码、第二分割掩码和目标人体的正面纹理;其中,所述第一分割掩码对应所述正面纹理的掩码区域,所述第二分割掩码对应于目标人体的背面纹理的掩码区域;将所述正面纹理、所述第一分割掩码和第二分割掩码,输入纹理生成网络,得到所述目标人体的所述背面纹理;基于所述背面纹理和正面纹理,得到所述目标人体对应的带有纹理的三维人体模型。In one example, the texture reconstruction module 1004 is specifically configured to: perform human body segmentation on the single human body image to obtain a first segmentation mask, a second segmentation mask and a frontal texture of the target human body; wherein the first segmentation mask The segmentation mask corresponds to the mask area of the front texture, and the second segmentation mask corresponds to the mask area of the back texture of the target human body; the front texture, the first segmentation mask and the second segmentation mask are code, input the texture generation network to obtain the back texture of the target body; based on the back texture and the front texture, obtain a textured three-dimensional body model corresponding to the target body.
在一个例子中,如图11所示,该装置还可以包括:模型训练模块1005。In an example, as shown in FIG. 11 , the apparatus may further include: a model training module 1005 .
模型训练模块1005,用于进行所述纹理生成网络的训练,包括:对训练样本图像集中人体样本的单张图像进行人体分割,得到第一样本分割掩码、第二样本分割掩码和所述人体样本的正面纹理,其中,所述第一样本分割掩码对应所述人体样本的正面纹理的掩码区域,所述第二样本分割掩码对应所述人体样本的背面纹理的掩码区域;根据辅助人体图像中人体的正面纹理、第三样本分割掩码和第四样本分割掩码,训练辅助纹理生成网络,其中,通过降低所述人体样本的单张图像的分辨率得到所述辅助人体图像,所述第三样本分割掩码对应所述辅助人体图像中人体的正面纹理的掩码区域,所述第四样本分割掩码对应所述辅助人体图像中的人体的背面纹理的掩码区域;在所述辅助纹理生成网络训练完成之后,基于所述人体样本的正面纹理、所述第一样本分割掩码和所述第二样本分割掩码,训练所述纹理生成网络,其中,所述纹理生成网络的网络参数包括:训练完成的所述辅助纹理生成网络的至少部分网络参数。The model training module 1005 is used to perform the training of the texture generation network, including: performing human body segmentation on a single image of a human body sample in the training sample image set to obtain a first sample segmentation mask, a second sample segmentation mask and all the front texture of the human sample, wherein the first sample segmentation mask corresponds to the mask area of the front texture of the human sample, and the second sample segmentation mask corresponds to the mask of the back texture of the human sample area; according to the frontal texture of the human body, the third sample segmentation mask and the fourth sample segmentation mask in the auxiliary human body image, the auxiliary texture generation network is trained, wherein the said human body sample is obtained by reducing the resolution of a single image of the human body sample. The auxiliary human body image, the third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human body image, and the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human body image. code area; after the auxiliary texture generation network is trained, the texture generation network is trained based on the frontal texture of the human sample, the first sample segmentation mask and the second sample segmentation mask, wherein , the network parameters of the texture generation network include: at least part of the network parameters of the auxiliary texture generation network that has been trained.
在一些实施例中,上述装置可以用于执行上文所述的对应任意方法,为了简洁,这里不再赘述。In some embodiments, the foregoing apparatus may be configured to execute any corresponding method described above, which is not repeated here for brevity.
本公开实施例还提供了一种电子设备,所述设备包括存储器、处理器,所述存储器用于存储计算机可读指令,所述处理器用于调用所述计算机指令,实现本说明书任一实施例的方法。An embodiment of the present disclosure further provides an electronic device, where the device includes a memory and a processor, where the memory is used to store computer-readable instructions, and the processor is used to invoke the computer instructions to implement any embodiment of this specification Methods.
本公开实施例还提供了一种计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现本说明书任一实施例的方法。An embodiment of the present disclosure further provides a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed by a processor, implements the method of any embodiment of the present specification.
本领域技术人员应明白,本公开一个或多个实施例可提供为方法、系统或计算机程序产品,该计算机程序产品包括计算机程序,该计算机程序被处理器执行时能够实现本说明书任一实施例的方法。因此,本公开一个或多个实施例可采用完全硬件实施例、 完全软件实施例或结合软件和硬件方面的实施例的形式。而且,本公开一个或多个实施例可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。Those skilled in the art will appreciate that one or more embodiments of the present disclosure may be provided as a method, a system or a computer program product, the computer program product comprising a computer program that, when executed by a processor, is capable of implementing any of the embodiments of the present specification Methods. Accordingly, one or more embodiments of the present disclosure may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, one or more embodiments of the present disclosure may employ a computer program implemented on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein form of the product.
其中,本公开实施例所述的“和/或”表示至少具有两者中的其中一个,例如,“A和/或B”包括三种方案:A、B、以及“A和B”。Wherein, "and/or" in the embodiments of the present disclosure means at least one of the two. For example, "A and/or B" includes three schemes: A, B, and "A and B".
本公开中的各个实施例均采用递进的方式描述,各个实施例之间相同相似的部分互相参见即可,每个实施例重点说明的都是与其他实施例的不同之处。尤其,对于数据处理设备实施例而言,由于其基本相似于方法实施例,所以描述的比较简单,相关之处参见方法实施例的部分说明即可。The various embodiments in the present disclosure are described in a progressive manner, and the same and similar parts between the various embodiments may be referred to each other, and each embodiment focuses on the differences from other embodiments. In particular, for the data processing device embodiment, since it is basically similar to the method embodiment, the description is relatively simple, and for related parts, please refer to the partial description of the method embodiment.
上述对本公开特定实施例进行了描述。其它实施例在所附权利要求书的范围内。在一些情况下,在权利要求书中记载的行为或步骤可以按照不同于实施例中的顺序来执行并且仍然可以实现期望的结果。另外,在附图中描绘的过程不一定要求示出的特定顺序或者连续顺序才能实现期望的结果。在某些实施方式中,多任务处理和并行处理也是可以的或者可能是有利的。The foregoing describes specific embodiments of the present disclosure. Other embodiments are within the scope of the appended claims. In some cases, the acts or steps recited in the claims can be performed in an order different from that in the embodiments and still achieve desirable results. Additionally, the processes depicted in the figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some embodiments, multitasking and parallel processing are also possible or may be advantageous.
本公开中描述的主题及功能操作的实施例可以在以下中实现:数字电子电路、有形体现的计算机软件或固件、包括本公开中公开的结构及其结构性等同物的计算机硬件、或者它们中的一个或多个的组合。本公开中描述的主题的实施例可以实现为一个或多个计算机程序,即编码在有形非暂时性程序载体上以被数据处理装置执行或控制数据处理装置的操作的计算机程序指令中的一个或多个模块。可替代地或附加地,程序指令可以被编码在人工生成的传播信号上,例如机器生成的电、光或电磁信号,该信号被生成以将信息编码并传输到合适的接收机装置以由数据处理装置执行。计算机存储介质可以是机器可读存储设备、机器可读存储基板、随机或串行存取存储器设备、或它们中的一个或多个的组合。Embodiments of the subject matter and functional operations described in this disclosure can be implemented in digital electronic circuitry, in tangible embodied computer software or firmware, in computer hardware including the structures disclosed in this disclosure and their structural equivalents, or in a combination of one or more. Embodiments of the subject matter described in this disclosure may be implemented as one or more computer programs, ie, one or more of computer program instructions encoded on a tangible, non-transitory program carrier for execution by, or to control the operation of, data processing apparatus. multiple modules. Alternatively or additionally, the program instructions may be encoded on an artificially generated propagated signal, such as a machine-generated electrical, optical or electromagnetic signal, which is generated to encode and transmit information to a suitable receiver device for interpretation by the data. The processing device executes. The computer storage medium may be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of one or more of these.
本公开中描述的处理及逻辑流程可以由执行一个或多个计算机程序的一个或多个可编程计算机执行,以通过根据输入数据进行操作并生成输出来执行相应的功能。所述处理及逻辑流程还可以由专用逻辑电路—例如FPGA(现场可编程门阵列)或ASIC(专用集成电路)来执行,并且装置也可以实现为专用逻辑电路。The processes and logic flows described in this disclosure can be performed by one or more programmable computers executing one or more computer programs to perform corresponding functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, eg, an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
适合用于执行计算机程序的计算机包括,例如通用和/或专用微处理器,或任何其他类型的中央处理单元。通常,中央处理单元将从只读存储器和/或随机存取存储器接收指令和数据。计算机的基本组件包括用于实施或执行指令的中央处理单元以及用于存储指令和数据的一个或多个存储器设备。通常,计算机还将包括用于存储数据的一个或多个大容量存储设备,例如磁盘、磁光盘或光盘等,或者计算机将可操作地与此大容量存储设备耦接以从其接收数据或向其传送数据,抑或两种情况兼而有之。然而,计算机不是必须具有这样的设备。此外,计算机可以嵌入在另一设备中,例如移动电话、个人数字助理(PDA)、移动音频或视频播放器、游戏操纵台、全球定位系统(GPS)接收机、或例如通用串行总线(USB)闪存驱动器的便携式存储设备,仅举几例。Computers suitable for the execution of a computer program include, for example, general and/or special purpose microprocessors, or any other type of central processing unit. Typically, the central processing unit will receive instructions and data from read only memory and/or random access memory. The basic components of a computer include a central processing unit for implementing or executing instructions and one or more memory devices for storing instructions and data. Typically, a computer will also include, or be operatively coupled to, one or more mass storage devices for storing data, such as magnetic, magneto-optical or optical disks, to receive data therefrom or to It transmits data, or both. However, the computer does not have to have such a device. Additionally, the computer may be embedded in another device, such as a mobile phone, personal digital assistant (PDA), mobile audio or video player, game console, global positioning system (GPS) receiver, or a universal serial bus (USB) ) flash drives for portable storage devices, to name a few.
适合于存储计算机程序指令和数据的计算机可读介质包括所有形式的非易失性存储器、媒介和存储器设备,例如包括半导体存储器设备(例如EPROM、EEPROM和 闪存设备)、磁盘(例如内部硬盘或可移动盘)、磁光盘以及CD ROM和DVD-ROM盘。处理器和存储器可由专用逻辑电路补充或并入专用逻辑电路中。Computer-readable media suitable for storage of computer program instructions and data include all forms of non-volatile memory, media, and memory devices including, for example, semiconductor memory devices (eg, EPROM, EEPROM, and flash memory devices), magnetic disks (eg, internal hard disks or memory devices). removable disks), magneto-optical disks, and CD-ROM and DVD-ROM disks. The processor and memory may be supplemented by or incorporated in special purpose logic circuitry.
虽然本公开包含许多具体实施细节,但是这些不应被解释为限制任何公开的范围或所要求保护的范围,而是主要用于描述特定公开的具体实施例的特征。本公开内在多个实施例中描述的某些特征也可以在单个实施例中被组合实施。另一方面,在单个实施例中描述的各种特征也可以在多个实施例中分开实施或以任何合适的子组合来实施。此外,虽然特征可以如上所述在某些组合中起作用并且甚至最初如此要求保护,但是来自所要求保护的组合中的一个或多个特征在一些情况下可以从该组合中去除,并且所要求保护的组合可以指向子组合或子组合的变型。Although this disclosure contains many specific implementation details, these should not be construed as limiting the scope of any disclosed or claimed, but rather as describing features of particular embodiments of particular disclosure. Certain features that are described in this disclosure in multiple embodiments can also be implemented in combination in a single embodiment. On the other hand, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Furthermore, although features may function as described above in certain combinations and even be originally claimed as such, one or more features from a claimed combination may in some cases be removed from the combination and the claimed A protected combination may point to a subcombination or a variation of a subcombination.
类似地,虽然在附图中以特定顺序描绘了操作,但是这不应被理解为要求这些操作以所示的特定顺序执行或顺次执行、或者要求所有例示的操作被执行,以实现期望的结果。在某些情况下,多任务和并行处理可能是有利的。此外,上述实施例中的各种系统模块和组件的分离不应被理解为在所有实施例中均需要这样的分离,并且应当理解,所描述的程序组件和系统通常可以一起集成在单个软件产品中,或者封装成多个软件产品。Similarly, although operations are depicted in the figures in a particular order, this should not be construed as requiring the operations to be performed in the particular order shown or sequentially, or that all illustrated operations be performed, to achieve the desired result. In some cases, multitasking and parallel processing may be advantageous. Furthermore, the separation of the various system modules and components in the above-described embodiments should not be construed as requiring such separation in all embodiments, and it should be understood that the described program components and systems may generally be integrated together in a single software product , or packaged into multiple software products.
由此,主题的特定实施例已被描述。其他实施例在所附权利要求书的范围以内。在某些情况下,权利要求书中记载的动作可以以不同的顺序执行并且仍实现期望的结果。此外,附图中描绘的处理并非必需所示的特定顺序或顺次顺序,以实现期望的结果。在某些实现中,多任务和并行处理可能是有利的。Thus, specific embodiments of the subject matter have been described. Other embodiments are within the scope of the appended claims. In some cases, the actions recited in the claims can be performed in a different order and still achieve desirable results. Furthermore, the processes depicted in the figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some implementations, multitasking and parallel processing may be advantageous.
以上所述仅为本公开一个或多个实施例的较佳实施例而已,并不用以限制本公开一个或多个实施例,凡在本公开一个或多个实施例的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本公开一个或多个实施例保护的范围之内。The above descriptions are only preferred embodiments of one or more embodiments of the present disclosure, and are not intended to limit one or more embodiments of the present disclosure. All within the spirit and principles of one or more embodiments of the present disclosure, Any modifications, equivalent replacements, improvements, etc. made should be included within the protection scope of one or more embodiments of the present disclosure.

Claims (20)

  1. 一种三维人体重建方法,包括:A three-dimensional human body reconstruction method, comprising:
    基于目标人体的人体图像进行人体几何重建,得到所述目标人体的三维网格模型;Performing human body geometry reconstruction based on the human body image of the target human body to obtain a three-dimensional mesh model of the target human body;
    基于所述目标人体的人体图像,对所述目标人体的局部部位进行局部几何重建,得到所述局部部位的三维网格模型;Based on the human body image of the target human body, perform local geometric reconstruction on the local part of the target human body to obtain a three-dimensional mesh model of the local part;
    将所述局部部位的三维网格模型与所述目标人体的三维网格模型进行融合,得到初始三维模型;fusing the three-dimensional mesh model of the local part with the three-dimensional mesh model of the target body to obtain an initial three-dimensional model;
    根据所述初始三维模型和所述人体图像,进行所述目标人体的人体纹理的重建,得到所述目标人体的三维人体模型。According to the initial three-dimensional model and the human body image, reconstruction of the human body texture of the target human body is performed to obtain a three-dimensional human body model of the target human body.
  2. 根据权利要求1所述的方法,其特征在于,所述基于目标人体的人体图像进行人体几何重建,得到所述目标人体的三维网格模型,包括:The method according to claim 1, wherein the performing geometric reconstruction of the human body based on the human body image of the target human body to obtain the three-dimensional mesh model of the target human body, comprising:
    通过第一深度神经网络分支对所述目标人体的人体图像进行三维重建,得到第一人体模型;Performing three-dimensional reconstruction on the human body image of the target human body through the first deep neural network branch to obtain a first human body model;
    通过第二深度神经网络分支对所述人体图像中的局部图像进行三维重建,得到第二人体模型;其中,所述局部图像包括所述目标人体的局部区域;A second human body model is obtained by performing three-dimensional reconstruction on a partial image in the human body image through a second deep neural network branch; wherein, the partial image includes a partial region of the target human body;
    将所述第一人体模型和所述第二人体模型进行融合,得到融合人体模型;Fusing the first human body model and the second human body model to obtain a fusion human body model;
    对所述融合人体模型进行网格化处理,得到所述目标人体的三维网格模型。Grid processing is performed on the fused human body model to obtain a three-dimensional grid model of the target human body.
  3. 根据权利要求2所述的方法,其特征在于,所述第一深度神经网络分支包括:全局特征子网络和第一拟合子网络;所述第二深度神经网络分支包括:局部特征子网络和第二拟合子网络;The method according to claim 2, wherein the first deep neural network branch comprises: a global feature sub-network and a first fitting sub-network; the second deep neural network branch comprises: a local feature sub-network and the second fitting sub-network;
    所述通过第一深度神经网络分支对所述目标人体的人体图像进行三维重建,得到第一人体模型,包括:通过所述全局特征子网络对所述人体图像进行特征提取,得到第一图像特征;通过所述第一拟合子网络基于所述第一图像特征得到所述第一人体模型;The three-dimensional reconstruction of the human body image of the target human body through the first deep neural network branch to obtain a first human body model includes: extracting features from the human body image through the global feature sub-network to obtain first image features ; Obtain the first human body model based on the first image feature through the first fitting sub-network;
    所述通过第二深度神经网络分支对所述人体图像中的局部图像进行三维重建,得到第二人体模型,包括:通过所述局部特征子网络对所述局部图像进行特征提取,得到第二图像特征;通过所述第二拟合子网络基于所述第二图像特征以及所述第一拟合子网络输出的中间特征,得到所述第二人体模型。The step of performing three-dimensional reconstruction on the partial image in the human body image through the second deep neural network branch to obtain the second human body model includes: extracting the feature of the partial image through the partial feature sub-network to obtain the second image feature; the second human body model is obtained by the second fitting sub-network based on the second image feature and the intermediate feature output by the first fitting sub-network.
  4. 根据权利要求1至3任一所述的方法,其特征在于,所述基于所述目标人体的人体图像,对所述目标人体的局部部位进行局部几何重建,得到所述局部部位的三维网格模型,包括:The method according to any one of claims 1 to 3, wherein, based on the human body image of the target human body, local geometric reconstruction is performed on a local part of the target human body to obtain a three-dimensional mesh of the local part. models, including:
    对所述目标人体的人体图像进行特征提取,得到第三图像特征;performing feature extraction on the human body image of the target human body to obtain a third image feature;
    根据所述第三图像特征、以及所述局部部位的三维拓扑模板,确定所述局部部位的三维网格模型。A three-dimensional mesh model of the partial part is determined according to the third image feature and the three-dimensional topology template of the partial part.
  5. 根据权利要求1至4任一所述的方法,其特征在于,所述将所述局部部位的三维网格模型与所述目标人体的三维网格模型进行融合,得到初始三维模型,包括:The method according to any one of claims 1 to 4, wherein the obtaining an initial 3D model by fusing the 3D mesh model of the local part with the 3D mesh model of the target body includes:
    根据所述目标人体的人体图像,获得所述局部部位的多个关键点;obtaining a plurality of key points of the local part according to the human body image of the target human body;
    确定所述多个关键点在所述目标人体的三维网格模型上对应的第一模型关键点的信息,以及,确定所述多个关键点在所述局部部位的三维网格模型上对应的第二模型关键点的信息;Determine the information of the first model key points corresponding to the multiple key points on the 3D mesh model of the target human body, and determine the corresponding key points of the multiple key points on the 3D mesh model of the local part Information about key points of the second model;
    基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型。Based on the information of the key points of the first model and the information of the key points of the second model, the three-dimensional mesh model of the local part is fused to the three-dimensional mesh model of the target body to obtain the initial three-dimensional model.
  6. 根据权利要求5所述的方法,其特征在于,所述基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型,包括:The method according to claim 5, wherein the three-dimensional mesh model of the local part is fused to the The three-dimensional mesh model of the target human body to obtain the initial three-dimensional model, including:
    基于所述第一模型关键点的信息和所述第二模型关键点的信息,确定所述目标人体的三维网格模型与所述局部部位的三维网格模型间的坐标变换关系;Based on the information of the key points of the first model and the information of the key points of the second model, determine the coordinate transformation relationship between the three-dimensional mesh model of the target human body and the three-dimensional mesh model of the local part;
    根据所述坐标变换关系,将所述局部部位的三维网格模型变换到所述目标人体的三维网格模型的坐标系下;According to the coordinate transformation relationship, transform the three-dimensional mesh model of the local part into the coordinate system of the three-dimensional mesh model of the target human body;
    在变换后的坐标系下将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型。The three-dimensional mesh model of the local part is fused to the three-dimensional mesh model of the target human body under the transformed coordinate system to obtain the initial three-dimensional model.
  7. 根据权利要求1至6任一所述的方法,其特征在于,所述人体图像包括:所述目标人体的正面纹理和背景图像;The method according to any one of claims 1 to 6, wherein the human body image comprises: a frontal texture and a background image of the target human body;
    所述根据所述初始三维模型和所述人体图像,进行所述目标人体的人体纹理的重建,得到所述目标人体的三维人体模型,包括:The reconstruction of the human body texture of the target human body is performed according to the initial three-dimensional model and the human body image to obtain the three-dimensional human body model of the target human body, including:
    对所述人体图像进行人体分割,得到第一分割掩码、第二分割掩码和所述目标人体的正面纹理;其中,所述第一分割掩码对应所述正面纹理的掩码区域,所述第二分割掩码对应于所述目标人体的背面纹理的掩码区域;Perform human body segmentation on the human body image to obtain a first segmentation mask, a second segmentation mask and a frontal texture of the target human body; wherein, the first segmentation mask corresponds to the mask area of the frontal texture, so the The second segmentation mask corresponds to the mask area of the back texture of the target human body;
    将所述正面纹理、所述第一分割掩码和所述第二分割掩码,输入纹理生成网络,得到所述目标人体的所述背面纹理;Inputting the front texture, the first segmentation mask and the second segmentation mask into a texture generation network to obtain the back texture of the target human body;
    基于所述背面纹理和所述正面纹理,得到所述目标人体对应的带有纹理的三维人体模型。Based on the back texture and the front texture, a textured three-dimensional human body model corresponding to the target human body is obtained.
  8. 根据权利要求7所述的方法,其特征在于,所述纹理生成网络的训练,包括如下处理:The method according to claim 7, wherein the training of the texture generation network includes the following processing:
    对训练样本图像集中人体样本的图像进行人体分割,得到第一样本分割掩码、第二样本分割掩码和所述人体样本的正面纹理,其中,所述第一样本分割掩码对应所述人体样本的正面纹理的掩码区域,所述第二样本分割掩码对应所述人体样本的背面纹理的掩码区域;Perform human body segmentation on the images of human body samples in the training sample image set to obtain a first sample segmentation mask, a second sample segmentation mask and the frontal texture of the human body sample, wherein the first sample segmentation mask corresponds to the the mask area of the front texture of the human sample, and the second sample segmentation mask corresponds to the mask area of the back texture of the human sample;
    根据辅助人体图像中人体的正面纹理、第三样本分割掩码和第四样本分割掩码,训练辅助纹理生成网络,其中,通过降低所述人体样本的图像的分辨率得到所述辅助人体图像,所述第三样本分割掩码对应所述辅助人体图像中人体的正面纹理的掩码区域,所述第四样本分割掩码对应所述辅助人体图像中人体的背面纹理的掩码区域;According to the frontal texture of the human body in the auxiliary human body image, the third sample segmentation mask and the fourth sample segmentation mask, an auxiliary texture generation network is trained, wherein the auxiliary human body image is obtained by reducing the resolution of the image of the human body sample, The third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human image, and the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human image;
    在所述辅助纹理生成网络训练完成之后,基于所述人体样本的正面纹理、所述第一 样本分割掩码和所述第二样本分割掩码,训练所述纹理生成网络,其中,所述纹理生成网络的网络参数包括:训练完成的所述辅助纹理生成网络的至少部分网络参数。After the auxiliary texture generation network is trained, the texture generation network is trained based on the frontal texture of the human sample, the first sample segmentation mask and the second sample segmentation mask, wherein the The network parameters of the texture generation network include: at least part of the network parameters of the auxiliary texture generation network that has been trained.
  9. 根据权利要求1至8任一所述的方法,其特征在于,The method according to any one of claims 1 to 8, wherein,
    所述目标人体的局部部位是所述目标人体的人脸;和/或,The partial part of the target human body is the face of the target human body; and/or,
    所述人体图像是RGB图像。The human body image is an RGB image.
  10. 根据权利要求1至9任一所述的方法,其特征在于,所述方法还包括:The method according to any one of claims 1 to 9, wherein the method further comprises:
    在所述基于目标人体的人体图像进行人体几何重建时,得到所述目标人体的人体骨骼结构;When the human body geometry reconstruction is performed based on the human body image of the target human body, the human skeleton structure of the target human body is obtained;
    在所述得到目标人体的三维人体模型之后,基于所述三维人体模型和所述人体骨骼结构,确定用于驱动所述三维人体模型的蒙皮权重。After the three-dimensional human body model of the target human body is obtained, skin weights for driving the three-dimensional human body model are determined based on the three-dimensional human body model and the human skeleton structure.
  11. 一种三维人体重建装置,其特征在于,所述装置包括:A three-dimensional human body reconstruction device, characterized in that the device comprises:
    整体重建模块,用于基于目标人体的人体图像进行人体几何重建,得到所述目标人体的三维网格模型;an overall reconstruction module, used for performing geometric reconstruction of the human body based on the human body image of the target human body to obtain a three-dimensional mesh model of the target human body;
    局部重建模块,用于基于所述目标人体的人体图像,对所述目标人体的局部部位进行局部几何重建,得到所述局部部位的三维网格模型;a local reconstruction module, configured to perform local geometric reconstruction on the local part of the target human body based on the human body image of the target human body to obtain a three-dimensional mesh model of the local part;
    融合处理模块,用于将所述局部部位的三维网格模型与所述目标人体的三维网格模型进行融合,得到初始三维模型;a fusion processing module, configured to fuse the 3D mesh model of the local part with the 3D mesh model of the target human body to obtain an initial 3D model;
    纹理重建模块,用于根据所述初始三维模型和所述人体图像,进行所述目标人体的人体纹理的重建,得到所述目标人体的三维人体模型。The texture reconstruction module is used for reconstructing the human body texture of the target human body according to the initial three-dimensional model and the human body image, so as to obtain the three-dimensional human body model of the target human body.
  12. 根据权利要求11所述的装置,其特征在于,The apparatus of claim 11, wherein:
    所述整体重建模块,在用于得到所述目标人体的三维网格模型时,包括:通过第一深度神经网络分支对所述目标人体的人体图像进行三维重建,得到第一人体模型;通过第二深度神经网络分支对所述人体图像中的局部图像进行三维重建,得到第二人体模型;其中,所述局部图像包括所述目标人体的局部区域;将所述第一人体模型和所述第二人体模型进行融合,得到融合人体模型;对所述融合人体模型进行网格化处理,得到所述目标人体的三维网格模型。When the overall reconstruction module is used to obtain the three-dimensional mesh model of the target human body, it includes: performing three-dimensional reconstruction on the human body image of the target human body through the first deep neural network branch to obtain a first human body model; The second deep neural network branch performs three-dimensional reconstruction on the partial image in the human body image to obtain a second human body model; wherein, the partial image includes a partial area of the target human body; the first human body model and the second human body model are combined. The two human body models are fused to obtain a fused human body model; the fused human body model is meshed to obtain a three-dimensional mesh model of the target human body.
  13. 根据权利要求11或12所述的装置,其特征在于,The device according to claim 11 or 12, characterized in that:
    所述局部重建模块,具体用于:对所述目标人体的人体图像进行特征提取,得到第三图像特征;根据所述第三图像特征、以及所述局部部位的三维拓扑模板,确定所述局部部位的三维网格模型。The local reconstruction module is specifically configured to: perform feature extraction on the human body image of the target human body to obtain a third image feature; determine the local part according to the third image feature and the three-dimensional topology template of the local part 3D mesh model of the part.
  14. 根据权利要求11至13任一所述的装置,其特征在于,The device according to any one of claims 11 to 13, characterized in that:
    所述融合处理模块,具体用于:根据所述目标人体的单张人体图像,获得所述局部部位的多个关键点;确定所述多个关键点在所述目标人体的三维网格模型上对应的第一模型关键点的信息,以及,确定所述多个关键点在所述局部部位的三维网格模型上对应的第二模型关键点的信息;基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初 始三维模型。The fusion processing module is specifically configured to: obtain multiple key points of the local part according to a single human body image of the target human body; determine that the multiple key points are on the three-dimensional mesh model of the target human body information of the corresponding key points of the first model, and, determining the information of the key points of the second model corresponding to the plurality of key points on the three-dimensional mesh model of the local part; based on the information of the key points of the first model and the information of the key points of the second model, and fuse the three-dimensional mesh model of the local part into the three-dimensional mesh model of the target body to obtain the initial three-dimensional model.
  15. 根据权利要求14所述的装置,其特征在于,The apparatus of claim 14, wherein:
    所述融合处理模块,在用于基于所述第一模型关键点的信息和所述第二模型关键点的信息,将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型时,包括:基于所述第一模型关键点的信息和所述第二模型关键点的信息,确定所述目标人体的三维网格模型与所述局部部位的三维网格模型间的坐标变换关系;根据所述坐标变换关系,将所述局部部位的三维网格模型变换到所述目标人体的三维网格模型的坐标系下;在变换后的坐标系下将所述局部部位的三维网格模型融合至所述目标人体的三维网格模型,得到所述初始三维模型。The fusion processing module is used to fuse the three-dimensional mesh model of the local part to the three-dimensional mesh of the target human body based on the information of the key points of the first model and the information of the key points of the second model model, when the initial three-dimensional model is obtained, including: based on the information of the key points of the first model and the information of the key points of the second model, determining the three-dimensional mesh model of the target human body and the three-dimensional grid model of the local part The coordinate transformation relationship between the grid models; according to the coordinate transformation relationship, the three-dimensional grid model of the local part is transformed into the coordinate system of the three-dimensional grid model of the target body; under the transformed coordinate system, the The three-dimensional mesh model of the local part is fused to the three-dimensional mesh model of the target human body to obtain the initial three-dimensional model.
  16. 根据权利要求11至15任一所述的装置,其特征在于,The device according to any one of claims 11 to 15, characterized in that:
    所述纹理重建模块,具体用于:对所述人体图像进行人体分割,得到第一分割掩码、第二分割掩码和所述目标人体的正面纹理;其中,所述第一分割掩码对应所述正面纹理的掩码区域,所述第二分割掩码对应于所述目标人体的背面纹理的掩码区域;将所述正面纹理、所述第一分割掩码和所述第二分割掩码,输入纹理生成网络,得到所述目标人体的所述背面纹理;基于所述背面纹理和所述正面纹理,得到所述目标人体对应的带有纹理的三维人体模型。The texture reconstruction module is specifically configured to: perform human body segmentation on the human body image to obtain a first segmentation mask, a second segmentation mask and the frontal texture of the target human body; wherein, the first segmentation mask corresponds to The mask area of the front texture, the second segmentation mask corresponds to the mask area of the back texture of the target human body; the front texture, the first segmentation mask and the second segmentation mask are code, input the texture generation network to obtain the back texture of the target body; based on the back texture and the front texture, obtain a textured three-dimensional body model corresponding to the target body.
  17. 根据权利要求16所述的装置,其特征在于,所述装置还包括:The apparatus of claim 16, wherein the apparatus further comprises:
    模型训练模块,用于进行所述纹理生成网络的训练,包括:对训练样本图像集中人体样本的图像进行人体分割,得到第一样本分割掩码、第二样本分割掩码和所述人体样本的正面纹理,其中,所述第一样本分割掩码对应所述人体样本的正面纹理的掩码区域,所述第二样本分割掩码对应所述人体样本的背面纹理的掩码区域;根据辅助人体图像中人体的正面纹理、第三分样本割掩码和第四样本分割掩码,训练辅助纹理生成网络,其中,通过降低所述人体样本的图像的分辨率得到所述辅助人体图像,所述第三样本分割掩码对应所述辅助人体图像中人体的正面纹理的掩码区域,所述第四样本分割掩码对应所述辅助人体图像中所述人体的背面纹理的掩码区域;在所述辅助纹理生成网络训练完成之后,基于所述人体样本的正面纹理、所述第一样本分割掩码和所述第二样本分割掩码,训练所述纹理生成网络,其中,所述纹理生成网络的网络参数包括:训练完成的所述辅助纹理生成网络的至少部分网络参数。A model training module for training the texture generation network, including: performing human body segmentation on images of human body samples in the training sample image set to obtain a first sample segmentation mask, a second sample segmentation mask and the human body sample , wherein the first sample segmentation mask corresponds to the mask area of the front texture of the human sample, and the second sample segmentation mask corresponds to the mask area of the back texture of the human sample; according to The frontal texture of the human body, the third sub-sample segmentation mask and the fourth sample segmentation mask in the auxiliary human body image, and the auxiliary texture generation network is trained, wherein the auxiliary human body image is obtained by reducing the resolution of the image of the human body sample, The third sample segmentation mask corresponds to the mask area of the front texture of the human body in the auxiliary human body image, and the fourth sample segmentation mask corresponds to the mask area of the back texture of the human body in the auxiliary human body image; After the auxiliary texture generation network is trained, the texture generation network is trained based on the frontal texture of the human sample, the first sample segmentation mask and the second sample segmentation mask, wherein the The network parameters of the texture generation network include: at least part of the network parameters of the auxiliary texture generation network that has been trained.
  18. 一种电子设备,包括:存储器、处理器,所述存储器用于存储计算机可读指令,所述处理器用于调用所述计算机指令,实现权利要求1至10任一所述的方法。An electronic device comprising: a memory and a processor, where the memory is used to store computer-readable instructions, and the processor is used to invoke the computer instructions to implement the method according to any one of claims 1 to 10.
  19. 一种计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现权利要求1至10任一所述的方法。A computer-readable storage medium on which a computer program is stored, and when the computer program is executed by a processor, implements the method of any one of claims 1 to 10.
  20. 一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时实现权利要求1至10任一所述的方法。A computer program product comprising a computer program which, when executed by a processor, implements the method of any one of claims 1 to 10.
PCT/CN2021/115122 2021-03-31 2021-08-27 Three-dimensional human body reconstruction method and apparatus, and device and storage medium WO2022205760A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110352199.4A CN113012282B (en) 2021-03-31 2021-03-31 Three-dimensional human body reconstruction method, device, equipment and storage medium
CN202110352199.4 2021-03-31

Publications (1)

Publication Number Publication Date
WO2022205760A1 true WO2022205760A1 (en) 2022-10-06

Family

ID=76387638

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/115122 WO2022205760A1 (en) 2021-03-31 2021-08-27 Three-dimensional human body reconstruction method and apparatus, and device and storage medium

Country Status (2)

Country Link
CN (1) CN113012282B (en)
WO (1) WO2022205760A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116645468A (en) * 2023-05-05 2023-08-25 阿里巴巴达摩院(杭州)科技有限公司 Human body three-dimensional modeling method, method and device for training human body structure to generate model
CN117726907A (en) * 2024-02-06 2024-03-19 之江实验室 Training method of modeling model, three-dimensional human modeling method and device
CN117726907B (en) * 2024-02-06 2024-04-30 之江实验室 Training method of modeling model, three-dimensional human modeling method and device

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113012282B (en) * 2021-03-31 2023-05-19 深圳市慧鲤科技有限公司 Three-dimensional human body reconstruction method, device, equipment and storage medium
CN113658309B (en) * 2021-08-25 2023-08-01 北京百度网讯科技有限公司 Three-dimensional reconstruction method, device, equipment and storage medium
WO2023024036A1 (en) * 2021-08-26 2023-03-02 华为技术有限公司 Method and apparatus for reconstructing three-dimensional model of person
CN113989434A (en) * 2021-10-27 2022-01-28 聚好看科技股份有限公司 Human body three-dimensional reconstruction method and device
CN115458128B (en) * 2022-11-10 2023-03-24 北方健康医疗大数据科技有限公司 Method, device and equipment for generating digital human body image based on key points
CN115810095B (en) * 2022-12-21 2023-08-29 首都师范大学 Surface reconstruction method, device, equipment and medium based on semicircular canal automatic positioning
CN116704097B (en) * 2023-06-07 2024-03-26 好易购家庭购物有限公司 Digitized human figure design method based on human body posture consistency and texture mapping

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110084884A (en) * 2019-04-28 2019-08-02 叠境数字科技(上海)有限公司 A kind of manikin facial area method for reconstructing
CN110378838A (en) * 2019-06-25 2019-10-25 深圳前海达闼云端智能科技有限公司 Become multi-view image generation method, device, storage medium and electronic equipment
US10621779B1 (en) * 2017-05-25 2020-04-14 Fastvdo Llc Artificial intelligence based generation and analysis of 3D models
CN111508079A (en) * 2020-04-22 2020-08-07 深圳追一科技有限公司 Virtual clothing fitting method and device, terminal equipment and storage medium
CN112950739A (en) * 2021-03-31 2021-06-11 深圳市慧鲤科技有限公司 Texture generation method, device, equipment and storage medium
CN112950769A (en) * 2021-03-31 2021-06-11 深圳市慧鲤科技有限公司 Three-dimensional human body reconstruction method, device, equipment and storage medium
CN113012282A (en) * 2021-03-31 2021-06-22 深圳市慧鲤科技有限公司 Three-dimensional human body reconstruction method, device, equipment and storage medium

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10776981B1 (en) * 2019-06-07 2020-09-15 Snap Inc. Entertaining mobile application for animating a single image of a human body and applying effects
CN109887077B (en) * 2019-03-07 2022-06-03 百度在线网络技术(北京)有限公司 Method and apparatus for generating three-dimensional model
CN109978930B (en) * 2019-03-27 2020-11-10 杭州相芯科技有限公司 Stylized human face three-dimensional model automatic generation method based on single image
CN110136243B (en) * 2019-04-09 2023-03-17 五邑大学 Three-dimensional face reconstruction method, system, device and storage medium thereof

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10621779B1 (en) * 2017-05-25 2020-04-14 Fastvdo Llc Artificial intelligence based generation and analysis of 3D models
CN110084884A (en) * 2019-04-28 2019-08-02 叠境数字科技(上海)有限公司 A kind of manikin facial area method for reconstructing
CN110378838A (en) * 2019-06-25 2019-10-25 深圳前海达闼云端智能科技有限公司 Become multi-view image generation method, device, storage medium and electronic equipment
CN111508079A (en) * 2020-04-22 2020-08-07 深圳追一科技有限公司 Virtual clothing fitting method and device, terminal equipment and storage medium
CN112950739A (en) * 2021-03-31 2021-06-11 深圳市慧鲤科技有限公司 Texture generation method, device, equipment and storage medium
CN112950769A (en) * 2021-03-31 2021-06-11 深圳市慧鲤科技有限公司 Three-dimensional human body reconstruction method, device, equipment and storage medium
CN113012282A (en) * 2021-03-31 2021-06-22 深圳市慧鲤科技有限公司 Three-dimensional human body reconstruction method, device, equipment and storage medium

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116645468A (en) * 2023-05-05 2023-08-25 阿里巴巴达摩院(杭州)科技有限公司 Human body three-dimensional modeling method, method and device for training human body structure to generate model
CN116645468B (en) * 2023-05-05 2024-03-26 阿里巴巴达摩院(杭州)科技有限公司 Human body three-dimensional modeling method, method and device for training human body structure to generate model
CN117726907A (en) * 2024-02-06 2024-03-19 之江实验室 Training method of modeling model, three-dimensional human modeling method and device
CN117726907B (en) * 2024-02-06 2024-04-30 之江实验室 Training method of modeling model, three-dimensional human modeling method and device

Also Published As

Publication number Publication date
CN113012282A (en) 2021-06-22
CN113012282B (en) 2023-05-19

Similar Documents

Publication Publication Date Title
WO2022205760A1 (en) Three-dimensional human body reconstruction method and apparatus, and device and storage medium
JP7249390B2 (en) Method and system for real-time 3D capture and live feedback using a monocular camera
WO2022205762A1 (en) Three-dimensional human body reconstruction method and apparatus, device, and storage medium
US10529137B1 (en) Machine learning systems and methods for augmenting images
CN109636831B (en) Method for estimating three-dimensional human body posture and hand information
CN109003325B (en) Three-dimensional reconstruction method, medium, device and computing equipment
US20200234482A1 (en) Systems and methods for photorealistic real-time portrait animation
WO2021184933A1 (en) Three-dimensional human body model reconstruction method
KR20200102114A (en) Method and appartus for learning-based generating 3d model
KR20220017900A (en) Single image-based real-time body animation
CN112037310A (en) Game character action recognition generation method based on neural network
US11928778B2 (en) Method for human body model reconstruction and reconstruction system
WO2022205755A1 (en) Texture generation method and apparatus, device, and storage medium
CN116977522A (en) Rendering method and device of three-dimensional model, computer equipment and storage medium
CN110458924A (en) A kind of three-dimensional facial model method for building up, device and electronic equipment
CN113705295A (en) Object posture migration method, device, equipment and storage medium
CN115018959A (en) Drive processing method, device, equipment and storage medium for three-dimensional virtual model
CN110415322A (en) The generation method and device of the action command of virtual objects model
CN114219001A (en) Model fusion method and related device
CN111739134B (en) Model processing method and device for virtual character and readable storage medium
CN115496864B (en) Model construction method, model reconstruction device, electronic equipment and storage medium
CN115272608A (en) Human hand reconstruction method and equipment
CN115222917A (en) Training method, device and equipment for three-dimensional reconstruction model and storage medium
KR20220149717A (en) Full skeletal 3D pose recovery from monocular camera
US20230274502A1 (en) Methods and systems for 3d modeling of a human subject having hair based on 2d imagery

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21934399

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2023574335

Country of ref document: JP

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 18.01.2024)