CN113554736A - Skeleton animation vertex correction method and model learning method, device and equipment - Google Patents

Skeleton animation vertex correction method and model learning method, device and equipment Download PDF

Info

Publication number
CN113554736A
CN113554736A CN202111103180.2A CN202111103180A CN113554736A CN 113554736 A CN113554736 A CN 113554736A CN 202111103180 A CN202111103180 A CN 202111103180A CN 113554736 A CN113554736 A CN 113554736A
Authority
CN
China
Prior art keywords
model
vertex
posture
joint
standard
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111103180.2A
Other languages
Chinese (zh)
Other versions
CN113554736B (en
Inventor
杨涛
唐浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Tishi Technology Co ltd
Original Assignee
Chengdu Tishi Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Tishi Technology Co ltd filed Critical Chengdu Tishi Technology Co ltd
Priority to CN202111103180.2A priority Critical patent/CN113554736B/en
Publication of CN113554736A publication Critical patent/CN113554736A/en
Application granted granted Critical
Publication of CN113554736B publication Critical patent/CN113554736B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • G06T13/403D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The application provides a skeleton animation vertex correction method and a model learning method, device and equipment, which comprises the steps of acquiring whole-body images and posture data of various actions including T-shaped postures of different people to obtain model training data: pose data P and P of a standard 3D mesh modelmThe method comprises the following steps of establishing and training a joint position prediction model, establishing a posture regression network model and a skeleton animation vertex correction model comprising a standard skin function, the trained joint position prediction model and the trained posture regression network model, and training, wherein the N vertex coordinates Gr and the N vertex coordinates T, N joint point coordinates J of the T-shaped posture 3D grid model are used for establishing and training a joint position prediction model. The skeleton animation vertex correction model provided by the application can compensate and correct unreasonable deformation generated on the original skin surface of the character in the motion process, so that the character angleThe color looks vivid and natural, the animation quality is improved, the workload of an animator is reduced, and the time for making the animation with the same quality is shorter.

Description

Skeleton animation vertex correction method and model learning method, device and equipment
Technical Field
The application relates to the technical field of character skinning, in particular to a learning method of a skeleton animation vertex correction model, a learning device of the skeleton animation vertex correction model, a skeleton animation vertex correction method and computer equipment.
Background
At present, AR intelligent technology is rapidly developing, content intake requirements of Internet users are more abundant, requirements for animation characters, AR and other contents are more intense, and in skeleton animation, after character roles complete skeleton binding, mapping brush weights from the character skeleton to grid vertexes are needed, so that reasonable deformation is formed on the surfaces of the animation characters in the motion process, and the character role covering is achieved. The traditional skeleton animation skin is completely made by manpower, different parts of different characters need to be manually endowed with different weights, and the process is called a weight brushing process.
In the process of covering the character, for different character characters and different parts, an animator needs to give different weights by repeatedly observing the surface deformation in the character motion process, the workload is very complicated, and meanwhile, the final output animation is difficult to be ensured to be consistent with the expected performance, such as the problems of collapse, skin surface shriveling at the joint in a linear hybrid covering model and the like. The animation engineer is caused to repeatedly reset the weight of the role skinning, so that the quality and the output efficiency of the animation are greatly reduced, and the cost of the animation is high. The existing method of IK constraint solves the problem of skin collapse and solves the weight through human tissue dynamics, but the methods have the problems of large workload, long calculation time, inaccurate result, dependence on the experience of an animator on the final effect and the like, and cannot be widely applied.
Disclosure of Invention
The application aims to provide a learning method of a skeleton animation vertex correction model, a learning device of the skeleton animation vertex correction model, a skeleton animation vertex correction method and computer equipment, which can compensate and correct unreasonable deformation generated on the original skin surface of a character in the motion process, and avoid the problems of skin surface collapse, distortion and the like of the character in certain postures in the motion process. The specific scheme is as follows:
the application provides a learning method of a skeleton animation vertex correction model, which comprises the following steps:
s11: collecting whole body images and posture data of different people including various actions of T-shaped postures, generating a standard 3D grid model with N vertexes and a T-shaped posture 3D grid model, and acquiring a posture data axial angle form P of the standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmN vertex coordinates Gr of the standard 3D grid model and N vertex coordinates T of the T-shaped posture 3D grid model, wherein N is a positive integer;
s12: joint skeleton binding is carried out on the T-shaped posture 3D grid model to obtain n joint point coordinates J of the T-shaped posture 3D grid model, wherein n is a positive integer;
s13: constructing a Joint position prediction model Joint _ net, wherein the output dimensionality of the Joint position prediction model is the same as the dimensionality of N Joint point coordinates of the T-shaped posture 3D model, taking N vertex coordinates T of the T-shaped posture 3D grid model as the input of the Joint position prediction model, and taking N Joint point coordinates J of the T-shaped posture 3D grid model as labels of a training Joint position prediction model to train the Joint position prediction model, so as to obtain the trained Joint position prediction model;
s14: constructing a posture regression network model Pose _ net, wherein the output dimensionality of the posture regression network model is the same as the dimensionality of N vertex coordinates of the standard 3D grid model, and the input of the posture regression network model is a posture data rotation matrix form P of the standard 3D grid modelmConstructing a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ net, and rotating a posture data axial angle form P of a standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmAnd taking N vertex coordinates T of the T-shaped posture 3D mesh model as input of the skeleton animation vertex correction model, and taking N vertex coordinates Gr of the standard 3D mesh model as labels for training the skeleton animation vertex correction model to train the skeleton animation vertex correction model, so as to obtain the trained skeleton animation vertex correction model.
Optionally, the S11 includes:
s111: acquiring whole-body images of different people including various actions of T-shaped postures through a camera matrix, and synchronously acquiring posture data through a movable capturing device;
s112: establishing a 3D point cloud model by a stereoscopic vision matching method according to the whole body images of different people including various actions of T-shaped postures, and registering and aligning the obtained 3D point cloud model by combining a geometric constraint and image projection constraint method to generate a standard 3D grid model and a T-shaped posture 3D grid model with a fixed vertex number N;
s113: establishing a coordinate system by taking human pelvis points as a coordinate origin, performing data normalization on the standard 3D grid model to obtain N vertex coordinates Gr of the standard 3D grid model, and performing data normalization on the T-shaped posture 3D grid model to obtain N vertex coordinates T of the T-shaped posture 3D grid model;
s114: establishing a one-to-one matching relation between the acquired attitude data and the standard 3D grid model to obtain an attitude data axial angle form P of the standard 3D grid model, and converting the attitude data axial angle form P of the standard 3D grid model into a rotation matrix form Pm
Optionally, the S13 includes:
s131: constructing a Joint position prediction model Joint _ net, wherein the input data dimension of the Joint position prediction model is N x 3, and the output data dimension is N x 3;
s132: taking a tensor of converting N vertex coordinates T of the T-shaped posture 3D grid model into N x 3 as input of a joint position prediction model, and taking a tensor of converting N joint point coordinates J of the T-shaped posture 3D grid model into N x 3 as a label of a training joint position prediction model to train the joint position prediction model;
s133: optimizing a joint position prediction model by adopting a mean square loss function, wherein the output of the joint position prediction model is JpreThe mean square loss calculation formula is:
Figure 994301DEST_PATH_IMAGE001
where loss represents the mean square loss, J [ i ]]Vector formed by the real coordinates of the ith joint point, Jpre[i]A vector formed by coordinates of the ith Joint point which represents the prediction output of the Joint position prediction model is obtained, and finally the trained Joint position prediction model J = Joint _ n is obtainedet(T)。
Optionally, the S14 includes:
s141: constructing a posture regression network model Pose _ net, wherein the input data dimension of the posture regression network model is N x 3, and the output data dimension is N x 3;
s142: constructing a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ net
Figure 279789DEST_PATH_IMAGE002
Where V represents the set of vertex coordinates of the skeletal animation vertex correction model prediction output, and V = [ V ]0, v1…, vN-1],viCoordinate vector representing ith vertex, LBS linear skin function, T vertex coordinate of T-pose 3D mesh model, P, PmRepresenting the attitude data axis angle form and the rotation matrix form, Pose _ net (P), respectively, of a standard 3D mesh modelm) Representing for each pose PmPredicting a vertex correction value, wherein Joint _ net (T) represents that n Joint point coordinates under the human T-shaped posture are predicted for each T-shaped posture, W represents the weight of a linear skin function, P represents the posture data of the T-shaped posture, W is set as a conductible parameter, and Joint _ net is set as an un-conductible parameter;
s143: converting N vertex coordinates T of T-shaped posture 3D grid model into tensor TinConverting the attitude data P of the standard 3D grid model into tensor Pin、PmInto a tensor PminThe method comprises the steps of training a skeleton animation vertex correction model by taking N vertex coordinates Gr of a standard 3D mesh model as input of the skeleton animation vertex correction model and taking the N vertex coordinates Gr as labels for training the skeleton animation vertex correction model;
s144: optimizing the skeleton animation vertex correction model by adopting a loss function comprising the loss of calculating the vertex coordinate difference of the standard 3D mesh model and the loss of the weight W of the skeleton animation vertex correction model, wherein the mean square loss function for calculating the loss of the vertex coordinate difference of the standard 3D mesh model is as follows:
Figure 282380DEST_PATH_IMAGE003
where loss _ v represents the loss value of the difference in the calculated vertex coordinates, G [ i]Coordinate vector representing the ith vertex of the standard 3D model, V [ i ]]A coordinate vector representing the ith vertex predicted by the skeleton animation vertex correction model; a loss function for calculating the loss of the weight W of the skeleton animation vertex correction model is
Figure 602503DEST_PATH_IMAGE004
Where loss _ w represents the skinning weight loss value, wi,jAnd representing the action weight of the j joint in the skinning function to the i vertex, wherein the combined loss function is as follows:
Figure 563506DEST_PATH_IMAGE005
and the loss represents the total loss value of the skeleton animation vertex correction model, represents the weight occupied by the loss of the vertex coordinate difference of the 3D mesh model, and finally obtains the trained skeleton animation vertex correction model.
The application also provides a bone animation vertex correction method, which comprises the following steps:
acquiring posture data and a T-shaped posture 3D grid model of a human body;
inputting the posture data of the human body and the T-shaped posture 3D mesh model into the trained skeleton animation vertex correction model to obtain N vertex coordinates of the standard 3D mesh model of the human body in the posture;
the skeleton animation vertex correction model is trained by adopting the learning method of the skeleton animation vertex correction model.
The present application further provides a learning device for a skeleton animation vertex correction model, including:
a training data obtaining module 31, configured to collect whole body images and posture data of various actions including T-shaped postures of different people, generate a standard 3D mesh model and a T-shaped posture 3D mesh model having N vertices, and obtain a posture data axis angle form P of the standard 3D mesh model and a posture data rotation matrix form P of the standard 3D mesh modelmN vertex coordinates Gr of the standard 3D grid model and N vertex coordinates T of the T-shaped posture 3D grid model, wherein N is a positive integer;
the joint data acquisition module 32 is used for binding joint bones of the T-shaped posture 3D grid model to obtain n joint point coordinates J of the T-shaped posture 3D grid model, wherein n is a positive integer;
the Joint model obtaining module 33 is configured to construct a Joint position prediction model Joint _ net, an output dimension of the Joint position prediction model is the same as a dimension of N Joint point coordinates of the T-type posture 3D model, N vertex coordinates T of the T-type posture 3D mesh model are used as an input of the Joint position prediction model, and N Joint point coordinates J of the T-type posture 3D mesh model are used as a label for training the Joint position prediction model to train the Joint position prediction model, so that a trained Joint position prediction model is obtained;
a vertex model obtaining module 34, configured to construct a Pose regression network model Pose _ net, where an output dimension of the Pose regression network model is the same as dimensions of N vertex coordinates of the standard 3D mesh model, and an input of the Pose regression network model is a Pose data rotation matrix form P of the standard 3D mesh modelmConstructing a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ net, and rotating a posture data axial angle form P of a standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmAnd taking N vertex coordinates T of the T-shaped posture 3D mesh model as input of the skeleton animation vertex correction model, and taking N vertex coordinates Gr of the standard 3D mesh model as labels for training the skeleton animation vertex correction model to train the skeleton animation vertex correction model, so as to obtain the trained skeleton animation vertex correction model.
Optionally, the training data obtaining module 31 includes:
the human body data acquisition unit is used for acquiring whole body images of different people with various actions including T-shaped gestures through a camera matrix and synchronously acquiring gesture data through the mobile capture equipment;
the 3D model acquisition unit is used for establishing a 3D point cloud model through a stereoscopic vision matching method according to the whole body images of various actions including T-shaped postures of different people, and registering and aligning the obtained 3D point cloud model by combining a geometric constraint and image projection constraint method to generate a standard 3D grid model and a T-shaped posture 3D grid model with a fixed vertex number N;
a vertex coordinate obtaining unit, configured to establish a coordinate system with the human pelvis point as a coordinate origin, and perform data normalization on the standard 3D mesh model to obtain N vertex coordinates Gr, Gr = [ g ] of the standard 3D mesh model0,g1,g2,…g(N-1)],giRepresenting the 3-dimensional coordinate value of each vertex of the standard 3D grid model, and carrying out data normalization on the T-shaped posture 3D grid model to obtain N vertex coordinates T of the T-shaped posture 3D grid model, wherein T = [ T ]0,t1,t2,…t(N-1)] ,ti3-dimensional coordinate values of each vertex of the T-shaped posture 3D grid model are represented;
the attitude data matching unit is used for establishing a one-to-one matching relationship between the acquired attitude data and the standard 3D grid model to obtain an attitude data axial angle form P of the standard 3D grid model, and converting the axial angle form P of the attitude data of the standard 3D grid model into a rotation matrix form Pm
Optionally, the joint model obtaining module 33 includes:
the Joint model building unit is used for building a Joint position prediction model Joint _ net, the input data dimension of the Joint position prediction model is N x 3, and the output data dimension is N x 3;
the joint model training unit is used for taking a tensor formed by converting N vertex coordinates T of the T-shaped posture 3D grid model into N x 3 as the input of the joint position prediction model, and taking a tensor formed by converting N joint point coordinates J of the T-shaped posture 3D grid model into N x 3 as a label for training the joint position prediction model to train the joint position prediction model;
a joint model optimization unit for optimizing a joint position prediction model with a mean square loss function, the joint position prediction model having an output of JpreThe mean square loss calculation formula is:
Figure 539552DEST_PATH_IMAGE006
where loss represents the mean square loss, J [ i ]]Vector formed by the real coordinates of the ith joint point, Jpre[i]Representing joint position prediction model prediction inputAnd finally obtaining a trained Joint position prediction model J = Joint _ net (T) by using a vector formed by coordinates of the ith Joint point.
Optionally, the vertex model obtaining module 34 includes:
the attitude model building unit is used for building an attitude regression network model Pose _ net, the input data dimension of the attitude regression network model is N x 3, and the output data dimension is N x 3;
a vertex model construction unit, configured to construct a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net, and a posture regression network model Pose _ net, where V represents a set of vertex coordinates output by the skeleton animation vertex correction model prediction, and V = [ V ], [0, v1…, vN-1],viCoordinate vector representing ith vertex, LBS linear skin function, T vertex coordinate of T-pose 3D mesh model, P, PmAxial angle form and rotation matrix form, Pose _ net (P), respectively representing Pose data of a standard 3D mesh modelm) Representing for each pose PmPredicting a vertex correction value, wherein Joint _ net (T) represents that n Joint point coordinates under the human T-shaped posture are predicted for each T-shaped posture, W represents the weight of a linear skin function, P represents the posture data of the T-shaped posture, W is set as a conductible parameter, and Joint _ net is set as an un-conductible parameter;
a vertex model training unit for converting the N vertex coordinates T of the T-shaped posture 3D grid model into tensor TinConverting the attitude data P of the standard 3D grid model into tensor Pin、PmInto a tensor PminThe method comprises the steps of training a skeleton animation vertex correction model by taking N vertex coordinates Gr of a standard 3D mesh model as input of the skeleton animation vertex correction model and taking the N vertex coordinates Gr as labels for training the skeleton animation vertex correction model;
a vertex model optimization unit for optimizing the skeleton animation vertex correction model by using a loss function including a loss of calculating a standard 3D mesh model vertex coordinate difference and a loss of a weight W of the skeleton animation vertex correction model, the mean square loss function of calculating the loss of the standard 3D mesh model vertex coordinate difference being:
Figure 345834DEST_PATH_IMAGE007
where loss _ v represents the loss value of the difference in the calculated vertex coordinates, G [ i]Coordinate vector representing the ith vertex of the standard 3D model, V [ i ]]A coordinate vector representing the ith vertex predicted by the skeleton animation vertex correction model; a loss function for calculating the loss of the weight W of the skeleton animation vertex correction model is
Figure 523393DEST_PATH_IMAGE008
Where loss _ w represents the skinning weight loss value, wi,jAnd representing the action weight of the j joint in the skinning function to the i vertex, wherein the combined loss function is as follows:
Figure 655297DEST_PATH_IMAGE009
and the loss represents the total loss value of the skeleton animation vertex correction model, represents the weight occupied by the loss of the vertex coordinate difference of the 3D mesh model, and finally obtains the trained skeleton animation vertex correction model.
The present application further provides a computer device, comprising:
a memory for storing a computer program;
a processor for executing the computer program to implement the steps of the learning method of the skeleton animation vertex correction model as described above.
The application provides a learning method of a skeleton animation vertex correction model, which comprises the steps of acquiring full-body images and posture data of various actions including T-shaped postures of different people to obtain posture data P of a standard 3D grid model and posture data rotation matrix form P of the standard 3D grid modelmThe method comprises the steps of obtaining N vertex coordinates Gr of a standard 3D grid model, N vertex coordinates T of a T-shaped posture 3D grid model, binding joint bones of the T-shaped posture 3D grid model to obtain N joint point coordinates J of the T-shaped posture 3D grid model, building and training a joint position prediction model, building a posture regression network model, building and training a skeleton animation vertex correction model comprising a standard skin function, the trained joint position prediction model and the posture regression network model, and obtaining and training a skeleton animation vertex correction model comprising the standard skin function, the trained joint position prediction model and the posture regression network modelAnd (5) training a bone animation vertex correction model.
Therefore, the skeleton animation vertex correction model provided by the application can compensate and correct the unreasonable deformation generated on the original skin surface of the character in the motion process, so that the character looks vivid, natural and reasonable in the motion process, the animation quality is improved, the workload of an animator in the animation post-production process is reduced, the time spent on producing the animation with the same quality is shorter, and the cost is lower.
The application also provides a learning device of the skeleton animation vertex correction model, a skeleton animation vertex correction method and computer equipment, and the learning device, the skeleton animation vertex correction method and the computer equipment all have the beneficial effects and are not repeated.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are only embodiments of the present application, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
FIG. 1 is a schematic flowchart illustrating a learning method of a skeleton animation vertex correction model according to an embodiment of the present disclosure;
fig. 2 is a schematic flowchart of a method for acquiring training data according to an embodiment of the present disclosure;
fig. 3 is a flowchart illustrating a method for obtaining a joint position prediction model according to an embodiment of the present disclosure;
fig. 4 is a schematic network structure diagram of a joint position prediction model according to an embodiment of the present application;
FIG. 5 is a flowchart illustrating a method for obtaining a skeleton animation vertex correction model according to an embodiment of the present disclosure;
fig. 6 is a schematic diagram of a network structure of a posture regression network model according to an embodiment of the present application
FIG. 7 is a schematic flowchart illustrating a method for correcting a vertex of a skeletal animation according to an embodiment of the present disclosure;
fig. 8 is a schematic structural diagram of a learning apparatus for a skeleton animation vertex correction model according to an embodiment of the present application.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present application clearer, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the scope of protection of the present application.
In the process of covering the character, for different character characters and different parts, an animator needs to give different weights by repeatedly observing the surface deformation in the character motion process, the workload is very complicated, and meanwhile, the final output animation is difficult to be ensured to be consistent with the expected performance, such as the problems of collapse, skin surface shriveling at the joint in a linear hybrid covering model and the like. The animation engineer is caused to repeatedly reset the weight of the role skinning, so that the quality and the output efficiency of the animation are greatly reduced, and the cost of the animation is high. The existing method of IK constraint solves the problem of skin collapse and solves the weight through human tissue dynamics, but the methods have the problems of large workload, long calculation time, inaccurate result, dependence on the experience of an animator on the final effect and the like, and cannot be widely applied.
Based on the above problems, the present embodiment provides a learning method for a skeleton animation vertex correction model, where the skeleton animation vertex correction model constructed based on a standard skinning function, a trained joint position prediction model and a posture regression network model can compensate and correct unreasonable deformation generated on the original skin surface of a character in a motion process, so that the character appears vivid, natural and reasonable in the motion process, the animation quality is improved, and meanwhile, the workload of an animator in an animation post-production process is reduced, so that the time spent on producing animations with the same quality is less, and the cost is lower.
Referring to fig. 1 specifically, fig. 1 is a method for learning a skeleton animation vertex correction model according to an embodiment of the present application, which specifically includes:
s11: collecting whole body images and posture data of different people including various actions of T-shaped postures, generating a standard 3D grid model with N vertexes and a T-shaped posture 3D grid model, and acquiring a posture data axial angle form P of the standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmThe three-dimensional model comprises N vertex coordinates Gr of a standard 3D grid model and N vertex coordinates T of a T-shaped posture 3D grid model, wherein N is a positive integer.
In one implementation, as shown in fig. 2, S11 includes:
s111: the whole-body images of different people including various actions of T-shaped postures are collected through a camera matrix, and meanwhile posture data are synchronously collected through a movable capturing device.
S112: according to the whole body images of different people including various actions of T-shaped postures, a 3D point cloud model is established through a stereoscopic vision matching method, and the obtained 3D point cloud model is registered and aligned by combining a geometric constraint method and an image projection constraint method to generate a standard 3D grid model and a T-shaped posture 3D grid model with a fixed vertex number N.
S113: establishing a coordinate system by taking human pelvis points as a coordinate origin, and performing data normalization on the standard 3D grid model to obtain N vertex coordinates Gr of the standard 3D grid model, wherein Gr = [ g ]0,g1,g2,…g(N-1)],giRepresenting the 3-dimensional coordinate value of each vertex of the standard 3D grid model, and carrying out data normalization on the T-shaped posture 3D grid model to obtain N vertex coordinates T of the T-shaped posture 3D grid model, wherein T = [ T ]0,t1,t2,…t(N-1)] ,tiAnd 3-dimensional coordinate values of each vertex of the T-shaped posture 3D grid model are represented.
S114: establishing a one-to-one matching relation between the acquired attitude data and the standard 3D grid model to obtain an attitude data axial angle form P, P = [ P ] of the standard 3D grid model0,p1, p2,…p(n-1)],piRepresenting Euler rotation angle of ith joint point of human skeleton relative to father node, converting shaft angle form P of attitude data of standard 3D grid model into rotation matrix form Pm,Pm=[p0 m, p1 m,…, pn-1 m],pi mAnd the rotation matrix of the ith joint point of the human skeleton relative to the parent node is represented.
In the present embodiment, whole-body images (shape data) and posture data of different persons are acquired using a camera array and a motion capture device, and are generated and registered as a standard grid. The annular array of the camera surrounds the shot collection object, the specific form of the camera array is not limited in the embodiment, and the camera array can be designed according to actual requirements. For example, in the embodiment, 48 cameras are arranged in an annular manner, the cameras are divided into 16 columns, 3 cameras in each column are vertically arranged at intervals of 0.7 m, the 16 columns of cameras uniformly surround the studio for one circle to achieve 360-degree dead-angle-free coverage on the human body, and the cameras must have a synchronous acquisition function to ensure that the acquired images can achieve synchronous alignment of each image and parameter calibration is performed on the cameras before data acquisition each time.
In this embodiment, the camera array is used to collect data of various movements of the human body (mainly including daily movement movements of the human body such as running, walking, squatting, dancing, etc.), and since the cameras adopt the frame synchronization technology, each frame of data between the cameras can be matched one by one. Each frame of data collected by the camera needs to store a time stamp which is accurate to millisecond. Then, a 3D point cloud model is established through a stereo vision matching method, and the obtained 3D point cloud model is registered and aligned by combining a geometric constraint and an image projection constraint method to generate a standard 3D mesh model with a fixed number N of vertices. Then, a coordinate system is established by taking the human pelvis point as the origin of coordinates (the directions of coordinate axes can be random, but need to be consistent front and back), and the standard is matchedThe 3D grid model is subjected to data normalization to obtain a three-dimensional coordinate value Gr = [ g ] of each vertex0,g1,g2,…g(N-1)],gi=[x,y,z]Representing 3-dimensional coordinate values of each vertex of the 3D model, acquiring T-type attitude data of each person in a T-Pose (T-type attitude) state to generate a T-type attitude 3D mesh model, normalizing the coordinates of each T-type attitude 3D mesh model to obtain three-dimensional coordinate values T = [ T ] of each vertex0,t1,t2,…t(N-1)];ti=[x,y,z]Representing the three-dimensional space coordinates of the vertices of the T-shaped pose 3D mesh model, T ∈ RN×3T _ psoe has a unified attitude data P = [ P ]0,p*1,p*2,...p*n-1]This amounts to a constant.
Specifically, before data is collected, the time of the movable capture device is adjusted to be consistent with the time of the camera, the frame rate of the movable capture data is set to be more than twice of the image collection frame rate of the camera, a timestamp of each frame of data is required to be stored when the movable capture device stores the data, and the timestamp of each frame of data can be stored when the movable capture device stores the data when the data is collected. After the acquisition is finished, comparing the time stamp of the dynamic capture data with the time stamp of the image data acquired by the camera array, and when the difference value of the two time stamps is smaller than a set threshold (the set threshold is 4 ms), considering that the dynamic capture data and the image data are acquired simultaneously to obtain a piece of matching data, so that the dynamic capture attitude motion data and a standard 3D grid model form a one-to-one matching relation, and simultaneously obtaining dynamic capture data P = [ P ] data0,p1,p2, …pn-1]Pi represents the Euler rotation angle of the ith joint point of the human skeleton relative to the father node, pi=[θ1,θ2,θ3],θ1,θ2,θ3Respectively representing a pitch angle, a yaw angle and a rolling angle (the sequence can be arbitrary, and the front and the back are consistent). Normalizing the rotation angle of the pose data P: p = [ P ]0,p1, p2,…p(n-1)]Handle PjConversion into rotation matrix P by standard rotation angle conversion formulaj m,Pj m∈R3×3,Pm=[p0 m, p1 m,…, pn-1 m],pj mA rotation matrix representing the jth joint point of the human skeleton relative to the parent node.
S12: and (3) carrying out joint skeleton binding on the T-shaped posture 3D grid model to obtain n joint point coordinates J of the T-shaped posture 3D grid model, wherein n is a positive integer.
In this embodiment, an animator can perform high-precision joint skeleton binding on a T-Pose 3D mesh model obtained under T _ dose by using a maya tool and derive joint point coordinates J, J = [ J ])0,j1,j2,…j(n-1)]Each T-type pose 3D grid model has n joint points, each joint contains x, y and z coordinates, and J belongs to Rn×3
S13: and constructing a Joint position prediction model Joint _ net, wherein the output dimensionality of the Joint position prediction model is the same as the dimensionality of N Joint point coordinates of the T-shaped posture 3D model, taking N vertex coordinates T of the T-shaped posture 3D grid model as the input of the Joint position prediction model, and taking N Joint point coordinates J of the T-shaped posture 3D grid model as a label for training the Joint position prediction model to train the Joint position prediction model, so as to obtain the trained Joint position prediction model.
In one implementation, as shown in fig. 3, S13 includes:
s131: constructing a Joint position prediction model Joint _ net, wherein the input data dimension of the Joint position prediction model is N x 3, and the output data dimension is N x 3;
s132: taking a tensor of converting N vertex coordinates T of the T-shaped posture 3D grid model into N x 3 as input of a joint position prediction model, and taking a tensor of converting N joint point coordinates J of the T-shaped posture 3D grid model into N x 3 as a label of a training joint position prediction model to train the joint position prediction model;
s133: optimizing a joint position prediction model by adopting a mean square loss function, wherein the output of the joint position prediction model is JpreThe mean square loss calculation formula is:
Figure 915377DEST_PATH_IMAGE010
where loss represents the mean square loss, J [ i ]]Vector formed by the real coordinates of the ith joint point, Jpre[i]And (3) a vector formed by coordinates of the ith Joint point which is output by the Joint position prediction model prediction is represented, and finally the trained Joint position prediction model J = Joint _ net (T) is obtained.
In this embodiment, a Joint position prediction model Joint _ net is constructed, as shown in fig. 4, where Input: n × 3 represents the input data dimension as N × 3, FC and ReLu represent the fully-connected layer and ReLu-activated layer of the neural network, respectively, the middle number represents the dimension of the Output data of each layer of the neural network, Output: n x 3 denotes the output data dimension is n x 3. Setting the input of Joint _ net to be Nx 3 (the registered 3D grid model has N vertexes, each vertex contains x, y and z coordinates), setting the output to be nx3 (each 3D grid model has N Joint points, each Joint contains x, y and z coordinates), using the N vertex coordinates T of the T-type posture 3D grid model obtained in the step S11 as the input, using the N Joint point coordinates J of the T-type posture 3D grid model as labels, using the mean square error function constraint to carry out the standardized learning training of the Joint position prediction model on the Joint _ net, and finally enabling the Joint position prediction model to predict the Joint coordinates of the T-type posture 3D grid model of any human body after the training convergence. I.e., J = Joint _ net (t).
The method comprises the following specific steps: (1) and constructing a Joint position prediction model Joint _ net, wherein the model structure is shown in figure 4. Joint _ net is a neural network with an input of N × 3 and an output of N × 3 (the data dimension of the T-type pose 3D mesh model established under T _ position is N × 3, and the dimension of Joint data J derived after bone binding is N × 3). (2) The N vertex coordinates T of the T-shaped posture 3D grid model are changed into an Nx 3 tensor, namely the N vertex coordinates T are spliced into an Nx 3 tensor which is used as the input of a Joint position prediction model Joint _ net, the N Joint point coordinates J of the T-shaped posture 3D grid model are changed into an N x 3 tensor, namely the N Joint point coordinates J are spliced into an N x 3 tensor which is used as a guide label for training the Joint position prediction model Joint _ net. (3) For Joint position prediction model Joint _ netThe mean square loss function is the optimization target, and the output of the network is JpreThe mean square loss calculation formula is as follows
Figure 525350DEST_PATH_IMAGE011
: where loss represents the mean square loss, J [ i ]]Vector formed by the real coordinates of the ith joint point, Jpre[i]And a vector composed of coordinates of the i-th joint point which represents the prediction output of the joint position prediction model. (4) Setting a training parameter of a Joint position prediction model Joint _ net, wherein the specific parameter in this embodiment is as follows: the learning rate is 0.001, the optimization method is Adam (adaptive time estimation method), the number of iterations is 100, and the Batch size is 8. (5) And training the Joint _ net by adopting a standard neural network training method to finally realize J = Joint _ net (T).
By constructing a Joint position prediction model Joint _ net, the coordinates of each Joint of any character role under T _ Point can be correctly predicted, the skeleton of the character role can be automatically bound through a simple program if the coordinate value of each Joint exists, the time spent on the skeleton binding of the character role is much, and the manufacturing time, the cost and the difficulty of the character role can be greatly reduced by the technology of the application.
S14: constructing a posture regression network model Pose _ net, wherein the output dimensionality of the posture regression network model is the same as the dimensionality of N vertex coordinates of the standard 3D grid model, and the input of the posture regression network model is a posture data rotation matrix form P of the standard 3D grid modelmConstructing a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ net, and rotating a posture data axial angle form P of a standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmAnd taking N vertex coordinates T of the T-shaped posture 3D mesh model as input of the skeleton animation vertex correction model, and taking N vertex coordinates Gr of the standard 3D mesh model as labels for training the skeleton animation vertex correction model to train the skeleton animation vertex correction model, so as to obtain the trained skeleton animation vertex correction model.
In one implementation, as shown in fig. 5, S14 includes:
s141: constructing a posture regression network model Pose _ net, wherein the input data dimension of the posture regression network model is N x 3, and the output data dimension is N x 3;
s142: and constructing a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ net, wherein V represents a set of vertex coordinates output by the skeleton animation vertex correction model in a prediction mode, and V = [ V ], [0, v1…, vN-1],viCoordinate vector representing ith vertex, LBS linear skin function, T vertex coordinate of T-pose 3D mesh model, P, PmAxial angle form and rotation matrix form, Pose _ net (P), respectively representing Pose data of a standard 3D mesh modelm) Representing for each pose PmPredicting a vertex correction value, wherein Joint _ net (T) represents that n Joint point coordinates under the human T-shaped posture are predicted for each T-shaped posture, W represents the weight of a linear skin function, P represents the posture data of the T-shaped posture, W is set as a conductible parameter, and Joint _ net is set as an un-conductible parameter;
s143: converting N vertex coordinates T of T-shaped posture 3D grid model into tensor TinConverting the attitude data P of the standard 3D grid model into tensor Pin、PmInto a tensor PminThe method comprises the steps of training a skeleton animation vertex correction model by taking N vertex coordinates Gr of a standard 3D mesh model as input of the skeleton animation vertex correction model and taking the N vertex coordinates Gr as labels for training the skeleton animation vertex correction model;
s144: optimizing the skeleton animation vertex correction model by adopting a loss function comprising the loss of calculating the vertex coordinate difference of the standard 3D mesh model and the loss of the weight W of the skeleton animation vertex correction model, wherein the mean square loss function for calculating the loss of the vertex coordinate difference of the standard 3D mesh model is as follows:
Figure 554486DEST_PATH_IMAGE012
where loss _ v represents the loss value of the difference in the calculated vertex coordinates, G [ i]Coordinate vector representing the ith vertex of the standard 3D model, V [ i ]]Vertex correction model for representing skeleton animationMeasuring the coordinate vector of the ith vertex; a loss function for calculating the loss of weight W of the skeleton animation vertex correction model is as follows, wherein loss _ W represents a skin weight loss value, Wi,jAnd representing the action weight of the j joint in the skinning function to the i vertex, wherein the combined loss function is as follows:
Figure 591712DEST_PATH_IMAGE013
and the loss represents the total loss value of the skeleton animation vertex correction model, represents the weight occupied by the loss of the vertex coordinate difference of the 3D mesh model, and finally obtains the trained skeleton animation vertex correction model.
In this embodiment, a posture regression network model Pose _ net is constructed, as shown in fig. 6, where Input: n × 3 × 3 represents the input data dimension as n × 3 × 3, FC and ReLu represent the fully-connected layer and the RuLu active layer of the neural network, respectively, the middle number represents the dimension of the Output data of each layer of the neural network, Output: n × 3 indicates an output data dimension of N × 3. The input to set the Pose _ net is nx3 × 3 (the registered 3D mesh model has N joint points, each joint point is represented by a rotation matrix and then is a 3 × 3 matrix), the output is nx3 (the registered 3D mesh model has N vertices, each vertex contains x, y, and z coordinates), and the final vertex coordinates V are calculated using a standard skinning function, where V is ═ V0,v1,v2,…,vN-1]。
In this embodiment, a skinning function is used, the embodiment does not specifically limit the type of the skinning function, and the method is applicable to any skinning algorithm in principle.
Specifically, a method of combining a linear skinning function (LBS) with a posture regression network model, Pose _ net, is used to improve the standard skinning function in combination with the posture regression network model, specifically, the standard linear skinning formula is V ═ LBS (T, P, J, W, P), where,i.e. where v isiCoordinate vector representing the ith vertex, wi,jAnd expressing the action weight of the j joint to the i vertex in the skin function, wherein the formula means that: the vertex of the 3D model in the current pose changes with respect to the rotation in the T-pose, and the meaning of the formula is: calculate the rotation of the jth joint point relative to the root node at the current pose, P, PmRespectively representing the form of the shaft angle and the form of a rotation matrix of the attitude data of the standard 3D grid model, J representing the coordinates of n joint points of the T-shaped attitude 3D grid model, Pj mThe posture data of j-th joint point in the character movement process is expressed in the form of rotation matrix, giAnd 3-dimensional coordinate values representing the ith vertex of the 3D character at T _ position. P represents the posture data of the T-shaped posture, is a constant, A (j) represents a path set formed by the shortest paths from the joint nodes to the root nodes, and is an ordered vector set from the root nodes to the nodes, wherein W = [ W = ]0,0,w0,1…w0,n;w1,0, w1,1…w1,n-1, … wN-1,n-1],wi,jRepresenting the action weight of the jth joint point of the 3D role model in the linear skin model on the ith vertex for wi,jComprises the following steps: 0<wi,j<1 and (b).
In this embodiment, a calculation formula of using a linear skin function in combination with a Pose regression network model Pose _ net is as follows:
Figure 339088DEST_PATH_IMAGE014
Figure 752752DEST_PATH_IMAGE015
in (d), since J can be derived from J = Joint _ net (t), the formula can be written as:
Figure 839657DEST_PATH_IMAGE016
namely:
Figure 313363DEST_PATH_IMAGE017
wherein W is the weight of the skin,Pj mIs attitude data PjIs represented by a rotation matrix of 3 x 3 (the rotation of each joint point is represented by a rotation matrix of 3 x 3), Pm=[p0 m, p1 m,…, pn-1 m]And setting W as a variable, so that in the process of training the skeleton animation vertex correction model, W becomes a variable parameter, and the skin function after improvement is a trainable neural network. Since the Joint position prediction model Joint _ net is trained, the Joint _ net parameter should be set to be non-conductive in the training LBS process, so that the Joint _ net parameter remains unchanged in the training LBS process.
Wherein, in the embodiment, each bit W of the weight value W in the LBS is improvedi,jSetting the skin function combined with Pose _ net as a neural network model, namely a skeleton animation vertex correction model, as a learnable variable, training the skeleton animation vertex correction model by using a neural network standardized training method, and optimizing w through the training and learning of the skeleton animation vertex correction modeli,jAnd (4) parameters. T _ pos model data T = [ T ]0,t1,t2,…t(N-1)]And corresponding rotation matrix representation form data P of attitude datam∈Rn×3×3As the input of the vertex correction model of the skeleton animation, the N vertex coordinates Gr of the standard 3D mesh model, Gr = [ g ] are used0,g1,g2,…g(N-1)]As a label for training the skeleton animation vertex correction model. In the improved model, the role of the posture regression network model Pose _ net is equivalent to predicting the correction value v of each vertex of the 3D model under each posture data P for each posture dataTherefore, a part of the shortage of the LBS is compensated.
Specifically, the acquired 3D model data T = [ T ] of T _ pos0,t1,t2,…t(N-1)]Constructing tensor Tin and tensor P constructed by attitude data PinAnd converting P into a representation form P of a rotation matrixmConstructed tensor PminAs input data, Gr is used as a label, and a loss function is designed in two parts, namely: standard of meritLoss of the vertex coordinate difference of the 3D mesh model and loss of the weight W of the improved linear skin model (skeleton animation vertex correction model), and the skin model is optimized and improved by adopting a standard neural network learning method. Loss of the standard 3D mesh model vertex coordinate difference using the mean square loss function, Gr = [ g =0,g1,g2,…g(N-1)]The output of the skeletal animation vertex correction model V = [ V ] for N vertex coordinates of the standard 3D mesh model0, v1…, vN-1]Calculating the mean square loss function of the loss of the vertex coordinate difference of the standard 3D mesh model as follows:
Figure 282456DEST_PATH_IMAGE018
where loss _ v represents the loss value of the difference in the calculated vertex coordinates, G [ i]Coordinate vector representing the ith vertex of the standard 3D model, V [ i ]]A coordinate vector representing the ith vertex predicted by the skeleton animation vertex correction model; a loss function for calculating the loss of weight W of the skeleton animation vertex correction model is as follows, wherein loss _ W represents a skin weight loss value, Wi,jAnd representing the action weight of the j-th joint on the i-th vertex in the skinning function. In the training phase, each value of W is monitored before updating the parameters for each iteration, and W = [ W =0,0, w0,1…w0,n;w1,0, w1,1…w1,n-1 … wN-1,n-1]W ini,jWhen w isi,j<At 0, w is set artificiallyi,j=0, when wi,j>At 1, w is set artificiallyi,jAnd = 1. The combined loss function is:
Figure 499811DEST_PATH_IMAGE019
and the loss represents the total loss value of the skeleton animation vertex correction model and the weight of the loss of the vertex coordinate difference of the 3D mesh model. Setting training parameters for improving the LBS, the specific parameters in this embodiment are: the learning rate is 0.001, the optimization method is Adam (adaptive time estimation method), the number of iterations is 100, the Batch size is 8, and = 0.7. Training the skeleton animation vertex correction model by adopting a standard neural network training method to finally obtain the trained skeleton animation vertex correction modelAnd (4) modeling.
The advantages of the present application include: the method has the advantages that the camera rectangular array is used for collecting and reconstructing the 3D high-dimensional grid curved surface of the human body, the posture regression network model learns the deformation rule of the surface of the human body caused by the action by the machine learning method, the skin model combined with the posture regression network model learns the proper weight value by the method, the posture regression network model can output the proper skin weight value by using the Pose _ net technology, the complicated weight brushing process is omitted in the animation manufacturing process, the workload of practitioners is greatly reduced, the accuracy is improved, the less time is spent to obtain higher animation quality, the time is saved, and the animation manufacturing cost is reduced.
In addition, the method outputs the offset of one vertex to an improved skinning function through a vertex correction model of the skeletal animation, leads the vertex correction model of the skeletal animation to learn the deformation rule of the skin surface of the human body under different actions, can compensate and correct the unreasonable deformation generated on the original skin surface of the character in the motion process through the automatic Joint _ net of the character skeleton and the automatic Pose _ net of the character vertex compensation and repair technology, avoids the problems of skin surface collapse, distortion and the like of the character in certain postures in the motion process, thereby avoiding the unreasonable deformation, leading the character to look vivid, natural and reasonable in the motion process, improving the animation quality, reducing the workload of an animator in the post-production process of the animation, leading the production of the animation with the same quality and less time spent by using the method, the cost is lower.
The present embodiment further provides a method for correcting a vertex of a bone animation, as shown in fig. 7, specifically including:
s21: acquiring posture data and a T-shaped posture 3D grid model of a human body;
s22: inputting the posture data of the human body and the T-shaped posture 3D mesh model into the trained skeleton animation vertex correction model to obtain N vertex coordinates of the standard 3D mesh model of the human body in the posture.
In this embodiment, the trained skeletal animation is savedNetwork parameters of the vertex correction model, loading parameters into the bone animated vertex correction model V = LBS (T _ + pos _ net (Pm _), P _, Joint _ net (T _), W, P _) and keeping all parameters fixed, where W is the learned skinning weight after training. Making a T-shaped posture 3D grid model T \undera person T _ Point, providing any reasonable human skeleton posture data P \, converting P _ into an expression form Pm _ of a rotation matrix as an input variable to be input into a skeleton animation vertex correction model, namely V = LBS (T _ + Point _ net (Pm _), P _, Joint _ net (T _), W, P _), and obtaining V = [ V _ [ ] [ V _ ] ]0, v1…, vN-1]The space coordinates of the vertexes of the human standard 3D mesh model in the posture are obtained.
The following describes a learning apparatus for a skeleton animation vertex correction model according to an embodiment of the present application, and the learning apparatus for a skeleton animation vertex correction model described below and the learning method for a skeleton animation vertex correction model described above may be referred to in correspondence.
Referring to fig. 8, fig. 8 is a schematic structural diagram of a learning apparatus for a skeleton animation vertex correction model according to an embodiment of the present application, which specifically includes:
a training data obtaining module 31, configured to collect whole body images and posture data of various actions including T-shaped postures of different people, generate a standard 3D mesh model and a T-shaped posture 3D mesh model having N vertices, and obtain a posture data axis angle form P of the standard 3D mesh model and a posture data rotation matrix form P of the standard 3D mesh modelmN vertex coordinates Gr of the standard 3D grid model and N vertex coordinates T of the T-shaped posture 3D grid model, wherein N is a positive integer;
the joint data acquisition module 32 is used for binding joint bones of the T-shaped posture 3D grid model to obtain n joint point coordinates J of the T-shaped posture 3D grid model, wherein n is a positive integer;
the Joint model obtaining module 33 is configured to construct a Joint position prediction model Joint _ net, an output dimension of the Joint position prediction model is the same as a dimension of N Joint point coordinates of the T-type posture 3D model, N vertex coordinates T of the T-type posture 3D mesh model are used as an input of the Joint position prediction model, and N Joint point coordinates J of the T-type posture 3D mesh model are used as a label for training the Joint position prediction model to train the Joint position prediction model, so that a trained Joint position prediction model is obtained;
a vertex model obtaining module 34, configured to construct a Pose regression network model Pose _ net, where an output dimension of the Pose regression network model is the same as dimensions of N vertex coordinates of the standard 3D mesh model, and an input of the Pose regression network model is a Pose data rotation matrix form P of the standard 3D mesh modelmConstructing a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ net, and rotating a posture data axial angle form P of a standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmAnd taking N vertex coordinates T of the T-shaped posture 3D mesh model as input of the skeleton animation vertex correction model, and taking N vertex coordinates Gr of the standard 3D mesh model as labels for training the skeleton animation vertex correction model to train the skeleton animation vertex correction model, so as to obtain the trained skeleton animation vertex correction model.
Preferably, the training data acquisition module 31 includes:
the human body data acquisition unit is used for acquiring whole body images of different people with various actions including T-shaped gestures through a camera matrix and synchronously acquiring gesture data through the mobile capture equipment;
the 3D model acquisition unit is used for establishing a 3D point cloud model through a stereoscopic vision matching method according to the whole body images of various actions including T-shaped postures of different people, and registering and aligning the obtained 3D point cloud model by combining a geometric constraint and image projection constraint method to generate a standard 3D grid model and a T-shaped posture 3D grid model with a fixed vertex number N;
a vertex coordinate obtaining unit, configured to establish a coordinate system with the human pelvis point as a coordinate origin, and perform data normalization on the standard 3D mesh model to obtain N vertex coordinates Gr, Gr = [ g ] of the standard 3D mesh model0,g1,g2,…g(N-1)],giExpressing the 3-dimensional coordinate value of each vertex of the standard 3D grid model, and performing data normalization on the T-shaped posture 3D grid model to obtain the T-shaped postureN vertex coordinates T, T = [ T ] of 3D mesh model0,t1,t2,…t(N-1)] ,ti3-dimensional coordinate values of each vertex of the T-shaped posture 3D grid model are represented;
the attitude data matching unit is used for establishing a one-to-one matching relationship between the acquired attitude data and the standard 3D grid model to obtain an attitude data axial angle form P of the standard 3D grid model, and converting the axial angle form P of the attitude data of the standard 3D grid model into a rotation matrix form Pm
Preferably, the joint model acquisition module 33 includes:
the Joint model building unit is used for building a Joint position prediction model Joint _ net, the input data dimension of the Joint position prediction model is N x 3, and the output data dimension is N x 3;
the joint model training unit is used for taking a tensor formed by converting N vertex coordinates T of the T-shaped posture 3D grid model into N x 3 as the input of the joint position prediction model, and taking a tensor formed by converting N joint point coordinates J of the T-shaped posture 3D grid model into N x 3 as a label for training the joint position prediction model to train the joint position prediction model;
a joint model optimization unit for optimizing a joint position prediction model with a mean square loss function, the joint position prediction model having an output of JpreThe mean square loss calculation formula is:
Figure 503539DEST_PATH_IMAGE020
where loss represents the mean square loss, J [ i ]]Vector formed by the real coordinates of the ith joint point, Jpre[i]And (3) a vector formed by coordinates of the ith Joint point which is output by the Joint position prediction model prediction is represented, and finally the trained Joint position prediction model J = Joint _ net (T) is obtained.
Preferably, the vertex model obtaining module 34 includes:
the attitude model building unit is used for building an attitude regression network model Pose _ net, the input data dimension of the attitude regression network model is N x 3, and the output data dimension is N x 3;
a vertex model construction unit for combining the standard skin function and trainingConstructing a skeleton animation vertex correction model by using a Joint position prediction model Joint _ net and a posture regression network model Pose _ net, wherein V represents a set of vertex coordinates predicted and output by the skeleton animation vertex correction model, and V = [ V ]0, v1…, vN-1],viCoordinate vector representing ith vertex, LBS linear skin function, T vertex coordinate of T-pose 3D mesh model, P, PmAxial angle form and rotation matrix form, Pose _ net (P), respectively representing Pose data of a standard 3D mesh modelm) Representing for each pose PmPredicting a vertex correction value, wherein Joint _ net (T) represents that n Joint point coordinates under the human T-shaped posture are predicted for each T-shaped posture, W represents the weight of a linear skin function, P represents the posture data of the T-shaped posture, W is set as a conductible parameter, and Joint _ net is set as an un-conductible parameter;
a vertex model training unit for converting N vertex coordinates T of the T-shaped posture 3D grid model into tensor Tin and converting posture data P of the standard 3D grid model into tensor Pin、PmInto a tensor PminThe method comprises the steps of training a skeleton animation vertex correction model by taking N vertex coordinates Gr of a standard 3D mesh model as input of the skeleton animation vertex correction model and taking the N vertex coordinates Gr as labels for training the skeleton animation vertex correction model;
a vertex model optimization unit for optimizing the skeleton animation vertex correction model by using a loss function including a loss of calculating a standard 3D mesh model vertex coordinate difference and a loss of a weight W of the skeleton animation vertex correction model, the mean square loss function of calculating the loss of the standard 3D mesh model vertex coordinate difference being: where loss _ v represents the loss value of the difference in the calculated vertex coordinates, G [ i]Coordinate vector representing the ith vertex of the standard 3D model, V [ i ]]A coordinate vector representing the ith vertex predicted by the skeleton animation vertex correction model; a loss function for calculating the loss of weight W of the skeleton animation vertex correction model is as follows, wherein loss _ W represents a skin weight loss value, Wi,jAnd representing the action weight of the j joint in the skinning function to the i vertex, wherein the combined loss function is as follows: wherein loss represents the total loss value of the skeleton animation vertex correction model and represents the vertex coordinate difference of the 3D mesh modelThe loss of the skeleton animation takes up weight, and finally the trained skeleton animation vertex correction model is obtained.
Since the embodiments of the apparatus portion and the method portion correspond to each other, please refer to the description of the embodiments of the method portion for the embodiments of the apparatus portion, which is not repeated here.
In the following, a computer device provided by the embodiments of the present application is introduced, and the computer device described below and the method described above may be referred to correspondingly.
The present application further provides a computer device, comprising: a memory for storing a computer program; a processor for executing the computer program to implement the steps of the learning method of the skeleton animation vertex correction model as described above.
Since the embodiment of the computer device portion and the embodiment of the method portion correspond to each other, please refer to the description of the embodiment of the method portion for the embodiment of the computer device portion, which is not repeated here.
The embodiments are described in a progressive manner in the specification, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other. The device disclosed by the embodiment corresponds to the method disclosed by the embodiment, so that the description is simple, and the relevant points can be referred to the method part for description.
Those of skill would further appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both, and that the various illustrative components and steps have been described above generally in terms of their functionality in order to clearly illustrate this interchangeability of hardware and software. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in Random Access Memory (RAM), memory, Read Only Memory (ROM), electrically programmable ROM, electrically erasable programmable ROM, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art.
The method for learning a skeleton animation vertex correction model, the device for learning a skeleton animation vertex correction model, the method for correcting a skeleton animation vertex, and the computer device provided by the present application are described in detail above. The principles and embodiments of the present application are explained herein using specific examples, which are provided only to help understand the method and the core idea of the present application. It should be noted that, for those skilled in the art, it is possible to make several improvements and modifications to the present application without departing from the principle of the present application, and such improvements and modifications also fall within the scope of the claims of the present application.

Claims (10)

1. A learning method of a skeleton animation vertex correction model is characterized by comprising the following steps:
s11: collecting whole body images and posture data of different people including various actions of T-shaped postures, generating a standard 3D grid model with N vertexes and a T-shaped posture 3D grid model, and acquiring a posture data axial angle form P of the standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmN vertex coordinates Gr of the standard 3D grid model and N vertex coordinates T of the T-shaped posture 3D grid model, wherein N is a positive integer;
s12: joint skeleton binding is carried out on the T-shaped posture 3D grid model to obtain n joint point coordinates J of the T-shaped posture 3D grid model, wherein n is a positive integer;
s13: constructing a Joint position prediction model Joint _ net, wherein the output dimensionality of the Joint position prediction model is the same as the dimensionality of N Joint point coordinates of the T-shaped posture 3D model, taking N vertex coordinates T of the T-shaped posture 3D grid model as the input of the Joint position prediction model, and taking N Joint point coordinates J of the T-shaped posture 3D grid model as labels of a training Joint position prediction model to train the Joint position prediction model, so as to obtain the trained Joint position prediction model;
s14: constructing a posture regression network model Pose _ net, wherein the output dimensionality of the posture regression network model is the same as the dimensionality of N vertex coordinates of the standard 3D grid model, and the input of the posture regression network model is a posture data rotation matrix form P of the standard 3D grid modelmConstructing a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ net, and rotating a posture data axial angle form P of a standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmAnd taking N vertex coordinates T of the T-shaped posture 3D mesh model as input of the skeleton animation vertex correction model, and taking N vertex coordinates Gr of the standard 3D mesh model as labels for training the skeleton animation vertex correction model to train the skeleton animation vertex correction model, so as to obtain the trained skeleton animation vertex correction model.
2. The learning method of the skeleton animation vertex correction model according to claim 1, wherein the step S11 includes:
s111: acquiring whole-body images of different people including various actions of T-shaped postures through a camera matrix, and synchronously acquiring posture data through a movable capturing device;
s112: establishing a 3D point cloud model by a stereoscopic vision matching method according to the whole body images of different people including various actions of T-shaped postures, and registering and aligning the obtained 3D point cloud model by combining a geometric constraint and image projection constraint method to generate a standard 3D grid model and a T-shaped posture 3D grid model with a fixed vertex number N;
s113: establishing a coordinate system by taking human pelvis points as a coordinate origin, performing data normalization on the standard 3D grid model to obtain N vertex coordinates Gr of the standard 3D grid model, and performing data normalization on the T-shaped posture 3D grid model to obtain N vertex coordinates T of the T-shaped posture 3D grid model;
s114: establishing a one-to-one matching relation between the acquired attitude data and the standard 3D grid model to obtain an attitude data axial angle form P of the standard 3D grid model, and converting the attitude data axial angle form P of the standard 3D grid model into a rotation matrix form Pm
3. The learning method of the skeleton animation vertex correction model according to claim 1, wherein the step S13 includes:
s131: constructing a Joint position prediction model Joint _ net, wherein the input data dimension of the Joint position prediction model is N x 3, and the output data dimension is N x 3;
s132: taking a tensor of converting N vertex coordinates T of the T-shaped posture 3D grid model into N x 3 as input of a joint position prediction model, and taking a tensor of converting N joint point coordinates J of the T-shaped posture 3D grid model into N x 3 as a label of a training joint position prediction model to train the joint position prediction model;
s133: optimizing a joint position prediction model by adopting a mean square loss function, wherein the output of the joint position prediction model is JpreThe mean square loss calculation formula is:
Figure 440014DEST_PATH_IMAGE001
where loss represents the mean square loss, J [ i ]]Vector formed by the real coordinates of the ith joint point, Jpre[i]And (3) a vector formed by coordinates of the ith Joint point which is output by the Joint position prediction model prediction is represented, and finally the trained Joint position prediction model J = Joint _ net (T) is obtained.
4. The learning method of an animated skeleton vertex correction model according to claim 1, wherein the step S14 comprises:
s141: constructing a posture regression network model Pose _ net, wherein the input data dimension of the posture regression network model is N x 3, and the output data dimension is N x 3;
s142: constructing a skeleton animation top by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ netPoint correction model
Figure 891243DEST_PATH_IMAGE002
Where V represents the set of vertex coordinates of the skeletal animation vertex correction model prediction output, and V = [ V ]0, v1…, vN-1],viCoordinate vector representing ith vertex, LBS linear skin function, T vertex coordinate of T-pose 3D mesh model, P, PmRepresenting the attitude data axis angle form and the rotation matrix form, Pose _ net (P), respectively, of a standard 3D mesh modelm) Representing for each pose PmPredicting a vertex correction value, wherein Joint _ net (T) represents that n Joint point coordinates under the human T-shaped posture are predicted for each T-shaped posture, W represents the weight of a linear skin function, P represents the posture data of the T-shaped posture, W is set as a conductible parameter, and Joint _ net is set as an un-conductible parameter;
s143: converting N vertex coordinates T of T-shaped posture 3D grid model into tensor TinConverting the attitude data P of the standard 3D grid model into tensor Pin、PmInto a tensor PminThe method comprises the steps of training a skeleton animation vertex correction model by taking N vertex coordinates Gr of a standard 3D mesh model as input of the skeleton animation vertex correction model and taking the N vertex coordinates Gr as labels for training the skeleton animation vertex correction model;
s144: optimizing the skeleton animation vertex correction model by adopting a loss function comprising the loss of calculating the vertex coordinate difference of the standard 3D mesh model and the loss of the weight W of the skeleton animation vertex correction model, wherein the mean square loss function for calculating the loss of the vertex coordinate difference of the standard 3D mesh model is as follows:
Figure 202138DEST_PATH_IMAGE003
where loss _ v represents the loss value of the difference in the calculated vertex coordinates, G [ i]Coordinate vector representing the ith vertex of the standard 3D model, V [ i ]]A coordinate vector representing the ith vertex predicted by the skeleton animation vertex correction model; a loss function for calculating the loss of the weight W of the skeleton animation vertex correction model is
Figure 128506DEST_PATH_IMAGE004
Where loss _ w represents the skinning weight loss value, wi,jAnd representing the action weight of the j joint in the skinning function to the i vertex, wherein the combined loss function is as follows:
Figure 208457DEST_PATH_IMAGE005
and the loss represents the total loss value of the skeleton animation vertex correction model, represents the weight occupied by the loss of the vertex coordinate difference of the 3D mesh model, and finally obtains the trained skeleton animation vertex correction model.
5. A bone animation vertex correction method is characterized by comprising the following steps:
acquiring posture data and a T-shaped posture 3D grid model of a human body;
inputting the posture data of the human body and the T-shaped posture 3D mesh model into the trained skeleton animation vertex correction model to obtain N vertex coordinates of the standard 3D mesh model of the human body in the posture;
wherein the skeleton animation vertex correction model is trained by adopting the learning method of the skeleton animation vertex correction model as claimed in any one of claims 1 to 4.
6. A learning device of a skeleton animation vertex correction model is characterized by comprising:
a training data acquisition module (31) for acquiring the whole body images and posture data of various actions including T-shaped postures of different people, generating a standard 3D grid model and a T-shaped posture 3D grid model with N vertexes, and acquiring a posture data axial angle form P of the standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmN vertex coordinates Gr of the standard 3D grid model and N vertex coordinates T of the T-shaped posture 3D grid model, wherein N is a positive integer;
the joint data acquisition module (32) is used for binding joint bones of the T-shaped posture 3D grid model to obtain n joint point coordinates J of the T-shaped posture 3D grid model, wherein n is a positive integer;
the Joint model obtaining module (33) is used for constructing a Joint position prediction model Joint _ net, the output dimensionality of the Joint position prediction model is the same as the dimensionality of N Joint point coordinates of the T-shaped posture 3D model, N vertex coordinates T of the T-shaped posture 3D grid model are used as the input of the Joint position prediction model, N Joint point coordinates J of the T-shaped posture 3D grid model are used as labels of a training Joint position prediction model to train the Joint position prediction model, and the trained Joint position prediction model is obtained;
a vertex model obtaining module (34) for constructing a posture regression network model Pose _ net, the output dimensionality of the posture regression network model is the same as the dimensionality of N vertex coordinates of the standard 3D grid model, and the input of the posture regression network model is a posture data rotation matrix form P of the standard 3D grid modelmConstructing a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ net, and rotating a posture data axial angle form P of a standard 3D grid model and a posture data rotation matrix form P of the standard 3D grid modelmAnd taking N vertex coordinates T of the T-shaped posture 3D mesh model as input of the skeleton animation vertex correction model, and taking N vertex coordinates Gr of the standard 3D mesh model as labels for training the skeleton animation vertex correction model to train the skeleton animation vertex correction model, so as to obtain the trained skeleton animation vertex correction model.
7. The learning device of a skeletal animation vertex correction model according to claim 6, wherein the training data acquisition module (31) comprises:
the human body data acquisition unit is used for acquiring whole body images of different people with various actions including T-shaped gestures through a camera matrix and synchronously acquiring gesture data through the mobile capture equipment;
the 3D model acquisition unit is used for establishing a 3D point cloud model through a stereoscopic vision matching method according to the whole body images of various actions including T-shaped postures of different people, and registering and aligning the obtained 3D point cloud model by combining a geometric constraint and image projection constraint method to generate a standard 3D grid model and a T-shaped posture 3D grid model with a fixed vertex number N;
a vertex coordinate obtaining unit, configured to establish a coordinate system with the human pelvis point as a coordinate origin, and perform data normalization on the standard 3D mesh model to obtain N vertex coordinates Gr, Gr = [ g ] of the standard 3D mesh model0,g1,g2,…g(N-1)],giRepresenting the 3-dimensional coordinate value of each vertex of the standard 3D grid model, and carrying out data normalization on the T-shaped posture 3D grid model to obtain N vertex coordinates T of the T-shaped posture 3D grid model, wherein T = [ T ]0,t1,t2,…t(N-1)] ,ti3-dimensional coordinate values of each vertex of the T-shaped posture 3D grid model are represented;
the attitude data matching unit is used for establishing a one-to-one matching relationship between the acquired attitude data and the standard 3D grid model to obtain an attitude data axial angle form P of the standard 3D grid model, and converting the axial angle form P of the attitude data of the standard 3D grid model into a rotation matrix form Pm
8. The learning device of a skeletal animation vertex correction model according to claim 6, wherein the joint model acquisition module (33) comprises:
the Joint model building unit is used for building a Joint position prediction model Joint _ net, the input data dimension of the Joint position prediction model is N x 3, and the output data dimension is N x 3;
the joint model training unit is used for taking a tensor formed by converting N vertex coordinates T of the T-shaped posture 3D grid model into N x 3 as the input of the joint position prediction model, and taking a tensor formed by converting N joint point coordinates J of the T-shaped posture 3D grid model into N x 3 as a label for training the joint position prediction model to train the joint position prediction model;
a joint model optimization unit for optimizing a joint position prediction model with a mean square loss function, the joint position prediction model having an output of JpreThe mean square loss calculation formula is:
Figure 562078DEST_PATH_IMAGE006
where loss represents the mean square loss, J [ i ]]To representVector formed by the real coordinates of the i-th joint point, Jpre[i]And (3) a vector formed by coordinates of the ith Joint point which is output by the Joint position prediction model prediction is represented, and finally the trained Joint position prediction model J = Joint _ net (T) is obtained.
9. The learning device of a skeletal animation vertex correction model according to claim 6, wherein the vertex model obtaining module (34) comprises:
the attitude model building unit is used for building an attitude regression network model Pose _ net, the input data dimension of the attitude regression network model is N x 3, and the output data dimension is N x 3;
a vertex model construction unit for constructing a skeleton animation vertex correction model by combining a standard skin function, a trained Joint position prediction model Joint _ net and a posture regression network model Pose _ net
Figure 94691DEST_PATH_IMAGE007
Where V represents the set of vertex coordinates of the skeletal animation vertex correction model prediction output, and V = [ V ]0, v1…, vN-1],viCoordinate vector representing ith vertex, LBS linear skin function, T vertex coordinate of T-pose 3D mesh model, P, PmAxial angle form and rotation matrix form, Pose _ net (P), respectively representing Pose data of a standard 3D mesh modelm) Representing for each pose PmPredicting a vertex correction value, wherein Joint _ net (T) represents that n Joint point coordinates under the human T-shaped posture are predicted for each T-shaped posture, W represents the weight of a linear skin function, P represents the posture data of the T-shaped posture, W is set as a conductible parameter, and Joint _ net is set as an un-conductible parameter;
a vertex model training unit for converting the N vertex coordinates T of the T-shaped posture 3D grid model into tensor TinConverting the attitude data P of the standard 3D grid model into tensor Pin、PmInto a tensor PminInputting the vertex correction model of the skeletal animation, and using the N vertex coordinates Gr of the standard 3D mesh model as the vertex correction of the training skeletal animationTraining the skeleton animation vertex correction model by using the label of the positive model;
a vertex model optimization unit for optimizing the skeleton animation vertex correction model by using a loss function including a loss of calculating a standard 3D mesh model vertex coordinate difference and a loss of a weight W of the skeleton animation vertex correction model, the mean square loss function of calculating the loss of the standard 3D mesh model vertex coordinate difference being:
Figure 824749DEST_PATH_IMAGE008
where loss _ v represents the loss value of the difference in the calculated vertex coordinates, G [ i]Coordinate vector representing the ith vertex of the standard 3D model, V [ i ]]A coordinate vector representing the ith vertex predicted by the skeleton animation vertex correction model; a loss function for calculating the loss of weight W of the skeleton animation vertex correction model is as follows, wherein loss _ W represents a skin weight loss value, Wi,jAnd representing the action weight of the j joint in the skinning function to the i vertex, wherein the combined loss function is as follows:
Figure 24787DEST_PATH_IMAGE009
and the loss represents the total loss value of the skeleton animation vertex correction model, represents the weight occupied by the loss of the vertex coordinate difference of the 3D mesh model, and finally obtains the trained skeleton animation vertex correction model.
10. A computer device, comprising:
a memory for storing a computer program;
a processor for executing said computer program for implementing the steps of the learning method of the skeletal animated vertex correction model according to any one of claims 1 to 5.
CN202111103180.2A 2021-09-22 2021-09-22 Skeleton animation vertex correction method and model learning method, device and equipment Active CN113554736B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111103180.2A CN113554736B (en) 2021-09-22 2021-09-22 Skeleton animation vertex correction method and model learning method, device and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111103180.2A CN113554736B (en) 2021-09-22 2021-09-22 Skeleton animation vertex correction method and model learning method, device and equipment

Publications (2)

Publication Number Publication Date
CN113554736A true CN113554736A (en) 2021-10-26
CN113554736B CN113554736B (en) 2021-12-21

Family

ID=78106407

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111103180.2A Active CN113554736B (en) 2021-09-22 2021-09-22 Skeleton animation vertex correction method and model learning method, device and equipment

Country Status (1)

Country Link
CN (1) CN113554736B (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114241100A (en) * 2022-02-25 2022-03-25 腾讯科技(深圳)有限公司 Skinning method, skinning device, skinning apparatus, skinning medium, and program product for virtual object
CN114827660A (en) * 2022-03-11 2022-07-29 华数传媒网络有限公司 AI body-building system based on set-top box and implementation method
CN115423907A (en) * 2022-09-29 2022-12-02 魔珐(上海)信息科技有限公司 Role model body binding method, device, equipment and storage medium
CN116486108A (en) * 2023-06-20 2023-07-25 腾讯科技(深圳)有限公司 Image processing method, device, equipment and storage medium
CN116597056A (en) * 2023-07-14 2023-08-15 腾讯科技(深圳)有限公司 Animation processing method, device, equipment and storage medium
CN117557700A (en) * 2024-01-12 2024-02-13 杭州优链时代科技有限公司 Method and equipment for modeling characters
CN117788775A (en) * 2024-02-28 2024-03-29 广州趣丸网络科技有限公司 Digital human palm skeleton posture correction method, device, equipment and medium
CN118470806A (en) * 2024-07-12 2024-08-09 河北省体育科学研究所(河北省体育局反兴奋剂服务中心) Athlete action recognition method based on gesture evaluation

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106780619A (en) * 2016-11-25 2017-05-31 青岛大学 A kind of human body dimension measurement method based on Kinect depth cameras
CN107067451A (en) * 2017-04-07 2017-08-18 广州爱九游信息技术有限公司 The implementation method and device of dynamic bone in animation
CN107924579A (en) * 2015-08-14 2018-04-17 麦特尔有限公司 The method for generating personalization 3D head models or 3D body models
CN111145319A (en) * 2019-11-07 2020-05-12 北京中科深智科技有限公司 Method and system for real-time automatic bone binding and skinning
CN112037312A (en) * 2020-11-04 2020-12-04 成都市谛视科技有限公司 Real-time human body posture inverse kinematics solving method and device
CN112753007A (en) * 2018-07-27 2021-05-04 奇跃公司 Gesture space dimension reduction for gesture space deformation of virtual characters
CN113345062A (en) * 2021-08-04 2021-09-03 成都市谛视科技有限公司 Three-dimensional virtual character action completion method and device and server
CN113345061A (en) * 2021-08-04 2021-09-03 成都市谛视科技有限公司 Training method and device for motion completion model, completion method and device, and medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107924579A (en) * 2015-08-14 2018-04-17 麦特尔有限公司 The method for generating personalization 3D head models or 3D body models
CN106780619A (en) * 2016-11-25 2017-05-31 青岛大学 A kind of human body dimension measurement method based on Kinect depth cameras
CN107067451A (en) * 2017-04-07 2017-08-18 广州爱九游信息技术有限公司 The implementation method and device of dynamic bone in animation
CN112753007A (en) * 2018-07-27 2021-05-04 奇跃公司 Gesture space dimension reduction for gesture space deformation of virtual characters
CN111145319A (en) * 2019-11-07 2020-05-12 北京中科深智科技有限公司 Method and system for real-time automatic bone binding and skinning
CN112037312A (en) * 2020-11-04 2020-12-04 成都市谛视科技有限公司 Real-time human body posture inverse kinematics solving method and device
CN113345062A (en) * 2021-08-04 2021-09-03 成都市谛视科技有限公司 Three-dimensional virtual character action completion method and device and server
CN113345061A (en) * 2021-08-04 2021-09-03 成都市谛视科技有限公司 Training method and device for motion completion model, completion method and device, and medium

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
GILLES, B等: "Creating and Animating Subject-Specific Anatomical Models", 《COMPUTER GRAPHICS FORUM》 *
杨东升: "基于位置动力学的骨骼动画蒙皮技术研究", 《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》 *
陈玉龙: "骨骼蒙皮动画的真实感处理技术研究", 《中国优秀博硕士学位论文全文数据库(硕士)社会科学Ⅱ辑》 *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114241100A (en) * 2022-02-25 2022-03-25 腾讯科技(深圳)有限公司 Skinning method, skinning device, skinning apparatus, skinning medium, and program product for virtual object
CN114827660A (en) * 2022-03-11 2022-07-29 华数传媒网络有限公司 AI body-building system based on set-top box and implementation method
CN115423907A (en) * 2022-09-29 2022-12-02 魔珐(上海)信息科技有限公司 Role model body binding method, device, equipment and storage medium
CN115423907B (en) * 2022-09-29 2023-05-26 魔珐(上海)信息科技有限公司 Character model body binding method, apparatus, device and storage medium
CN116486108B (en) * 2023-06-20 2023-09-01 腾讯科技(深圳)有限公司 Image processing method, device, equipment and storage medium
CN116486108A (en) * 2023-06-20 2023-07-25 腾讯科技(深圳)有限公司 Image processing method, device, equipment and storage medium
CN116597056A (en) * 2023-07-14 2023-08-15 腾讯科技(深圳)有限公司 Animation processing method, device, equipment and storage medium
CN116597056B (en) * 2023-07-14 2023-11-03 腾讯科技(深圳)有限公司 Animation processing method, device, equipment and storage medium
CN117557700A (en) * 2024-01-12 2024-02-13 杭州优链时代科技有限公司 Method and equipment for modeling characters
CN117557700B (en) * 2024-01-12 2024-03-22 杭州优链时代科技有限公司 Method and equipment for modeling characters
CN117788775A (en) * 2024-02-28 2024-03-29 广州趣丸网络科技有限公司 Digital human palm skeleton posture correction method, device, equipment and medium
CN117788775B (en) * 2024-02-28 2024-05-07 广州趣丸网络科技有限公司 Digital human palm skeleton posture correction method, device, equipment and medium
CN118470806A (en) * 2024-07-12 2024-08-09 河北省体育科学研究所(河北省体育局反兴奋剂服务中心) Athlete action recognition method based on gesture evaluation

Also Published As

Publication number Publication date
CN113554736B (en) 2021-12-21

Similar Documents

Publication Publication Date Title
CN113554736B (en) Skeleton animation vertex correction method and model learning method, device and equipment
CN103093490B (en) Based on the real-time face animation method of single video camera
CN110189399B (en) Indoor three-dimensional layout reconstruction method and system
KR101796190B1 (en) Apparatus and method for generating digital clone
CN110599573B (en) Method for realizing real-time human face interactive animation based on monocular camera
CN109829972B (en) Three-dimensional human standard skeleton extraction method for continuous frame point cloud
CN108022278B (en) Character animation drawing method and system based on motion tracking in video
CN113379904B (en) Hidden space motion coding-based multi-person human body model reconstruction method
CN110060329B (en) Mobile terminal human body model reconstruction method based on color depth video stream data
US20240257462A1 (en) Method, apparatus, and storage medium for three-dimensional reconstruction of buildings based on missing point cloud data
CN103854306A (en) High-reality dynamic expression modeling method
CN106408653B (en) Real-time robust cluster adjustment method for large-scale three-dimensional reconstruction
JP2023524252A (en) Generative nonlinear human shape model
CN114581613A (en) Trajectory constraint-based human body model posture and shape optimization method and system
WO2024198747A1 (en) Processing method and apparatus for motion capture data, and device and storage medium
CN112401369A (en) Body parameter measuring method, system, equipment, chip and medium based on human body reconstruction
CN111612887B (en) Human body measuring method and device
CN114550292A (en) High-physical-reality human body motion capture method based on neural motion control
CN115937374B (en) Digital human modeling method, device, equipment and medium
CN105321205B (en) A kind of parameterized human body model method for reconstructing based on sparse key point
CN116797733A (en) Real-time three-dimensional object dynamic reconstruction method
CN115049764B (en) Training method, device, equipment and medium of SMPL parameter prediction model
CN111553954A (en) Direct method monocular SLAM-based online luminosity calibration method
CN116310103A (en) Human body posture estimation and grid recovery method based on skin multi-person linear model
CN114758205B (en) Multi-view feature fusion method and system for 3D human body posture estimation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant