CN114898020A - 3D character real-time face driving method and device, electronic equipment and storage medium - Google Patents

3D character real-time face driving method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN114898020A
CN114898020A CN202210589964.9A CN202210589964A CN114898020A CN 114898020 A CN114898020 A CN 114898020A CN 202210589964 A CN202210589964 A CN 202210589964A CN 114898020 A CN114898020 A CN 114898020A
Authority
CN
China
Prior art keywords
actor
role
character
picture set
facial
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210589964.9A
Other languages
Chinese (zh)
Inventor
邱戴飞
范勇
吴永辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Weiwu Hangzhou Technology Co Ltd
Original Assignee
Weiwu Hangzhou Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Weiwu Hangzhou Technology Co Ltd filed Critical Weiwu Hangzhou Technology Co Ltd
Priority to CN202210589964.9A priority Critical patent/CN114898020A/en
Publication of CN114898020A publication Critical patent/CN114898020A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • G06T13/403D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/002Image coding using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The invention discloses a real-time face driving method and device for a 3D role, electronic equipment and a storage medium, wherein the method comprises the steps of obtaining a role face animation file and a facial expression video file corresponding to an actor; acquiring a controller value corresponding to each frame of animation based on the character facial animation file; rendering a role video file from the role facial animation file, and extracting a corresponding role picture set and an actor facial picture set from the role video file and an actor facial expression video file; constructing a VAE model; training the VAE model based on the role picture set, the actor facial picture set and the controller value corresponding to the role picture set; after the training is finished, inputting the actor face picture into a trained VAE model to obtain the coefficient of a controller; and transmitting the coefficient of the controller to rendering software, and driving the 3D virtual image by the rendering software in real time to obtain high-quality facial animation. The invention can obtain high-precision facial animation through the VAE model.

Description

3D character real-time face driving method and device, electronic equipment and storage medium
Technical Field
The invention relates to the technical field of 3D image processing, in particular to a method and a device for driving a 3D role real-time face, electronic equipment and a storage medium.
Background
With the advent of the meta-universe era, a great deal of requirements for real-time face driving of high-precision 3D face models (3-4 thousands of model vertices) are generated. Current video face driving schemes: the problem of the mobile phone terminal represented by iphoneX is that the expressiveness is not enough, such as the change of the mouth shape of a speaking person, and the mobile phone terminal cannot cross over the uncany valley (terrorist effect) when being used for driving a high-precision real virtual person; the surface capture helmet generally has a complicated 3D role, has a higher use threshold in the calibration stage of an actor, and needs to be recalibrated each time the actor is changed.
Disclosure of Invention
In order to solve the above problems in the prior art, the present invention provides a method and an apparatus for driving a 3D character real-time face, an electronic device, and a storage medium.
The invention discloses a real-time face driving method for a 3D role, which comprises the following steps:
acquiring a 3D character facial animation file and a facial expression video file of a corresponding actor;
acquiring a controller value corresponding to each frame of animation based on the 3D character facial animation file;
rendering the 3D character facial animation file into a 3D character video file, and extracting a corresponding 3D character picture set and an actor facial picture set from the 3D character video file and the actor facial expression video file;
constructing a VAE model;
training the VAE model based on the 3D role picture set, the actor facial picture set and controller values corresponding to the 3D role picture set;
after the training is finished, inputting the actor face picture into the VAE model after the training is finished, and acquiring the coefficient of the controller;
and transmitting the coefficient of the controller to rendering software, wherein the rendering software drives the 3D virtual image in real time to obtain high-quality facial animation.
Preferably, the VAE model includes an encoder and two decoders;
the encoder encodes the input pictures in the 3D character picture set and the actor face picture set;
the decoder performs decoding optimization training on the pictures in the 3D role picture set and the actor face picture set after being encoded to obtain the optimal weight values of the encoder and the decoder;
and the other decoder performs decoding training on the encoded 3D role picture set and the controller value corresponding to the 3D role picture set to obtain the controller coefficient.
Preferably, the encoder encoding function is:
Enc(x)=(f z (x),f id (x));
in the formula: x is an input picture; f. of z (x) Encoding for a VAE; z is the code of expression information in a hidden space; f. of id (x) Encoding for AE; id is the code of the identity information in the hidden space.
Preferably, a loss function of the decoding optimization training is:
L=L vae +L rec +L cycle
Figure BDA0003664337630000028
Figure BDA0003664337630000027
Figure BDA0003664337630000021
in the formula: x is the number of 1 The character is a 3D character picture; x is the number of 2 An actor face picture;
Figure BDA0003664337630000022
is the output of the decoder;
Figure BDA0003664337630000023
is f z (x);p θ (x | z) is the reconstruction of a photograph by z; p (z) is the prior distribution of z; l is dssim Is x and
Figure BDA0003664337630000025
structural similarity error of (2).
Preferably, the error function of the other decoding training is:
Figure BDA0003664337630000024
in the formula: y is the controller value vector corresponding to a frame of animation,
Figure BDA0003664337630000026
is a vector of predicted controller values.
Preferably, the 3D character face animation file is an FBX file, and the controller value corresponding to each frame of animation is read out through FBX sdk of the autodesk.
The present invention further provides a device for driving the real-time face of the 3D character, including:
the acquiring module is used for acquiring a 3D role facial animation file and a facial expression video file corresponding to an actor;
the reading module is used for acquiring a controller value corresponding to each frame of animation based on the 3D role facial animation file;
the extraction module is used for rendering the 3D role facial animation file into a 3D role video file and extracting a corresponding 3D role picture set and an actor facial picture set from the 3D role video file and the actor facial expression video file;
the building module is used for building the VAE model;
a training module for training the VAE model based on the 3D character picture set, the actor face picture set, and a controller value corresponding to the 3D character picture set;
the computing module is used for inputting the actor face picture into the trained VAE model after the training is finished and acquiring the coefficient of the controller;
and the driving module is used for transmitting the coefficient of the controller to rendering software, and the rendering software drives the 3D virtual image in real time to obtain high-quality facial animation.
The invention also provides an electronic device comprising at least one processing unit and at least one memory unit, wherein the memory unit stores a computer program which, when executed by the processing unit, causes the processing unit to perform the above-mentioned method.
The invention also provides a storage medium storing a computer program executable by an electronic device, which when run on the electronic device causes the electronic device to perform the above-mentioned method.
Compared with the prior art, the invention has the beneficial effects that:
according to the invention, high-precision facial animation can be obtained through the VAE model, and the threshold is lower than that of a facial helmet.
Drawings
FIG. 1 is a schematic flow structure diagram of a 3D role real-time face driving method according to the present invention;
FIG. 2 is a schematic structural diagram of a model in the 3D role real-time face driving method of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, are within the scope of the present invention.
The invention is described in further detail below with reference to the attached drawing figures:
referring to fig. 1, the invention discloses a 3D character real-time face driving method, including:
acquiring a 3D character facial animation file and a facial expression video file of a corresponding actor;
acquiring a controller value corresponding to each frame of animation based on the 3D character facial animation file;
specifically, the 3D character face animation file is an FBX file, and the controller name used by the 3D character model and the controller value corresponding to each frame of animation are read from the FBX sdk of the autodesk.
Rendering the 3D role facial animation file into a 3D role video file, and extracting a corresponding 3D role picture set and an actor facial picture set from the 3D role video file and the actor facial expression video file;
specifically, frames are extracted from the 3D character video file and the facial expression video file of the actor according to a certain frame rate to form a picture set, and the picture of each 3D character animation frame corresponds to the value of the controller.
Constructing a VAE model, as shown in FIG. 2;
specifically, the model comprises an encoder and two decoders, wherein the encoder and the decoders are all multilayer convolutional neural networks, and gradient descent is adopted to update the model weight during training. The VAE model comprises an encoder and two decoders; the encoder encodes the input 3D role picture set and pictures in the actor face picture set; a decoder 1 carries out decoding optimization training on the pictures in the 3D role picture set and the actor face picture set after being coded to obtain the optimal weight values of the encoder and the decoder; and fixing the parameters of the encoder, discarding the decoder 1 and the id part, and performing decoding training on the encoded 3D role picture set and the controller value corresponding to the 3D role picture set by the other decoder 2 to obtain a controller coefficient. An encoder encodes an input picture into a hidden space (spatial space), enc (x) ═ f z (x),f id (x) X represents an input picture, and the entire Enc represents the encoder, where f z (x) Z is the typical VAE encoding; f. of id (x) Id is the code of a typical AE. f. of z (x) And f id (x) Sharing the structure of the previous layers of the encoder except for the last layer andand (4) weighting. Wherein z is the code of the expression information in the hidden space, and id is the code of the identity information in the hidden space.
Training the VAE model based on the 3D role picture set, the actor face picture set and the controller value corresponding to the 3D role picture set;
specifically, during training, the 3D character picture and the actor face picture do not need to be in one-to-one correspondence with each frame, the training is divided into two stages, the first stage is self-supervision learning, and the optimal weight values of the encoder and the decoder 1 are learned by reconstructing the input photos. The loss function of the first stage model includes:
L=L vae +L rec +L cycle
L vae ELBO loss for VAE:
Figure BDA0003664337630000051
wherein the content of the first and second substances,
Figure BDA0003664337630000052
i.e. f z (x) Representing an input picture x, and acquiring a VAE code z; p is a radical of θ (x | z) denotes the reconstruction of a picture by z, corresponding to decoder 1; p (z) is the prior distribution of z. Where the first term represents the expected value of the x-likelihood and the second term, called KL divergence, is equivalent to a regularization term, such that
Figure BDA0003664337630000059
Can approximate p (z) as closely as possible.
L rec Error for reconstructed picture and original picture:
Figure BDA0003664337630000053
wherein the content of the first and second substances,
Figure BDA0003664337630000054
is the output of the decoder, L dssim Is x and
Figure BDA0003664337630000055
the first term of the error focuses on the consistency of each pixel, and the second term focuses on the similarity of the structures of the two photographs.
L cycle For cycle consistency error:
Figure BDA0003664337630000056
wherein x is 1 The character is a 3D character picture; x is the number of 2 An actor face picture; the first term represents x 1 The hidden space expression code f of the picture z (x 1 ) And x 2 Identity code f id (x 2 ) Pictures obtained after splicing through the decoder Dec should be summed f after passing through the encoder z (x 1 ) The distance is as close as possible; the second equality is simply from x 2 Obtaining expression code f z (x 2 ),x 1 Obtaining an identity code f id (x 1 )。
The second stage of training is a supervised learning, and the data is 3D character pictures and corresponding controller values. Inputting the 3D role picture into an encoder, inputting the output implicit code z into a decoder 2 to obtain a controller value of model prediction, wherein an error function is
Figure BDA0003664337630000057
Wherein y is a controller value vector corresponding to a frame of animation,
Figure BDA0003664337630000058
is a vector of predicted controller values.
Further, in order to ensure the robustness of the model, the training data, i.e., the 3D character picture and the actor's face picture, are processed, including but not limited to rotation, flipping, cropping, adding noise, changing brightness, chrominance, contrast, saturation, simulated illumination, distortion, etc., so as to improve the robustness of the model.
Namely, training the VAE in the first stage, inputting the actor facial picture and the 3D role picture into the model, and obtaining the optimal weight values of the encoder and the decoder through the optimal loss function. In order not to spill the expression information encoded in z into id, during training, the id obtained for the same actor photo in a batch is randomized before entering the encoder, and the id obtained for the character photo in a batch is randomized before entering the encoder. And the z and the id are spliced by a re-parameterization skill during training.
And the second stage trains the controller decoding module. And fixing parameters of an encoder, discarding a decoder 1 and an id part, training an output z of the encoder connected with a decoder 2, namely performing supervision training on the values of an input 3D role picture and a corresponding controller, wherein the controller parameters are used as output label values. By optimizing L reg A penalty function to obtain optimal weight values for the controller decoder. After the training is completed, the VAE model is used only by the encoder and decoder 2.
After the training is finished, inputting the actor face picture into a trained VAE model to obtain the coefficient of a controller;
and transmitting the coefficient of the controller to rendering software, and driving the 3D virtual image by the rendering software in real time to obtain high-quality facial animation.
Specifically, when the system is used, actor face video stream is obtained from a camera, each frame of picture of the actor face is extracted and input to a 3D role model, the obtained result of a decoder 2 is input to a 3D rendering engine as a controller coefficient, animation is rendered, the performance of the whole process is high, and under the condition that a GPU display card is accelerated, if the frame rate of the camera is high enough, the whole process can reach a frame rate more than 60. Rendering software includes, but is not limited to, UE, maya.
In this embodiment, the facial expression video files of the actors may be the same actor or may be multiple actors.
Examples
The well-pinched realistic faces on Metahuman creator are downloaded from qualel Bridge and imported into Maya and UE as 3D characters.
A performance script is prepared, which includes common expressions, Chinese pronunciations, and Chinese sentences. And the actor A performs according to the performance script, records a front face video with the frame rate of 30 frames, and takes the front face video as a reference template for the performance of the subsequent actor.
Sequence frame animation was performed from actor a performance video K using Metahuman's controller inside Maya for a total of about 8 minutes; exporting the animation into an FBX file for storage, and rendering the animation into a video file in Maya, wherein the video file comprises videos of front faces and side face angles, and the frame rate is 30 frames;
performing by the actor A or other actors B according to the performance script, recording front face and side face videos, wherein the frame rate is 30 frames;
reading the controller coefficient of each frame in the FBX animation file by FBX sdk, dimension 169; extracting photos from the 3D role video and the actor A or other actor B videos, aligning the face of each photo, and cutting out 256x256 pictures;
building a VAE model, wherein an encoder comprises a plurality of layers of rolling blocks, a decoder 1 comprises a plurality of layers of reverse rolling blocks, and a decoder 2 is a plurality of layers of MLPs;
training the VAE model by using the obtained picture, and obtaining the coefficient of the controller after the training is finished;
starting a model client, enabling the actor A or other actors B to face a camera, reading video streams of the camera, inputting each frame of actor picture into a model, and transmitting the coefficient of a controller to UE (user equipment) through levellink; therefore, the Metahuman 3D character in the UE is driven to carry out facial animation in real time, and the effect exceeds the driving effect of IphoneX.
The present invention further provides a device for driving the real-time face of the 3D character, including:
the acquiring module is used for acquiring a 3D role facial animation file and a facial expression video file corresponding to an actor;
the reading module is used for acquiring a controller value corresponding to each frame of animation based on the 3D role facial animation file;
the extraction module is used for rendering the 3D role facial animation file into a 3D role video file and extracting a corresponding 3D role picture set and an actor facial picture set from the 3D role video file and the actor facial expression video file;
the building module is used for building the VAE model;
the training module is used for training the VAE model based on the 3D role picture set, the actor face picture set and the controller value corresponding to the 3D role picture set;
the computing module is used for inputting the actor face picture into the trained VAE model after the training is finished and acquiring the coefficient of the controller;
and the driving module is used for transmitting the coefficient of the controller to rendering software, and the rendering software drives the 3D virtual image in real time to obtain high-quality facial animation.
The invention also provides an electronic device comprising at least one processing unit and at least one memory unit, wherein the memory unit stores a computer program which, when executed by the processing unit, causes the processing unit to perform the above-mentioned method.
The present invention also provides a storage medium storing a computer program executable by an electronic device, which when run on the electronic device causes the electronic device to perform the above-mentioned method.
The above is only a preferred embodiment of the present invention, and is not intended to limit the present invention, and various modifications and changes will occur to those skilled in the art. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (9)

1. A real-time face driving method for a 3D character is characterized by comprising the following steps:
acquiring a 3D character facial animation file and a facial expression video file of a corresponding actor;
acquiring a controller value corresponding to each frame of animation based on the 3D character facial animation file;
rendering the 3D character facial animation file into a 3D character video file, and extracting a corresponding 3D character picture set and an actor facial picture set from the 3D character video file and the actor facial expression video file;
constructing a VAE model;
training the VAE model based on the 3D role picture set, the actor facial picture set and a controller value corresponding to the 3D role picture set;
after the training is finished, inputting the actor face picture into the VAE model after the training is finished, and acquiring the coefficient of the controller;
and transmitting the coefficient of the controller to rendering software, wherein the rendering software drives the 3D virtual image in real time to obtain high-quality facial animation.
2. The 3D character real-time face driving method according to claim 1, wherein the VAE model includes an encoder and two decoders;
the encoder encodes the input pictures in the 3D character picture set and the actor face picture set;
the decoder performs decoding optimization training on the pictures in the 3D role picture set and the actor face picture set after being encoded to obtain the optimal weight values of the encoder and the decoder;
and the other decoder carries out decoding training on the encoded 3D role picture set and the controller value corresponding to the 3D role picture set to obtain the controller coefficient.
3. The method of real-time face driving of a 3D character according to claim 2, wherein the encoder encoding function is:
Enc(x)=(f z (x),f id (x));
in the formula: x is an input picture; f. of z (x) Encoding for a VAE; z is the code of expression information in a hidden space; f. of id (x) Encoding for AE; id is the code of the identity information in the hidden space.
4. The 3D character real-time face driving method according to claim 3, wherein the loss function of the decoding optimization training is:
L=L vae +L rec +L cycle
Figure FDA0003664337620000011
Figure FDA0003664337620000021
Figure FDA0003664337620000022
in the formula: x is the number of 1 The character is a 3D character picture; x is the number of 2 An actor face picture;
Figure FDA0003664337620000023
is the output of the decoder;
Figure FDA0003664337620000026
is f z (x);p θ (x | z) is the reconstruction of a photograph by z; p (z) is the prior distribution of z; l is dssim Is x and
Figure FDA0003664337620000027
structural similarity error of (2).
5. The method of claim 4, wherein the error function of the decoding training is:
Figure FDA0003664337620000024
in the formula: y is the controller value vector corresponding to a frame of animation,
Figure FDA0003664337620000025
is a vector of predicted controller values.
6. The method of claim 1, wherein the 3D character face animation file is an FBX file, and the controller value corresponding to each frame of animation is read out through FBX sdk of autodesk.
7. A 3D character real-time face-driving apparatus, comprising:
the acquiring module is used for acquiring a 3D role facial animation file and a facial expression video file corresponding to an actor;
the reading module is used for acquiring a controller value corresponding to each frame of animation based on the 3D role facial animation file;
the extraction module is used for rendering the 3D role facial animation file into a 3D role video file and extracting a corresponding 3D role picture set and an actor facial picture set from the 3D role video file and the actor facial expression video file;
the building module is used for building the VAE model;
a training module for training the VAE model based on the 3D character picture set, the actor face picture set, and a controller value corresponding to the 3D character picture set;
the computing module is used for inputting the actor face picture into the trained VAE model after the training is finished and acquiring the coefficient of the controller;
and the driving module is used for transmitting the coefficient of the controller to rendering software, and the rendering software drives the 3D virtual image in real time to obtain high-quality facial animation.
8. An electronic device, comprising at least one processing unit and at least one memory unit, wherein the memory unit stores a computer program that, when executed by the processing unit, causes the processing unit to perform the method of any of claims 1 to 6.
9. A storage medium storing a computer program executable by an electronic device, the program, when run on the electronic device, causing the electronic device to perform the method of any one of claims 1 to 6.
CN202210589964.9A 2022-05-26 2022-05-26 3D character real-time face driving method and device, electronic equipment and storage medium Pending CN114898020A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210589964.9A CN114898020A (en) 2022-05-26 2022-05-26 3D character real-time face driving method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210589964.9A CN114898020A (en) 2022-05-26 2022-05-26 3D character real-time face driving method and device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
CN114898020A true CN114898020A (en) 2022-08-12

Family

ID=82725991

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210589964.9A Pending CN114898020A (en) 2022-05-26 2022-05-26 3D character real-time face driving method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN114898020A (en)

Citations (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100141663A1 (en) * 2008-12-04 2010-06-10 Total Immersion Software, Inc. System and methods for dynamically injecting expression information into an animated facial mesh
US9104908B1 (en) * 2012-05-22 2015-08-11 Image Metrics Limited Building systems for adaptive tracking of facial features across individuals and groups
CN106600667A (en) * 2016-12-12 2017-04-26 南京大学 Method for driving face animation with video based on convolution neural network
CN109934767A (en) * 2019-03-06 2019-06-25 中南大学 A kind of human face expression conversion method of identity-based and expressive features conversion
CN111598979A (en) * 2020-04-30 2020-08-28 腾讯科技(深圳)有限公司 Method, device and equipment for generating facial animation of virtual character and storage medium
US20200294294A1 (en) * 2019-03-15 2020-09-17 NeoCortext Inc. Face-swapping apparatus and method
CN112200894A (en) * 2020-12-07 2021-01-08 江苏原力数字科技股份有限公司 Automatic digital human facial expression animation migration method based on deep learning framework
CN112541445A (en) * 2020-12-16 2021-03-23 中国联合网络通信集团有限公司 Facial expression migration method and device, electronic equipment and storage medium
CN112541958A (en) * 2020-12-21 2021-03-23 清华大学 Parametric modeling method and device for three-dimensional face
CN112700524A (en) * 2021-03-25 2021-04-23 江苏原力数字科技股份有限公司 3D character facial expression animation real-time generation method based on deep learning
CN112700523A (en) * 2020-12-31 2021-04-23 魔珐(上海)信息科技有限公司 Virtual object face animation generation method and device, storage medium and terminal
CN113255457A (en) * 2021-04-28 2021-08-13 上海交通大学 Animation character facial expression generation method and system based on facial expression recognition
US20210327038A1 (en) * 2020-04-16 2021-10-21 Disney Enterprises, Inc. Tunable models for changing faces in images
CN113633983A (en) * 2021-08-16 2021-11-12 上海交通大学 Method, device, electronic equipment and medium for controlling expression of virtual character
WO2021232876A1 (en) * 2020-05-18 2021-11-25 北京搜狗科技发展有限公司 Method and apparatus for driving virtual human in real time, and electronic device and medium
CN113781616A (en) * 2021-11-08 2021-12-10 江苏原力数字科技股份有限公司 Facial animation binding acceleration method based on neural network
CN113807265A (en) * 2021-09-18 2021-12-17 山东财经大学 Diversified human face image synthesis method and system
CN113822437A (en) * 2020-06-18 2021-12-21 辉达公司 Deep layered variational automatic encoder
US20220005248A1 (en) * 2020-07-02 2022-01-06 Weta Digital Limited Generating an Animation Rig for Use in Animating a Computer-Generated Character Based on Facial Scans of an Actor and a Muscle Model
CN114494542A (en) * 2022-01-24 2022-05-13 广州喳喳科技有限公司 Character driving animation method and system based on convolutional neural network
CN114531561A (en) * 2022-01-25 2022-05-24 阿里巴巴(中国)有限公司 Face video coding method, decoding method and device

Patent Citations (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100141663A1 (en) * 2008-12-04 2010-06-10 Total Immersion Software, Inc. System and methods for dynamically injecting expression information into an animated facial mesh
US9104908B1 (en) * 2012-05-22 2015-08-11 Image Metrics Limited Building systems for adaptive tracking of facial features across individuals and groups
CN106600667A (en) * 2016-12-12 2017-04-26 南京大学 Method for driving face animation with video based on convolution neural network
CN109934767A (en) * 2019-03-06 2019-06-25 中南大学 A kind of human face expression conversion method of identity-based and expressive features conversion
US20200294294A1 (en) * 2019-03-15 2020-09-17 NeoCortext Inc. Face-swapping apparatus and method
US20210327038A1 (en) * 2020-04-16 2021-10-21 Disney Enterprises, Inc. Tunable models for changing faces in images
CN111598979A (en) * 2020-04-30 2020-08-28 腾讯科技(深圳)有限公司 Method, device and equipment for generating facial animation of virtual character and storage medium
WO2021232876A1 (en) * 2020-05-18 2021-11-25 北京搜狗科技发展有限公司 Method and apparatus for driving virtual human in real time, and electronic device and medium
CN113822437A (en) * 2020-06-18 2021-12-21 辉达公司 Deep layered variational automatic encoder
US20220005248A1 (en) * 2020-07-02 2022-01-06 Weta Digital Limited Generating an Animation Rig for Use in Animating a Computer-Generated Character Based on Facial Scans of an Actor and a Muscle Model
CN112200894A (en) * 2020-12-07 2021-01-08 江苏原力数字科技股份有限公司 Automatic digital human facial expression animation migration method based on deep learning framework
CN112541445A (en) * 2020-12-16 2021-03-23 中国联合网络通信集团有限公司 Facial expression migration method and device, electronic equipment and storage medium
CN112541958A (en) * 2020-12-21 2021-03-23 清华大学 Parametric modeling method and device for three-dimensional face
CN112700523A (en) * 2020-12-31 2021-04-23 魔珐(上海)信息科技有限公司 Virtual object face animation generation method and device, storage medium and terminal
CN112700524A (en) * 2021-03-25 2021-04-23 江苏原力数字科技股份有限公司 3D character facial expression animation real-time generation method based on deep learning
CN113255457A (en) * 2021-04-28 2021-08-13 上海交通大学 Animation character facial expression generation method and system based on facial expression recognition
CN113633983A (en) * 2021-08-16 2021-11-12 上海交通大学 Method, device, electronic equipment and medium for controlling expression of virtual character
CN113807265A (en) * 2021-09-18 2021-12-17 山东财经大学 Diversified human face image synthesis method and system
CN113781616A (en) * 2021-11-08 2021-12-10 江苏原力数字科技股份有限公司 Facial animation binding acceleration method based on neural network
CN114494542A (en) * 2022-01-24 2022-05-13 广州喳喳科技有限公司 Character driving animation method and system based on convolutional neural network
CN114531561A (en) * 2022-01-25 2022-05-24 阿里巴巴(中国)有限公司 Face video coding method, decoding method and device

Non-Patent Citations (7)

* Cited by examiner, † Cited by third party
Title
JIANG Z H, WU Q, CHEN K, ET AL.: "Disentangled representation learning for 3d face shape", CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, 31 December 2019 (2019-12-31), pages 11957 - 11966 *
李嘉豪: "基于深度学习的三维人脸变形研究与应用", 中国优秀硕士学位论文全文数据库, 信息科技辑, no. 2021, 15 December 2021 (2021-12-15), pages 3 *
王浩童: "基于隐向量控制的人脸卡通漫画风格迁移研究", 中国优秀硕士学位论文全文数据库, 信息科技辑, no. 2022, 15 January 2022 (2022-01-15), pages 1 - 78 *
蔡倩倩: "卡通漫画头像的个性化风格迁移算法研究", 中国优秀硕士学位论文全文数据库, 信息科技辑, no. 2021, 15 December 2021 (2021-12-15), pages 1 - 56 *
费建伟, 夏志华, 余佩鹏等: "人脸合成技术综述", 计算机科学与探索, vol. 15, no. 11, 15 July 2021 (2021-07-15), pages 2025 - 2047 *
陈松, 袁训明: "动态人脸表情合成的模型特征驱动算法综述", 计算机与现代化, no. 2019, 4 July 2019 (2019-07-04), pages 47 - 54 *
陈柯宇: "面向数字人应用的人脸表示与动画驱动", 中国优秀硕士学位论文全文数据库, 信息科技辑, no. 2021, 15 August 2021 (2021-08-15), pages 1 - 83 *

Similar Documents

Publication Publication Date Title
CN108596024B (en) Portrait generation method based on face structure information
WO2022267641A1 (en) Image defogging method and system based on cyclic generative adversarial network
CN110751649B (en) Video quality evaluation method and device, electronic equipment and storage medium
CN113261013A (en) System and method for realistic head rotation and facial animation synthesis on mobile devices
CN113901894A (en) Video generation method, device, server and storage medium
CN115914505B (en) Video generation method and system based on voice-driven digital human model
WO2023185395A1 (en) Facial expression capturing method and apparatus, computer device, and storage medium
US11640687B2 (en) Volumetric capture and mesh-tracking based machine learning 4D face/body deformation training
US20220414838A1 (en) Image dehazing method and system based on cyclegan
WO2024098685A1 (en) Face driving method and apparatus for virtual character, and terminal device and readable storage medium
CN114820341A (en) Image blind denoising method and system based on enhanced transform
Elgharib et al. Egocentric videoconferencing
CN114202615A (en) Facial expression reconstruction method, device, equipment and storage medium
CN114187165A (en) Image processing method and device
CN115082300A (en) Training method of image generation model, image generation method and device
Zhang et al. Adaptive affine transformation: A simple and effective operation for spatial misaligned image generation
CN116091955A (en) Segmentation method, segmentation device, segmentation equipment and computer readable storage medium
US20230206955A1 (en) Re-Timing Objects in Video Via Layered Neural Rendering
CN114898020A (en) 3D character real-time face driving method and device, electronic equipment and storage medium
CN117119123A (en) Method and system for generating digital human video based on video material
CN115578298A (en) Depth portrait video synthesis method based on content perception
CN112950501B (en) Noise field-based image noise reduction method, device, equipment and storage medium
CN115496843A (en) Local realistic-writing cartoon style migration system and method based on GAN
Oshiba et al. Face image generation of anime characters using an advanced first order motion model with facial landmarks
Wang Cartoon‐Style Image Rendering Transfer Based on Neural Networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination