CN108363973B - Unconstrained 3D expression migration method - Google Patents

Unconstrained 3D expression migration method Download PDF

Info

Publication number
CN108363973B
CN108363973B CN201810124168.1A CN201810124168A CN108363973B CN 108363973 B CN108363973 B CN 108363973B CN 201810124168 A CN201810124168 A CN 201810124168A CN 108363973 B CN108363973 B CN 108363973B
Authority
CN
China
Prior art keywords
face
expression
model
shape
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810124168.1A
Other languages
Chinese (zh)
Other versions
CN108363973A (en
Inventor
程洪
谢非
郝家胜
赵洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201810124168.1A priority Critical patent/CN108363973B/en
Publication of CN108363973A publication Critical patent/CN108363973A/en
Application granted granted Critical
Publication of CN108363973B publication Critical patent/CN108363973B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2135Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Processing Or Creating Images (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an unconstrained 3D expression migration method, which is realized by using a computer vision and probability statistics based method; firstly, detecting a face region by using the adaboost + harr characteristics, then extracting face geometric characteristics in the face region according to a Constrained Local Model (CLM) method, then extracting face expression parameters by using Support Vector Regression (SVR), finally inputting the expression parameters to control the face BlendShape of the 3D Model, synthesizing expression animation, and realizing the unconstrained 3D expression migration method.

Description

Unconstrained 3D expression migration method
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to an unconstrained 3D expression migration method.
Background
The facial expression has an important role in communication between people, and the facial expression has the advantages of being more intuitive and accurate in expressing human emotion compared with media such as character voice and the like. The emotion interaction mode of people is used in the fields of virtual reality, digital entertainment, communication and video conference, man-machine interaction and the like, and has the advantages of strong expressive force, more natural interaction and the like compared with the traditional interaction modes such as voice, control panels and the like. The expression migration method roughly comprises the following three aspects: capturing the facial expression, extracting facial expression parameters, and synthesizing parameterized target facial animation.
At present, facial expression capturing technologies can be divided into an expression capturing technology without mark points and an expression capturing technology with mark points, the expression capturing technology based on the mark points can capture accurate expression details, but complex hardware equipment is often needed for assistance, a painting brush and the like need to be used for marking the face of a user in the capturing process, and the facial expression capturing technology has certain invasiveness. The unmarked expression capturing technology has less constraint on hardware and a user, and facial expression information is extracted through a 2D facial image.
Most of the expression feature parameterization methods use a machine learning algorithm, a model is trained through a large number of data sets, the mapping relation between captured expression information and expression control parameters is learned, however, the method is limited by individual differences of users, the expression result of a classifier is dependent on training data and the detection capability of the facial expression in a natural state to a great extent, and for the problem, a learner learns the expression features of the current user through an additional initialization step, so that the influence caused by personalized differences is effectively weakened, but the algorithm execution complexity is increased.
The animation synthesis method can be divided into 2D face animation and 3D model face animation according to the type of the target face, the 2D face animation is based on images, and high-reality 2D face animation can be obtained, but in the synthesized animation, the illumination condition of the face and the posture of the face are difficult to change, and seamless splicing into a 3D scene is difficult to achieve. The muscle model driving mode in the 3D model face animation is difficult to obtain control parameters through a computer vision algorithm, the manual control and the manual consumption for establishing the 3D model are very high, the face animation mode of a mixed sample needs to preset an expression library of each basic expression, the expression library needs to meet the orthogonality and the comprehensiveness, and the manual cost for establishing the expression library is very high.
The invention uses a label-free expression capture algorithm, uses a machine learning algorithm to realize automatic parameterized expression feature extraction by using comprehensive training data according to the universality feature of the facial expression, and finally uses an orthogonalized mixed sample to synthesize the facial animation.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides an unconstrained 3D expression migration method, which comprises the steps of extracting user expression parameters according to the facial image characteristics of a user, mapping the user expression parameters to a virtual 3D model, synthesizing facial animation and realizing unconstrained 3D expression migration.
In order to achieve the above object, the present invention provides an unconstrained 3D expression migration method, comprising:
(1) off-line training of the face shape model and obtaining relevant parameters in the model
(1.1) downloading a face image with characteristic point labels from a face database to serve as a face image set;
(1.2) constructing a human face shape model:
Figure BDA0001573010930000021
wherein,
Figure BDA0001573010930000022
representing the average face shape, P is a matrix of principal components of the face shape variation, P ═ P1,P2...Pk]B is the weight vector of the face shape change, B ═ B1,b2.,..,bk]T
(1.3) calculating relevant parameters of the face shape model by using the face image set
Setting M face images in the face image set, wherein each face image has N feature points, and the coordinate of the ith feature point is recorded as (x)i,yi),i=1,2,…,N;
The feature point vector composed by jth face image uses x(j)=[x1 y1 x2 y2...xN yN]TIndicate, average face shape
Figure BDA0001573010930000023
Comprises the following steps:
Figure BDA0001573010930000024
subtracting the average face shape from the feature point vector formed by each face image to obtain a shape change matrix with the average value of 0
Figure BDA0001573010930000025
Figure BDA0001573010930000026
Then, PCA extraction matrix is used for principal component analysis
Figure BDA0001573010930000027
Characteristic vector P ofcAnd a corresponding characteristic value lambdacC is 1,2, …, min (M, N), and then the first k eigenvectors are selected to form a matrix P consisting of main components of the face shape change in a column discharge mode;
(1.4) respectively modeling in a neighborhood range of m multiplied by m by taking the coordinate position of each feature point as a center, then training a scoring model by using a Support Vector Machine (SVM), scoring m multiplied by m points in the neighborhood range by using the trained scoring model to obtain m multiplied by m scoring results, and forming a scoring response graph of each feature point by using the m multiplied by m scoring results;
fitting each scored response graph to a quadratic function r (x, y) ═ a (x-x)0)2+b(y-y0)2+ c, wherein (x)0,y0) The position coordinates of the center point of the response diagram, namely the coordinates of each characteristic point, a, b and c are fitting parameters, and (x, y) the position coordinates of the rest points except the center point in the response diagram;
the position coordinates of the other points are respectively substituted into the quadratic function to obtain the maximum value R (x, y) in the response diagram, and then sigma (R (x, y) -R (x, y))2Solving the values of a, b and c;
(2) acquiring an image containing the face of a user in real time, and marking the image as a source image;
(3) detecting a face region of a user in the source image by using a haar feature + adaboost face detection algorithm;
(4) solved using a training phase
Figure BDA0001573010930000031
P,λjA, b, c, detecting the characteristic points of the current face area
(4.1) according to the human face shape model:
Figure BDA0001573010930000032
and parameters
Figure BDA0001573010930000033
P, using a shape-changing weight vector B ═ B1,b2.,..,bk]TAcquiring all initial feature points in the face region;
(4.2) obtaining a scoring response graph of each initial characteristic point according to the method in the step (1.4);
(4.3) by
Figure BDA0001573010930000034
As an objective function, wherein (x)p,yp) The position coordinate of the p-th point in the scoring response graph is represented, and beta refers to the weight value of the shape constraint; enabling f (x) to converge to the maximum value through iteration, and substituting the weight vector of the corresponding shape change into the face shape model at the moment to acquire all feature points in the face region;
(4.4) taking the updated shape change weight vector B as a new shape model parameter, repeating the steps (4.2) - (4.3) until the updated shape change weight vector B is kept unchanged, and outputting the position of a face feature point corresponding to the final face shape model;
(5) parameterizing the expression characteristics of the user by adopting an Action Unit (AU) detection algorithm with normalized personal characteristics according to the positions of the facial feature points
(5.1) extracting apparent features of the human face
Using a face image set with AU parameter labels in a plurality of databases, scaling the face region scale of each face image to a fixed size, then extracting a gradient histogram of each face image, extracting gradient features in the gradient histogram by using a Principal Component Analysis (PCA), and selecting the first k features as face apparent features;
(5.2) extracting geometric features of human face
Extracting human face feature points from the step (4) as human face geometric features;
(5.3) extracting the expression in the natural state
Calculating the average value of the apparent characteristics and the geometric characteristics in the volume image through the face image in the sequence natural state in the SEMAINE data set to be used as a natural expression descriptor, and then carrying out normalization operation on the natural expression descriptor to generate natural expression characteristics;
(5.4) according to the apparent features, the geometric features and the natural expression features extracted from the face image sets in the multiple databases, training an AU detection model by using a Support Vector Regression (SVR) method of a linear kernel;
(5.5) extracting apparent characteristics, geometric characteristics and natural expression characteristics of the user face area in the source image, inputting the extracted apparent characteristics, geometric characteristics and natural expression characteristics into an AU detection model, and acquiring AU expression parameters of the user face in the current source image;
(6) inputting the AU expression parameters into a 3D model containing a shape difference model (BlendShape), and driving the model expression by using a Unity3D engine to realize the unconstrained 3D model expression migration of the user.
The invention aims to realize the following steps:
the invention relates to an unconstrained 3D expression migration method, which is realized by using a computer vision and probability statistics based method; firstly, detecting a face region by using the adaboost + harr characteristics, then extracting face geometric characteristics in the face region according to a Constrained Local Model (CLM) method, then extracting face expression parameters by using Support Vector Regression (SVR), finally inputting the expression parameters to control the face BlendShape of the 3D Model, synthesizing expression animation, and realizing the unconstrained 3D expression migration method.
Meanwhile, the unconstrained 3D expression migration method provided by the invention also has the following beneficial effects:
(1) the expression migration method applied to 3D movie making at present needs to utilize additional hardware auxiliary equipment to capture expression information, so that the problems of high system complexity, difficulty in popularization and the like are caused, and the method only adopts a common network camera and has good migration performance;
(2) and when the expression parameters of the user are extracted by a plurality of methods, the specific position of the face of the user needs to be marked, which is invasive to the user. (ii) a
(3) The invention has the advantages of low hardware cost, good real-time performance, high corresponding speed, no restriction to users and the like;
drawings
FIG. 1 is a flow chart of an unconstrained 3D expression migration method of the present invention;
FIG. 2 is a flow chart of detecting facial region features;
FIG. 3 is a flow chart of expression parameter extraction;
fig. 4 is a diagram illustrating the expression state of the prefabricated 3D model.
Detailed Description
The following description of the embodiments of the present invention is provided in order to better understand the present invention for those skilled in the art with reference to the accompanying drawings. It is to be expressly noted that in the following description, a detailed description of known functions and designs will be omitted when it may obscure the subject matter of the present invention.
Examples
FIG. 1 is a flow chart of an unconstrained 3D expression migration method according to the present invention.
In this embodiment, as shown in fig. 1, an unconstrained 3D expression migration method according to the present invention includes the following steps:
s1, training the face shape model off line and obtaining relevant parameters in the model
S1.1, training face images of which relevant parameters of a face shape model need to be manually calibrated at key point coordinates, wherein an AFLW database is used for downloading the face images with characteristic point labels from the AFLW database to serve as a face image set;
s1.2, constructing a human face shape model:
Figure BDA0001573010930000051
wherein,
Figure BDA0001573010930000052
representing the average face shape, P is a matrix of principal components of the face shape variation, P ═ P1,P2...Pk]B is the weight vector of the face shape change, B ═ B1,b2.,..,bk]T
S1.3, using the face image set as input, comparing the true value of the characteristic points marked in the image set with the output result of the model, and iteratively updating the value of the face model parameter to be close to the true value:
setting M face images in the face image set, wherein each face image has N feature points, and the coordinate of the ith feature point is recorded as (x)i,yi),i=1,2,…,N;
The feature point vector composed by jth face image uses x(j)=[x1 y1 x2 y2...xN yN]TIndicate, average face shape
Figure BDA0001573010930000061
Comprises the following steps:
Figure BDA0001573010930000062
subtracting the average face shape from the feature point vector formed by each face image to obtain a shape change matrix with the average value of 0
Figure BDA0001573010930000063
Figure BDA0001573010930000064
Then, PCA extraction matrix is used for principal component analysis
Figure BDA0001573010930000065
Characteristic vector P ofcAnd a corresponding characteristic value lambdacC is 1,2, …, min (M, N), and then the first k eigenvectors are selected to form a matrix P consisting of main components of the face shape change in a column discharge mode;
s1.4, respectively modeling in a neighborhood range of m multiplied by m by taking the coordinate position of each feature point as a center, then training a scoring model by using a Support Vector Machine (SVM), scoring m multiplied by m points in the neighborhood range by using the trained scoring model to obtain m multiplied by m scoring results, and forming a scoring response graph of each feature point by using the m multiplied by m scoring results;
fitting each scored response graph to a quadratic function r (x, y) ═ a (x-x)0)2+b(y-y0)2+ c, wherein (x)0,y0) The position coordinates of the center point of the response diagram, namely the coordinates of each characteristic point, a, b and c are fitting parameters, and (x, y) the position coordinates of the rest points except the center point in the response diagram;
the position coordinates of the other points are respectively substituted into the quadratic function to obtain the maximum value R (x, y) in the response diagram, and then sigma (R (x, y) -R (x, y))2Solving the values of a, b and c;
s2, acquiring an image containing a user face in real time by a camera, inputting the image as a data source, and marking the image as a source image;
s3, detecting the face area of the user in the source image by using a haar feature and adaboost face detection algorithm;
s4, using the shape model correlation parameters solved in the training phase:
Figure BDA0001573010930000066
P,λjand a, b and c, inputting the initialized key point position into a model by combining with the image information of the face area, and outputting the model so as to detect the feature point of the current face area, wherein the specific detection process is described below by combining with the figure 2.
According to the constructed human face shape model, a human face shape model can be initialized on a detected human face area, and each point is enabled to search for an optimal matching point in a neighborhood range according to the position of a human face characteristic point calculated by the human face shape model, and the specific steps are as follows:
s4.1, according to the human face shape model:
Figure BDA0001573010930000071
and parameters
Figure BDA0001573010930000072
P, using a shape-changing weight vector B ═ B1,b2.,..,bk]TAcquiring all initial feature points in the face region;
s4.2, obtaining a scoring response graph of each initial characteristic point according to the method in the step S1.4;
s4.3, by
Figure BDA0001573010930000073
As an objective function, wherein (x)p,yp) The position coordinate of the p-th point in the scoring response graph is represented, and beta refers to the weight value of the shape constraint; converging f (x) to a maximum value through iteration, wherein a point corresponding to the maximum value is an optimal matching point searched in a neighborhood range, and substituting a weight vector of the corresponding shape change into the face shape model at the moment to acquire all feature points in the face region;
and S4.4, taking the updated shape change weight vector B as a new shape model parameter, repeating the steps S4.2-S4.3 until the updated shape change weight vector B is basically unchanged, and outputting the position of the face feature point corresponding to the final face shape model.
S5, parameterizing the expression features of the user by using an Action Unit (AU) detection algorithm with normalized personal characteristics according to the positions of the facial feature points, and the specific flowchart is shown in fig. 3.
After all the feature points in the face area of the user are acquired, the feature points need to be further extracted as parametric information under a fixed rule, and then expression animation synthesis can be further performed. The invention uses the commonly used human face expression motion Unit (AU) rule to parameterize the human face expression. The actual effect of AU classifiers is very dependent on the training data due to the large variability between individual users. Aiming at the problem, the invention combines the geometric characteristics and the apparent characteristics of the facial expression of the user and trains an AU classifier under a cross-database, and the specific process is as follows:
s5.1, extracting apparent features of human faces
Using a face image set with AU parameter labels in a plurality of databases, scaling the face region of each face image to 112 × 112, and then extracting a gradient histogram of each face image, specifically, using 2 × 2 unit blocks, each unit block being 8 × 8 pixels, to form 4464-dimensional vector features describing the face. In order to reduce the quantity of HOGs (human body features), extracting gradient feature principal components in a gradient histogram by using a Principal Component Analysis (PCA) method on the basis, namely selecting the first k features as apparent features of a human face; in the embodiment, the data of a plurality of databases are used for extracting apparent features with stronger generalization capability when the PCA performs data dimension reduction;
s5.2, extracting geometric features of human face
Extracting human face feature points from the step S4 to be used as human face geometric features;
s5.3, extracting expression in natural state
Because of individual differences, for example, some faces can present different conditions like smiling or frowning in a relaxed state, and it is very difficult to acquire certain facial expressions under the condition that the natural expressions of the individual are unknown, after the apparent features and the geometric features are acquired, the expressions in the natural state need to be extracted. The method relies on that the human face in the sequence image is basically in a natural state, so that in the embodiment, the average value of the apparent characteristic and the geometric characteristic in the volume image is calculated through the human face image in the sequence natural state in the SEMAINE data set to be used as a natural expression descriptor, and then the natural expression descriptor is subjected to normalization operation to generate a natural expression characteristic;
s5.4, training an AU detection model by using a Support Vector Regression (SVR) method of a linear kernel according to the apparent characteristics, the geometric characteristics and the natural expression characteristics extracted from the face image set in the multiple databases;
s5.5, respectively extracting the apparent characteristics, the geometric characteristics and the natural expression characteristics of the user face area of the source image according to S5.1-S5.3, inputting the extracted apparent characteristics, the geometric characteristics and the natural expression characteristics into an AU detection model, and acquiring AU expression parameters of the user face of the current source image;
s6, the animation synthesis method comprises a muscle model driving mode, a shape difference model driving mode and the like, and as the facial expression is higher in deformation complexity relative to limb actions, a plurality of skeleton nodes need to be arranged in the bone driving mode, the driving links are quite complex, the shape difference model is simple and effective, one shape difference model is shown in figure 4, only partial expression states need to be preset, and the final expression animation is synthesized in a linear fusion mode. Therefore, the method uses AU expression parameters to be input into a 3D model containing a shape difference model (blend shape), and uses a Unity3D engine to drive the model expression, so that the unconstrained 3D model expression migration of the user is realized. .
Although illustrative embodiments of the present invention have been described above to facilitate the understanding of the present invention by those skilled in the art, it should be understood that the present invention is not limited to the scope of the embodiments, and various changes may be made apparent to those skilled in the art as long as they are within the spirit and scope of the present invention as defined and defined by the appended claims, and all matters of the invention which utilize the inventive concepts are protected.

Claims (1)

1. An unconstrained 3D expression migration method is characterized by comprising the following steps:
(1) off-line training a face shape model and obtaining relevant parameters in the model;
(1.1) downloading a face image with characteristic point labels from a face database to serve as a face image set;
(1.2) constructing a human face shape model:
Figure FDA0003441767730000011
wherein,
Figure FDA0003441767730000012
representing the average face shape, P is a matrix of principal components of the face shape variation, P ═ P1,P2...Pk]B is a weight vector of the face shape change,
Figure FDA0003441767730000018
(1.3) calculating relevant parameters of the face shape model by using the face image set;
setting M face images in the face image set, wherein each face image has N feature points, and the coordinate of the ith feature point is recorded as (x)i,yi),i=1,2,…,N;
The feature point vector composed by jth face image uses x(j)=[x1 y1 x2 y2...xN yN]TIndicate, average face shape
Figure FDA0003441767730000013
Comprises the following steps:
Figure FDA0003441767730000014
subtracting the average face shape from the feature point vector formed by each face image to obtain a shape change matrix with the average value of 0
Figure FDA0003441767730000015
Figure FDA0003441767730000016
Then, PCA extraction matrix is used for principal component analysis
Figure FDA0003441767730000017
Characteristic vector P ofcAnd a corresponding characteristic value lambdacC is 1,2, …, min (M, N), and then the first k eigenvectors are selected to form a matrix P consisting of main components of the face shape change in a column discharge mode;
respectively modeling in a neighborhood range of m multiplied by m by taking the coordinate position of each feature point as a center, then training a scoring model by using a Support Vector Machine (SVM), scoring m multiplied by m points in the neighborhood range by using the trained scoring model to obtain m multiplied by m scoring results, and forming a scoring response graph of each feature point by using the m multiplied by m scoring results;
fitting each scored response graph to a quadratic function r (x, y) ═ a (x-x)0)2+b(y-y0)2+ c, wherein (x)0,y0) The position coordinates of the center point of the response diagram, namely the coordinates of each characteristic point, a, b and c are fitting parameters, and (x, y) the position coordinates of the rest points except the center point in the response diagram;
the position coordinates of the other points are respectively substituted into the quadratic function to obtain the maximum value R (x, y) in the response diagram, and then sigma (R (x, y) -R (x, y))2Solving the values of a, b and c;
(2) acquiring an image containing the face of a user in real time, and marking the image as a source image;
(3) detecting a face region of a user in the source image by using a haar feature + adaboost face detection algorithm;
(4) solved using a training phase
Figure FDA0003441767730000021
P,λcA, b and c, detecting the feature points of the current face area;
(4.1) according to the human face shape model:
Figure FDA0003441767730000022
and parameters
Figure FDA0003441767730000023
P, using weight vectors of shape changes
Figure FDA0003441767730000025
Acquiring all initial feature points in the face region;
(4.2) obtaining a scoring response graph of each initial characteristic point according to the method in the step (1.3);
(4.3) by
Figure FDA0003441767730000024
As an objective function, wherein (x)p,yp) The position coordinate of the p-th point in the scoring response graph is represented, and beta refers to the weight value of the shape constraint; enabling f (x) to converge to the maximum value through iteration, and substituting the weight vector of the corresponding shape change into the face shape model at the moment to acquire all feature points in the face region;
(4.4) taking the updated shape change weight vector B as a new shape model parameter, repeating the steps (4.2) - (4.3) until the updated shape change weight vector B is kept unchanged, and outputting the position of a face feature point corresponding to the final face shape model;
(5) adopting an expression parameter extraction algorithm based on an Action Unit (AU) to parameterize the expression characteristics of the user according to the positions of the facial feature points;
(5.1) extracting the apparent features of the human face;
using a face image set with AU parameter labels in a plurality of databases, scaling the face region scale of each face image to a fixed size, then extracting a gradient histogram of each face image, extracting gradient features in the gradient histogram by using a Principal Component Analysis (PCA), and selecting the first k features as face apparent features;
(5.2) extracting geometric features of the human face;
extracting human face feature points from the step (4) as human face geometric features;
(5.3) extracting the expression in a natural state;
calculating the average value of the apparent characteristics and the geometric characteristics in the volume image through the face image in the sequence natural state in the SEMAINE data set to be used as a natural expression descriptor, and then carrying out normalization operation on the natural expression descriptor to generate natural expression characteristics;
(5.4) according to the apparent features, the geometric features and the natural expression features extracted from the face image sets in the multiple databases, training an AU detection model by using a Support Vector Regression (SVR) method of a linear kernel;
(5.5) extracting apparent characteristics, geometric characteristics and natural expression characteristics of the user face area in the source image, inputting the extracted apparent characteristics, geometric characteristics and natural expression characteristics into an AU detection model, and acquiring AU expression parameters of the user face in the current source image;
(6) inputting the AU expression parameters into a 3D model containing a shape difference model (BlendShape), and driving the model expression by using a Unity3D engine to realize the unconstrained 3D model expression migration of the user.
CN201810124168.1A 2018-02-07 2018-02-07 Unconstrained 3D expression migration method Active CN108363973B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810124168.1A CN108363973B (en) 2018-02-07 2018-02-07 Unconstrained 3D expression migration method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810124168.1A CN108363973B (en) 2018-02-07 2018-02-07 Unconstrained 3D expression migration method

Publications (2)

Publication Number Publication Date
CN108363973A CN108363973A (en) 2018-08-03
CN108363973B true CN108363973B (en) 2022-03-25

Family

ID=63005138

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810124168.1A Active CN108363973B (en) 2018-02-07 2018-02-07 Unconstrained 3D expression migration method

Country Status (1)

Country Link
CN (1) CN108363973B (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109493403A (en) * 2018-11-13 2019-03-19 北京中科嘉宁科技有限公司 A method of human face animation is realized based on moving cell Expression Mapping
CN110197167B (en) * 2019-06-05 2021-03-26 清华大学深圳研究生院 Video motion migration method
CN110647636B (en) * 2019-09-05 2021-03-19 深圳追一科技有限公司 Interaction method, interaction device, terminal equipment and storage medium
CN110728193B (en) * 2019-09-16 2022-10-04 连尚(新昌)网络科技有限公司 Method and device for detecting richness characteristics of face image
CN110889894A (en) * 2019-10-25 2020-03-17 中国科学院深圳先进技术研究院 Three-dimensional face reconstruction method and device and terminal equipment
CN111460945A (en) * 2020-03-25 2020-07-28 亿匀智行(深圳)科技有限公司 Algorithm for acquiring 3D expression in RGB video based on artificial intelligence
CN112489689B (en) * 2020-11-30 2024-04-30 东南大学 Cross-database voice emotion recognition method and device based on multi-scale difference countermeasure
CN112541445B (en) * 2020-12-16 2023-07-18 中国联合网络通信集团有限公司 Facial expression migration method and device, electronic equipment and storage medium
CN114677739A (en) * 2022-03-30 2022-06-28 北京字跳网络技术有限公司 Facial expression capturing method and device, computer equipment and storage medium
CN115908655B (en) * 2022-11-10 2023-07-14 北京鲜衣怒马文化传媒有限公司 Virtual character facial expression processing method and device

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104951743A (en) * 2015-03-04 2015-09-30 苏州大学 Active-shape-model-algorithm-based method for analyzing face expression
CN106096557A (en) * 2016-06-15 2016-11-09 浙江大学 A kind of semi-supervised learning facial expression recognizing method based on fuzzy training sample

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104951743A (en) * 2015-03-04 2015-09-30 苏州大学 Active-shape-model-algorithm-based method for analyzing face expression
CN106096557A (en) * 2016-06-15 2016-11-09 浙江大学 A kind of semi-supervised learning facial expression recognizing method based on fuzzy training sample

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
《Markerless 3D Facial Motion Capture System》;Hwang, Y;et al;《ENGINEERING REALITY OF VIRTUAL REALITY 2012》;20121231;第8289卷;全文 *
《Spontaneous facial micro-expression detection based on deep learning》;XiaoHong 等;《IEEE 13th International Conference on Signal Processing (ICSP)》;20161231;第1131-1133页第三节 *
《人脸特征点跟踪方法的研究》;周帆;《中国优秀硕士学位论文全文数据库 信息科技辑》;20170315(第3期);第18页第2.6小节,第21-27页第3.2小节 *
《真实感3D人脸表情合成技术研究进展》;万贤美,等;《计算机辅助设计与图形学学报》;20140228;第26卷(第2期);全文 *

Also Published As

Publication number Publication date
CN108363973A (en) 2018-08-03

Similar Documents

Publication Publication Date Title
CN108363973B (en) Unconstrained 3D expression migration method
Lim et al. Isolated sign language recognition using convolutional neural network hand modelling and hand energy image
Deng et al. MVF-Net: A multi-view fusion network for event-based object classification
Cao et al. Expressive speech-driven facial animation
CN109472198B (en) Gesture robust video smiling face recognition method
Murtaza et al. Analysis of face recognition under varying facial expression: a survey.
CN111028319B (en) Three-dimensional non-photorealistic expression generation method based on facial motion unit
CN112329525A (en) Gesture recognition method and device based on space-time diagram convolutional neural network
CN111680550B (en) Emotion information identification method and device, storage medium and computer equipment
Liang et al. Resolving ambiguous hand pose predictions by exploiting part correlations
CN113255457A (en) Animation character facial expression generation method and system based on facial expression recognition
Jia et al. Two-stream temporal convolutional networks for skeleton-based human action recognition
Mohanty et al. Rasabodha: Understanding Indian classical dance by recognizing emotions using deep learning
Tur et al. Evaluation of hidden markov models using deep cnn features in isolated sign recognition
Wang et al. A comprehensive survey of rgb-based and skeleton-based human action recognition
Xu et al. A high resolution grammatical model for face representation and sketching
Tur et al. Isolated sign recognition with a siamese neural network of RGB and depth streams
Liang et al. Egocentric hand pose estimation and distance recovery in a single RGB image
Zhang et al. Linearly augmented real-time 4D expressional face capture
Kwolek et al. Recognition of JSL fingerspelling using deep convolutional neural networks
Aiman et al. Angle based hand gesture recognition using graph convolutional network
RU2755396C1 (en) Neural network transfer of the facial expression and position of the head using hidden position descriptors
Al Farid et al. Single Shot Detector CNN and Deep Dilated Masks for Vision-Based Hand Gesture Recognition From Video Sequences
Rao et al. Sign Language Recognition using LSTM and Media Pipe
Angelopoulou et al. Evaluation of different chrominance models in the detection and reconstruction of faces and hands using the growing neural gas network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant