CN102254336B - Method and device for synthesizing face video - Google Patents

Method and device for synthesizing face video Download PDF

Info

Publication number
CN102254336B
CN102254336B CN 201110197873 CN201110197873A CN102254336B CN 102254336 B CN102254336 B CN 102254336B CN 201110197873 CN201110197873 CN 201110197873 CN 201110197873 A CN201110197873 A CN 201110197873A CN 102254336 B CN102254336 B CN 102254336B
Authority
CN
China
Prior art keywords
face
people
expression
image
facial image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN 201110197873
Other languages
Chinese (zh)
Other versions
CN102254336A (en
Inventor
刘烨斌
李凯
王好谦
徐枫
戴琼海
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN 201110197873 priority Critical patent/CN102254336B/en
Publication of CN102254336A publication Critical patent/CN102254336A/en
Application granted granted Critical
Publication of CN102254336B publication Critical patent/CN102254336B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

The invention discloses a method and a device for synthesizing a face video. The method for synthesizing the face video comprises the following steps of: establishing a target character expression database comprising a plurality of frames of face images; pre-treating the plurality of frames of face images so as to extract a face posture position and a face expression on each frame of face image; retrieving the joint similarity between the face posture position and the face expression in a user-defined face image sequence and the face posture position and the face expression in the plurality of frames of face images so as to obtain a retrieval image sequence matched with the user-defined face image sequence; and performing winding transform and smoothing treatment on the retrieval image sequence. According to the method and the device for synthesizing the face video disclosed by the embodiment of the invention, for any user-defined expression, a lifelike image sequence relative to a target character can be synthesized conveniently; and the automation degree is high.

Description

People's face image synthesizing method and device
Technical field
The present invention relates to field of Computer Graphics, particularly a kind of people's face image synthesizing method and device.
Background technology
In game making, film making and virtual reality, the analogue technique of human face expression has obtained fast development, has proposed the synthetic method of many expressions, yet synthesizing of true personage's expression still can not be satisfied actual demand.On the one hand, some lack the sense of reality based on the expression picture that the picture deformation technology of sample produces, and can not satisfy the requirement of a large amount of expression sequence senses of reality; On the other hand, the facial expression capturing technology can be transferred to another personage with performer's expression, this technology is comparatively ripe, in many film special efficacys (such as 3D film " A Fanda "), application is arranged, but implement comparatively complicated, and the performer need to wear a helmet, and is very inconvenient; Simultaneously, the performer also often need to repeat to perform certain expression until satisfy quality requirements, and this bothers for the performer very much.
Want personage's expression synthetic true to nature, main difficult point is the geometry of people's face and synthesizing of textural characteristics, and when the people did certain expression, his profile changed according to muscle, and the while is affected by ambient lighting also and produces bright alternately dark.Wherein, encoding facial movement system (FACS) is used widely in expression is analyzed and be synthetic, it is divided into a series of moving cells (AU) with people's face, the respective muscle motion of some basic facial expressions is provided, by changing the correlation combiner of these moving cells, just can produce different human face expressions.Although this method can access comparatively real human face expression, can not realize the robotization processing, namely a given personage's expression can not arrive another personage with Expression Mapping under the prerequisite that reduces as far as possible people's intervention.
Summary of the invention
Purpose of the present invention is intended to solve at least one of above-mentioned technological deficiency.For this reason, the present invention need to provide a kind of people's face image synthesizing method and device, and the advantage of this people's face image synthesizing method and device is: for user-defined any expression, can both synthesize easily the corresponding photorealism sequence of target person; Reduce the actor number of times; And automaticity is high.
According to an aspect of the present invention, provide a kind of people's face image synthesizing method, it is characterized in that, comprised step: set up target person expression database, comprise the multiframe facial image in the described target person expression database; Described multiframe facial image is carried out pre-service, with the posture position that extracts people's face on each frame facial image and the expression of people's face; Carry out the associating similarity retrieval of the expression of the posture position of people's face on the expression of the posture position of people's face in the user-defined human face image sequence and people's face and the described multiframe facial image and people's face, to obtain the retrieving images sequence with described user-defined human face image sequence coupling; And to described retrieving images sequence reel conversion and smoothing processing.
According to people's face image synthesizing method of the embodiment of the invention, for user-defined any expression, can both synthesize easily the corresponding photorealism sequence of target person.
According to one embodiment of present invention, the described step of setting up target person expression database comprises: gather a plurality of basic facial expression image sequences of described target person from a plurality of visual angles and a plurality of posture position, each the basic facial expression image sequence in wherein said a plurality of basic facial expression image sequences is expressed one's feelings by the change procedure of neutrality expression and described neutral expression and is consisted of.
According to people's face image synthesizing method of the embodiment of the invention, can comprise great amount of images in the target person expression database.
According to one embodiment of present invention, describedly described multiframe facial image is carried out pretreated step comprise: extract in the described multiframe facial image posture position of people's face in each frame facial image and describe the position of people's face with crab angle, helix angle and side rake angle; People's face in each frame facial image in the described multiframe facial image aimed at positive standard attitude and carry out feature and describe point and detect, describe the expression of people's face with described feature description.
According to people's face image synthesizing method of the embodiment of the invention, posture position and expression that can accurate description people face.
According to one embodiment of present invention, carry out described feature based on active shape model and describe the some detection.
The step of associating similarity retrieval of according to one embodiment of present invention, carrying out the expression of the posture position of people's face on the expression of the posture position of people's face in the described user-defined human face image sequence and people's face and the described multiframe facial image and people's face comprises: according to D Pose=L (| Y i-Y j|)+L (| R i-R j|)+L (| P i-P j|) similarity of carrying out the posture position of people's face on the posture position of people's face in the described user-defined human face image sequence and the described multiframe facial image calculates, wherein Y represents crab angle, and R represents helix angle, and P represents side rake angle, I iRepresent the arbitrary frame facial image in the described user-defined human face image sequence, I jRepresent the arbitrary frame facial image in the described multiframe facial image, L (d) is the sigmoid function, is defined as
Figure BDA0000075898320000021
γ=ln99 wherein, T and σ are respectively average statistical and the standard deviations of variable d; According to described crab angle, described helix angle and described side rake angle, the posture position of people's face in the described user-defined human face image sequence is registered to positive attitude, then basis
Figure BDA0000075898320000022
Carry out the similarity of the expression of people's face on the expression of people's face in the described user-defined human face image sequence and the described multiframe facial image and calculate, wherein A I, kIt is image I iK feature behind the aligning described some A J, kIt is image I jK feature behind the aligning described a little, and J is the number of unique point, w kBe the weight of k unique point, L (d) is the sigmoid function, is defined as
Figure BDA0000075898320000023
The unique point range normalization that calculates is arrived [0,1]; According to D (i, j)=D Pose(i, j)+λ D Expression(i, j) obtains the associating similarity, and wherein parameter lambda is to regulate the weight proportion of expression similarity and posture position similarity; And obtain image I in the user-defined human face image sequence according to described associating similarity iA plurality of candidate images in described multiframe facial image.
According to people's face image synthesizing method of the embodiment of the invention, can obtain posture position and the suitable associating similarity between the expression of people's face and obtain qualified a plurality of candidate image.
According to one embodiment of present invention, described a plurality of candidate images for described associating similarity retrieval acquisition, further calculate the associating similarity between described a plurality of candidate images, and use the Dijkstra shortest path first to obtain described retrieving images sequence to guarantee time continuity and Space Consistency.
According to people's face image synthesizing method of the embodiment of the invention, can guarantee accuracy and the flatness of retrieving images sequence.
According to an aspect of the present invention, provide a kind of people's face video synthesizer, it is characterized in that, having comprised: target person expression database, described target person expression database is used for storage multiframe facial image; Pretreatment module, described pretreatment module are used for extracting the posture position of people's face on described each frame facial image of target person expression database and the expression of people's face; Associating similarity retrieval module, the associating similarity retrieval of the expression of the posture position of people's face and people's face on the expression that described associating similarity retrieval module is used for carrying out the posture position of user-defined human face image sequence people face and people's face and the described multiframe facial image is with the retrieving images sequence that obtains mating with described user-defined human face image sequence; And post-processing module, described post-processing module is for described retrieving images sequence is reeled conversion and smoothing processing.
According to people's face video synthesizer of the embodiment of the invention, for user-defined any expression, can both synthesize easily the corresponding photorealism sequence of target person.
According to one embodiment of present invention, described multiframe facial image comprises a plurality of basic facial expression image sequences that gather described target person from a plurality of visual angles and a plurality of posture position, and each the basic facial expression image sequence in wherein said a plurality of basic facial expression image sequences is expressed one's feelings by the change procedure of neutrality expression and described neutral expression and consisted of.
According to people's face video synthesizer of the embodiment of the invention, can comprise great amount of images in the target person expression database.
According to one embodiment of present invention, described pretreatment module is further used for: extract in the described multiframe facial image posture position of people's face in each frame facial image and describe the position of people's face with crab angle, helix angle and side rake angle; An and expression of the people's face in each frame facial image in the described multiframe facial image being aimed at and described with described feature description point people's face with positive standard attitude.
According to people's face video synthesizer of the embodiment of the invention, posture position and expression that can accurate description people face.
According to one embodiment of present invention, described pretreatment module is extracted and is carried out described feature based on active shape model and describe point and detect.
According to one embodiment of present invention, described associating similarity retrieval module is further used for: according to D Pose=L (| Y i-Y j|)+L (| R i-R j|)+L (| P i-P j|) similarity of carrying out the posture position of people's face on the posture position of people's face in the described user-defined human face image sequence and the described multiframe facial image calculates, wherein Y represents crab angle, and R represents helix angle, and P represents side rake angle, I iRepresent the arbitrary frame facial image in the described user-defined human face image sequence, I jRepresent the arbitrary frame facial image in the described multiframe facial image, L (d) is the sigmoid function, is defined as
Figure BDA0000075898320000041
γ=ln99 wherein, T and σ are respectively average statistical and the standard deviations of variable d; According to described crab angle, described helix angle and described side rake angle, the posture position of people's face in the described user-defined human face image sequence is registered to positive attitude, then basis
Figure BDA0000075898320000042
Carry out the similarity of the expression of people's face on the expression of people's face in the described user-defined human face image sequence and the described multiframe facial image and calculate, wherein A I, kIt is image I iK feature behind the aligning described some A J, kIt is image I jK feature behind the aligning described a little, and J is the number of unique point, w kBe the weight of k unique point, L (d) is the sigmoid function, is defined as The unique point range normalization that calculates is arrived [0,1]; According to D (i, j)=D Pose(i, j)+λ D Expression(i, j) obtains the associating similarity, and wherein parameter lambda is to regulate the weight proportion of expression similarity and posture position similarity; And obtain image I in the user-defined human face image sequence according to described associating similarity iA plurality of candidate images in described multiframe facial image.
According to people's face video synthesizer of the embodiment of the invention, can obtain posture position and the suitable associating similarity between the expression of people's face and obtain qualified a plurality of candidate image.
According to one embodiment of present invention, for described a plurality of candidate images, described associating similarity retrieval module further uses the Dijkstra shortest path first to obtain described retrieving images sequence to guarantee time continuity and Space Consistency according to the associating similarity between described a plurality of candidate images.
According to people's face video synthesizer of the embodiment of the invention, can guarantee accuracy and the flatness of retrieving images sequence.
Additional aspect of the present invention and advantage in the following description part provide, and part will become obviously from the following description, or recognize by practice of the present invention.
Description of drawings
Above-mentioned and/or the additional aspect of the present invention and advantage are from obviously and easily understanding becoming the description of embodiment below in conjunction with accompanying drawing, wherein:
Fig. 1 is the synoptic diagram of people's face image synthesizing method according to an embodiment of the invention;
Fig. 2 is the process flow diagram of people's face image synthesizing method according to an embodiment of the invention;
Fig. 3 is the process flow diagram of uniting according to an embodiment of the invention the method for similarity retrieval;
Fig. 4 extracts the synoptic diagram that face characteristic is described point based on active shape model (ASM) according to an embodiment of the invention;
Fig. 5 is the synoptic diagram that obtains according to an embodiment of the invention the retrieving images sequence; And
Fig. 6 is the block diagram of people's face video synthesizer according to an embodiment of the invention.
Embodiment
The below describes embodiments of the invention in detail, and the example of described embodiment is shown in the drawings, and wherein identical or similar label represents identical or similar element or the element with identical or similar functions from start to finish.Be exemplary below by the embodiment that is described with reference to the drawings, only be used for explaining the present invention, and can not be interpreted as limitation of the present invention.
Need to prove that in addition, term " first ", " second ", " the 3rd " only are used for describing purpose, and can not be interpreted as indication or hint relative importance or the implicit quantity that indicates indicated technical characterictic.Thus, one or more these features can be expressed or impliedly be comprised to the feature that is limited with " first ", " second ", " the 3rd ".Further, in description of the invention, except as otherwise noted, the implication of " a plurality of " is two or more.
Below with reference to accompanying drawing specific embodiments of the invention are described.
Fig. 1 is the synoptic diagram of people's face image synthesizing method according to an embodiment of the invention.As shown in Figure 1, comprise the multiframe facial image at target person expression database.Expression according to posture position and people's face of people's face in the user-defined human face image sequence, searched targets personage the express one's feelings posture position of people's face of multiframe facial image and the expression of people's face in the database, to retrieve the image sequence of coupling, afterwards to image sequence reel with smoothly.
Fig. 2 is the process flow diagram of people's face image synthesizing method according to an embodiment of the invention.As shown in Figure 2, people's face image synthesizing method may further comprise the steps.
Step S201 sets up target person expression database, comprises the multiframe facial image in the target person expression database.Particularly, target person is made a plurality of basic facial expression image sequences under the various visual angles acquisition system, and target person repeats to do for several times for each basic facial expression in the situation of the posture position that changes head.Gather as far as possible the expression under a plurality of visual angles and a plurality of posture position.Each basic facial expression image sequence is made of the change procedure expression of neutrality expression and neutral expression.Basic facial expression comprises happiness, anger, sadness, detest, fear, surprised etc.
Step S202 carries out pre-service to the multiframe facial image, with the posture position that extracts people's face on each frame facial image and the expression of people's face.Specifically comprise: extract in the multiframe facial image posture position of people's face in each frame facial image and describe the position of people's face with crab angle, helix angle and side rake angle; People's face in each frame facial image in the multiframe facial image aimed at positive standard attitude and carry out feature and describe point and detect, use characteristic is described an expression of describing people's face.
Step S203, carry out the associating similarity retrieval of the expression of the posture position of people's face on the expression of the posture position of people's face in the user-defined human face image sequence and people's face and the multiframe facial image and people's face, to obtain the retrieving images sequence with user-defined human face image sequence coupling.The concrete steps of associating similarity retrieval are described with reference to figure 3.Fig. 3 is the process flow diagram of uniting according to an embodiment of the invention the method for similarity retrieval.As shown in Figure 3, the associating similarity retrieval comprises the steps.
Step S2031 is according to D Pose=L (| Y i-Y j|)+L (| R i-R j|)+L (| P i-P j|) similarity of carrying out the posture position of people's face on the posture position of people's face in the user-defined human face image sequence and the multiframe facial image calculates, wherein Y represents crab angle, and R represents helix angle, and P represents side rake angle, I iRepresent the arbitrary frame facial image in the user-defined human face image sequence, I jArbitrary frame facial image in the expression multiframe facial image, L (d) is the sigmoid function, is defined as
Figure BDA0000075898320000061
γ=ln99 wherein, T and σ are respectively average statistical and the standard deviations of variable d.
Step S2032 according to described crab angle, described helix angle and described side rake angle, is registered to positive attitude, then basis with the posture position of people's face in the described user-defined human face image sequence
Figure BDA0000075898320000062
Carry out the similarity of the expression of people's face on the expression of people's face in the user-defined human face image sequence and the multiframe facial image and calculate, wherein A I, kIt is image I iK feature behind the aligning described some A J, kIt is image I jK feature behind the aligning described a little, and J is the number of unique point, w kBe the weight of k unique point, L (d) is the sigmoid function, is defined as
Figure BDA0000075898320000063
The unique point range normalization that calculates is arrived [0,1].The extraction of point is wherein described referring to Fig. 4 Expressive Features.
Fig. 4 extracts the synoptic diagram that face characteristic is described point based on active shape model (ASM) according to an embodiment of the invention.As shown in Figure 4, target person is expressed one's feelings face alignment on each frame facial image in the database to positive standard attitude, then carry out the active shape model feature and describe the some detection, detect a plurality of features to describe a little, these a plurality of features are described the expression that point defines people's face.
Step S2033 is according to D (i, j)=D Pose(i, j)+λ D Expression(i, j) obtains the associating similarity of posture position and expression, and wherein parameter lambda is to regulate the weight proportion of expression similarity and posture position similarity.
Step S2034, according to the associating similarity that obtains among the step S2033, for user-defined human face image sequence, retrieve one by one the nearest K frame of target person expression database middle distance image, the K frame of namely uniting the similarity minimum, namely each user images constantly has K candidate image.
Step S2035 for the K that obtains among a step S2034 candidate image, further according to the associating similarity between K the candidate image, uses the Dijkstra shortest path first to obtain the retrieving images sequence to guarantee time continuity and Space Consistency.Wherein, use with above-mentioned steps in associating similarity between the identical method calculating K candidate image.Particularly, utilize the Dijkstra shortest path first to retrieve one and be carved into t=m shortest path constantly during from t=1, wherein m is the length of user-defined human face image sequence.Fig. 5 is the synoptic diagram that obtains according to an embodiment of the invention the retrieving images sequence.Before implementing the Dijkstra shortest path first, need to set up a digraph, as shown in Figure 5, the node of this digraph is by the candidate face image construction of all target persons that retrieve, only allow to have between the candidate face image of adjacent moment directed edge to link to each other, and the length on limit (or being called cost) is defined as:
L(C t,i,C t+1,j)=D(C t,i,U t)+D(C t+1,j,U t+1)+μD(C t,i,C t+1,j),
Wherein, C T, iWith C T+1, jRepresent respectively t i candidate image and j the candidate image in the t+1 moment constantly, U tWith U T+1T and t+1 facial image of the definition of difference representative of consumer.The tolerance of this costs between nodes is so that select as far as possible the candidate of rank forward (near as far as possible with the distance of image in the user-defined human face image sequence), again can be so that not significantly sudden change between different constantly candidate image, μ regulates the accuracy of the human face image sequence that retrieves and the parameter of flatness.
Utilize shortest path first, draw node C 1, iWith C M, jBetween shortest path, can find out that from the permutation and combination of first and last node always total K*K in such path is at this K 2Obtain again that of path minimum in the individual shortest path, just obtained the retrieving images sequence that needs.
Step S204 is to retrieving images sequence reel conversion and smoothing processing.Although the retrieving images sequence that retrieval obtains among the step S203 has kept time continuity and Space Consistency to a certain extent, also have certain difference but compare with the expression of people's face on the target person image, the conversion of therefore need to reeling comes so that the expression of the people's face expression of people's face on the image of the human face image sequence of match user definition more in the retrieving images sequence.Simultaneously, the retrieving images sequence may exist video jitter to jump, and therefore needs further smoothing processing.The concrete steps of smoothing processing are as follows.
For t user images constantly, utilize the feature of extracting to describe the people's face of naming a person for a particular job and be divided into the parts such as eyebrow, eyes, nose, face, calculate respectively U tAnd C tThese provincial characteristicss are described the distance of point, if this distance, thinks then that this regional aim personage's action and user's action have certain gap greater than a certain constant δ.For example the action of face not too meets the requirements, and then needs to be used in the target person expression database candidate image that again retrieval and user's face mate most, then the face that mates most that detects is substituted original C tFace, can use to mend based on people's face of Poisson equation and paint algorithm, thereby when guaranteeing to substitute and the consistance of peripheral region texture, be unlikely to produce obvious feeling of unreality.
On time domain, may there be shake through people's face of mending after painting, needs the distance of adjacent moment between the image after the calculating benefit is painted, if detect shake, take the method for light stream optimization to realize seamlessly transitting, substitute shake constantly some image of front and back occurs.
Fig. 6 is the block diagram of people's face video synthesizer according to an embodiment of the invention.As shown in Figure 6, people's face video synthesizer 10 comprises target person expression database 110, pretreatment module 120, associating similarity retrieval module 130 and post-processing module 140.
Particularly, target person expression database 110 is used for storage multiframe facial image.Pretreatment module 120 is used for extracting the posture position of people's face on target person expression database 110 each frame facial image and the expression of people's face.The associating similarity retrieval of the expression of the posture position of people's face and people's face on the expression that associating similarity retrieval module 130 is used for carrying out the posture position of user-defined human face image sequence people face and people's face and the multiframe facial image is with the retrieving images sequence that obtains mating with user-defined human face image sequence.Post-processing module 140 is for the retrieving images sequence is reeled conversion and smoothing processing.
In one embodiment of the invention, the multiframe facial image comprises that from a plurality of basic facial expression image sequences of a plurality of visual angles and a plurality of posture position collection target person wherein each the basic facial expression image sequence in a plurality of basic facial expression image sequences is expressed one's feelings by the change procedure of neutrality expression and neutral expression and consisted of.Describe a little based on the active shape model detected characteristics.
In one embodiment of the invention, pretreatment module 120 is further used for: extract in the multiframe facial image posture position of people's face in each frame facial image and describe the position of people's face with crab angle, helix angle and side rake angle; And the people's face in each frame facial image in the multiframe facial image is aimed at also use characteristic description put an expression of describing people's face with positive standard attitude.
In one embodiment of the invention, associating similarity retrieval module 130 is further used for: according to D Pose=L (| Y i-Y j|)+L (| R i-R j|)+L (| P i-P j|) similarity of carrying out the posture position of people's face on the posture position of people's face in the user-defined human face image sequence and the multiframe facial image calculates, wherein Y represents crab angle, and R represents helix angle, and P represents side rake angle, I iRepresent the arbitrary frame facial image in the user-defined human face image sequence, I jArbitrary frame facial image in the expression multiframe facial image, L (d) is the sigmoid function, is defined as
Figure BDA0000075898320000081
γ=ln99 wherein, T and σ are respectively average statistical and the standard deviations of variable d; According to described crab angle, described helix angle and described side rake angle, the posture position of people's face in the described user-defined human face image sequence is registered to positive attitude, then basis
Figure BDA0000075898320000082
Carry out the similarity of the expression of people's face on the expression of people's face in the user-defined human face image sequence and the multiframe facial image and calculate, wherein A I, kIt is image I iK feature behind the aligning described some A J, kIt is image I jK feature behind the aligning described a little, and J is the number of unique point, w kBe the weight of k unique point, L (d) is the sigmoid function, is defined as
Figure BDA0000075898320000083
The unique point range normalization that calculates is arrived [0,1]; According to D (i, j)=D Pose(i, j)+λ D Expression(i, j) obtains the associating similarity, and wherein parameter lambda is to regulate the weight proportion of expression similarity and posture position similarity; And according to the associating similarity obtain image I in the user-defined human face image sequence iA plurality of candidate images in the multiframe facial image.For a plurality of candidate images, associating similarity retrieval module 130 further uses the Dijkstra shortest path first to obtain the retrieving images sequence to guarantee time continuity and Space Consistency according to the associating similarity between a plurality of candidate images.
People's face image synthesizing method and device according to the embodiment of the invention have following advantage: for user-defined any expression, can both synthesize easily the corresponding photorealism sequence of target person; Reduce the actor number of times; And automaticity is high.
In the description of this instructions, the description of reference term " embodiment ", " some embodiment ", " example ", " concrete example " or " some examples " etc. means to be contained at least one embodiment of the present invention or the example in conjunction with specific features, structure, material or the characteristics of this embodiment or example description.In this manual, the schematic statement of above-mentioned term not necessarily referred to identical embodiment or example.And the specific features of description, structure, material or characteristics can be with suitable mode combinations in any one or more embodiment or example.
Although illustrated and described embodiments of the invention, those having ordinary skill in the art will appreciate that: can carry out multiple variation, modification, replacement and modification to these embodiment in the situation that does not break away from principle of the present invention and aim, scope of the present invention is limited by claim and equivalent thereof.
Although illustrated and described embodiments of the invention, for the ordinary skill in the art, be appreciated that without departing from the principles and spirit of the present invention and can carry out multiple variation, modification, replacement and modification to these embodiment that scope of the present invention is by claims and be equal to and limit.

Claims (8)

1. people's face image synthesizing method is characterized in that, comprises step:
Set up target person expression database, comprise the multiframe facial image in the described target person expression database;
Extract in the described multiframe facial image posture position of people's face in each frame facial image and describe the position of people's face with crab angle, helix angle and side rake angle;
People's face in each frame facial image in the described multiframe facial image aimed at positive standard attitude and carry out feature and describe point and detect, describe the expression of people's face with described feature description;
According to D Pose=L (| Y i-Y j|)+L (| R i-R j|)+L (| P i-P j|) similarity of carrying out the posture position of people's face on the posture position of people's face in the user-defined human face image sequence and the described multiframe facial image calculates, wherein Y represents crab angle, and R represents helix angle, and P represents side rake angle, I iRepresent the arbitrary frame facial image in the described user-defined human face image sequence, I jRepresenting the arbitrary frame facial image in the described multiframe facial image, is the sigmoid function L(d), is defined as
Figure FDA00002098841600011
γ=ln99 wherein, T and σ are respectively average statistical and the standard deviations of variable d;
According to described crab angle, described helix angle and described side rake angle, the posture position of people's face in the described user-defined human face image sequence is registered to positive attitude, then basis Carry out the similarity of the expression of people's face on the expression of people's face in the described user-defined human face image sequence and the described multiframe facial image and calculate, wherein A I, kIt is image I iK feature behind the aligning described some A J, kIt is image I jK feature behind the aligning described a little, and J is the number of unique point, w kBe the weight of k unique point, L (d) is the sigmoid function, is defined as
Figure FDA00002098841600013
The unique point range normalization that calculates is arrived [0,1];
According to D (i, j)=D Pose(i, j)+λ D Expression(i, j) obtains the associating similarity, and wherein parameter lambda is to regulate the weight proportion of expression similarity and posture position similarity;
Obtain image I in the user-defined human face image sequence according to described associating similarity iA plurality of candidate images in described multiframe facial image, thus retrieving images sequence with described user-defined human face image sequence coupling obtained; And
To described retrieving images sequence reel conversion and smoothing processing.
2. people's face image synthesizing method according to claim 1 is characterized in that, the described step of setting up target person expression database comprises:
Gather a plurality of basic facial expression image sequences of described target person from a plurality of visual angles and a plurality of posture position, each the basic facial expression image sequence in wherein said a plurality of basic facial expression image sequences is expressed one's feelings by the change procedure of neutrality expression and described neutral expression and is consisted of.
3. people's face image synthesizing method according to claim 1 is characterized in that, extracts described feature based on active shape model and describes a little.
4. people's face image synthesizing method according to claim 1, it is characterized in that, described a plurality of candidate images for described associating similarity retrieval acquisition, further calculate the associating similarity between described a plurality of candidate images, and use the Dijkstra shortest path first to obtain described retrieving images sequence to guarantee time continuity and Space Consistency.
5. people's face video synthesizer is characterized in that, comprising:
Target person expression database, described target person expression database is used for storage multiframe facial image;
Pretreatment module, the position that described pretreatment module is used for extracting the posture position of people's face in each frame facial image of described multiframe facial image and describes people's face with crab angle, helix angle and side rake angle, and the people's face in each frame facial image in the described multiframe facial image aimed at positive standard attitude and carry out feature and describe point and detect, the expression of people's face is described with described feature description;
Associating similarity retrieval module, described associating similarity retrieval module is used for carrying out the associating similarity retrieval of the expression of the posture position of people's face on the expression of the posture position of user-defined human face image sequence people face and people's face and the described multiframe facial image and people's face, with the retrieving images sequence that obtains mating with described user-defined human face image sequence, wherein said associating similarity retrieval comprises according to D Pose=L (| Y i-Y j|)+L (| R i-R j|)+L (| P i-P j|) similarity of carrying out the posture position of people's face on the posture position of people's face in the described user-defined human face image sequence and the described multiframe facial image calculates, wherein Y represents crab angle, and R represents helix angle, and P represents side rake angle, I iRepresent the arbitrary frame facial image in the described user-defined human face image sequence, I jRepresenting the arbitrary frame facial image in the described multiframe facial image, is the sigmoid function L(d), is defined as
Figure FDA00002098841600021
γ=ln99 wherein, T and σ are respectively average statistical and the standard deviations of variable d; According to described crab angle, described helix angle and described side rake angle, the posture position of people's face in the described user-defined human face image sequence is registered to positive attitude, then basis
Figure FDA00002098841600022
Carry out the similarity of the expression of people's face on the expression of people's face in the described user-defined human face image sequence and the described multiframe facial image and calculate, wherein A I, kIt is image I iK feature behind the aligning described some A J, kIt is image I jK feature behind the aligning described a little, and J is the number of unique point, w kBe the weight of k unique point, L (d) is the sigmoid function, is defined as
Figure FDA00002098841600023
The unique point range normalization that calculates is arrived [0,1]; According to D (i, j)=D Pose(i, j)+λ D Expression(i, j) obtains the associating similarity, and wherein parameter lambda is to regulate the weight proportion of expression similarity and posture position similarity; With obtain image I in the user-defined human face image sequence according to described associating similarity iA plurality of candidate images in described multiframe facial image; And
Post-processing module, described post-processing module is for described retrieving images sequence is reeled conversion and smoothing processing.
6. people's face video synthesizer according to claim 5, it is characterized in that, described multiframe facial image comprises a plurality of basic facial expression image sequences that gather described target person from a plurality of visual angles and a plurality of posture position, and each the basic facial expression image sequence in wherein said a plurality of basic facial expression image sequences is expressed one's feelings by the change procedure of neutrality expression and described neutral expression and consisted of.
7. people's face video synthesizer according to claim 5 is characterized in that, described pretreatment module is extracted based on active shape model and carried out the detection of described feature description point.
8. people's face video synthesizer according to claim 5, it is characterized in that, for described a plurality of candidate images, described associating similarity retrieval module further uses the Dijkstra shortest path first to obtain described retrieving images sequence to guarantee time continuity and Space Consistency according to the associating similarity between described a plurality of candidate images.
CN 201110197873 2011-07-14 2011-07-14 Method and device for synthesizing face video Active CN102254336B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 201110197873 CN102254336B (en) 2011-07-14 2011-07-14 Method and device for synthesizing face video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 201110197873 CN102254336B (en) 2011-07-14 2011-07-14 Method and device for synthesizing face video

Publications (2)

Publication Number Publication Date
CN102254336A CN102254336A (en) 2011-11-23
CN102254336B true CN102254336B (en) 2013-01-16

Family

ID=44981577

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 201110197873 Active CN102254336B (en) 2011-07-14 2011-07-14 Method and device for synthesizing face video

Country Status (1)

Country Link
CN (1) CN102254336B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11595617B2 (en) 2012-04-09 2023-02-28 Intel Corporation Communication using interactive avatars

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014194439A1 (en) * 2013-06-04 2014-12-11 Intel Corporation Avatar-based video encoding
CN103647922A (en) * 2013-12-20 2014-03-19 百度在线网络技术(北京)有限公司 Virtual video call method and terminals
WO2016101131A1 (en) 2014-12-23 2016-06-30 Intel Corporation Augmented facial animation
CN104679832A (en) * 2015-02-05 2015-06-03 四川长虹电器股份有限公司 System and method for searching single or multi-body combined picture based on face recognition
CN105184249B (en) * 2015-08-28 2017-07-18 百度在线网络技术(北京)有限公司 Method and apparatus for face image processing
WO2017101094A1 (en) 2015-12-18 2017-06-22 Intel Corporation Avatar animation system
CN106303233B (en) * 2016-08-08 2019-03-15 西安电子科技大学 A kind of video method for secret protection based on expression fusion
CN109670386A (en) * 2017-10-16 2019-04-23 深圳泰首智能技术有限公司 Face identification method and terminal
CN109784123A (en) * 2017-11-10 2019-05-21 浙江思考者科技有限公司 The analysis and judgment method of real's expression shape change
CN108510435A (en) * 2018-03-28 2018-09-07 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium
CN109886091B (en) * 2019-01-08 2021-06-01 东南大学 Three-dimensional facial expression recognition method based on weighted local rotation mode
CN109993102B (en) * 2019-03-28 2021-09-17 北京达佳互联信息技术有限公司 Similar face retrieval method, device and storage medium
CN110675433A (en) * 2019-10-31 2020-01-10 北京达佳互联信息技术有限公司 Video processing method and device, electronic equipment and storage medium
CN111274447A (en) * 2020-01-13 2020-06-12 深圳壹账通智能科技有限公司 Target expression generation method, device, medium and electronic equipment based on video
CN111652121B (en) * 2020-06-01 2023-11-03 腾讯科技(深圳)有限公司 Training method of expression migration model, and method and device for expression migration
CN111833413B (en) * 2020-07-22 2022-08-26 平安科技(深圳)有限公司 Image processing method, image processing device, electronic equipment and computer readable storage medium
CN113269872A (en) * 2021-06-01 2021-08-17 广东工业大学 Synthetic video generation method based on three-dimensional face reconstruction and video key frame optimization
CN113436302B (en) * 2021-06-08 2024-02-13 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) Face animation synthesis method and system
CN114429611B (en) * 2022-04-06 2022-07-08 北京达佳互联信息技术有限公司 Video synthesis method and device, electronic equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920886A (en) * 2006-09-14 2007-02-28 浙江大学 Video flow based three-dimensional dynamic human face expression model construction method
CN1920880A (en) * 2006-09-14 2007-02-28 浙江大学 Video flow based people face expression fantasy method
CN101179665A (en) * 2007-11-02 2008-05-14 腾讯科技(深圳)有限公司 Method and device for transmitting face synthesized video

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920886A (en) * 2006-09-14 2007-02-28 浙江大学 Video flow based three-dimensional dynamic human face expression model construction method
CN1920880A (en) * 2006-09-14 2007-02-28 浙江大学 Video flow based people face expression fantasy method
CN101179665A (en) * 2007-11-02 2008-05-14 腾讯科技(深圳)有限公司 Method and device for transmitting face synthesized video

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Ira Kemelmacher-Shlizerman, Aditya Sankar, Eli Shechtman, and St.Being John Malkovich.《European Conference on Computer Vision》.2010,1-13. *
Qingshan Zhang, et al..geometry-driven photorealistic facial expression synthesis.《IEEE Transactions on Visualization and Computer Graphics》.2006,第12卷(第1期),48-60. *
雷林华.基于视频序列的虚拟人脸合成.《中国优秀硕士学位论文全文数据库 信息科技辑》.2008,(第1期),全文. *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11595617B2 (en) 2012-04-09 2023-02-28 Intel Corporation Communication using interactive avatars

Also Published As

Publication number Publication date
CN102254336A (en) 2011-11-23

Similar Documents

Publication Publication Date Title
CN102254336B (en) Method and device for synthesizing face video
CN111652828B (en) Face image generation method, device, equipment and medium
Ge et al. 3d convolutional neural networks for efficient and robust hand pose estimation from single depth images
Liu et al. Fusedream: Training-free text-to-image generation with improved clip+ gan space optimization
Zhou et al. Photorealistic facial expression synthesis by the conditional difference adversarial autoencoder
Toshev et al. Deeppose: Human pose estimation via deep neural networks
CN106600626B (en) Three-dimensional human motion capture method and system
Lei et al. Whole-body humanoid robot imitation with pose similarity evaluation
KR101711736B1 (en) Feature extraction method for motion recognition in image and motion recognition method using skeleton information
CN102567716B (en) Face synthetic system and implementation method
CN104008564A (en) Human face expression cloning method
KR101563297B1 (en) Method and apparatus for recognizing action in video
US10147023B1 (en) Markerless face tracking with synthetic priors
CN102509333A (en) Action-capture-data-driving-based two-dimensional cartoon expression animation production method
Uddin et al. Human Activity Recognition via 3-D joint angle features and Hidden Markov models
CN115083015B (en) 3D human body posture estimation data labeling mode and corresponding model construction method
Lin et al. Facial expression recognition with data augmentation and compact feature learning
Lu et al. Multi-task learning for single image depth estimation and segmentation based on unsupervised network
Seddik et al. Unsupervised facial expressions recognition and avatar reconstruction from Kinect
Camurri et al. Visual Gesture Recognition: from motion tracking to expressive gesture
Neverova Deep learning for human motion analysis
CN106778576B (en) Motion recognition method based on SEHM characteristic diagram sequence
Zhang et al. Styleavatar3d: Leveraging image-text diffusion models for high-fidelity 3d avatar generation
CN103116901A (en) Motion characteristic based human motion interpolation calculation method
Tan Evolution of art form of video animation design under the background of computer graphics system development

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant