CN103530900B - Modeling method, face tracking method and the equipment of three-dimensional face model - Google Patents

Modeling method, face tracking method and the equipment of three-dimensional face model Download PDF

Info

Publication number
CN103530900B
CN103530900B CN201210231897.XA CN201210231897A CN103530900B CN 103530900 B CN103530900 B CN 103530900B CN 201210231897 A CN201210231897 A CN 201210231897A CN 103530900 B CN103530900 B CN 103530900B
Authority
CN
China
Prior art keywords
frame
expression
face
model
tracking
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201210231897.XA
Other languages
Chinese (zh)
Other versions
CN103530900A (en
Inventor
沈晓璐
冯雪涛
张辉
金培亭
金智渊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Samsung Telecommunications Technology Research Co Ltd
Samsung Electronics Co Ltd
Original Assignee
Beijing Samsung Telecommunications Technology Research Co Ltd
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Samsung Telecommunications Technology Research Co Ltd, Samsung Electronics Co Ltd filed Critical Beijing Samsung Telecommunications Technology Research Co Ltd
Priority to CN201210231897.XA priority Critical patent/CN103530900B/en
Priority to KR1020130043463A priority patent/KR102024872B1/en
Priority to US13/936,001 priority patent/US20140009465A1/en
Publication of CN103530900A publication Critical patent/CN103530900A/en
Application granted granted Critical
Publication of CN103530900B publication Critical patent/CN103530900B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • G06T13/403D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • G06V20/653Three-dimensional objects by matching three-dimensional models, e.g. conformal mapping of Riemann surfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/167Detection; Localisation; Normalisation using comparisons between temporally consecutive images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • G06V40/176Dynamic expression

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Geometry (AREA)
  • Computer Graphics (AREA)
  • Image Analysis (AREA)
  • Processing Or Creating Images (AREA)

Abstract

Modeling method, face tracking method and the equipment of a kind of three-dimensional face model are provided.A kind of modeling method of three-dimensional face model is provided, it include: using preset standard three-dimensional faceform as working model, and specified start frame is set as first frame, face tracking is carried out since the specified start frame of the multiple video frames continuously inputted based on working model, tracking result corresponding with the video frame of predetermined number is generated from the video frame extraction human face characteristic point of tracking, expression parameter and head pose parameter, and according to predetermined condition;Tracking result based on generation updates working model, wherein, if when updating working model, the difference of appearance parameter before determining the appearance parameter of the working model updated and updating is not less than scheduled limit value, continue to continue to execute face tracking to subsequent video frame, the tracking until completing all videos frame;Output services model.

Description

Modeling method, face tracking method and the equipment of three-dimensional face model
Technical field
This application involves a kind of modeling method of three-dimensional face model, face tracking method and for executing the modeling The equipment of method and face tracking method more particularly to a kind of face tracking is concurrently carried out from the face video frame continuously inputted With the modeling of three-dimensional face model, to provide the three-dimensional face model closer to user's face, and/or high-precision is exported Human facial expression information.
Background technique
With continuous videos be input, existing face tracking/modeling technique can export differing complexity as a result, example Such as: the classification of expression parameter and its intensity, face two-dimensional shapes, the face 3D shape of low-res or the people of high-res Face 3D shape.
Face tracking/modeling basic technology is broadly divided into three classes: the technology based on identification, the technology based on fitting and base In the technology of reconstruction.
But there are the following problems for existing face tracking/modeling technique: firstly, some technologies need to rely on volume Outer equipment, such as binocular camera, depth camera;Secondly, most of technology needs the cooperation of user, such as manual markings to close Key point, using preceding progress user's registration, amimia or fixed expression etc. is kept in modeling;In addition, some technologies cannot be defeated The tracking result of high-res out can not avoid the influence of appearance, posture to expression parameters precision.
Summary of the invention
The purpose of the present invention is to provide a kind of modeling methods of the three-dimensional face model of personalization, from the people continuously inputted Face video frame is made iteratively face tracking, and updates three-dimensional face model on the basis of face tracking result, to mention For the three-dimensional face model closer to user's face.
Another object of the present invention is to provide a kind of face tracking methods, iteratively from the face video frame continuously inputted Face tracking is carried out, three-dimensional face model is updated based on tracking result, to export high-precision human facial expression information.
According to an aspect of the present invention, a kind of modeling method of three-dimensional face model, the three-dimensional face model packet are provided The information of 3D shape, appearance parameter, expression parameter and head pose is included, the modeling method is the following steps are included: by pre- If standard three-dimensional faceform be set as first frame as working model, and by specified start frame, perform the following operations: a) base Face tracking is carried out in specified start frame of the working model since the multiple video frames continuously inputted, is mentioned from the video frame of tracking Human face characteristic point, expression parameter and head pose parameter are taken, and the video frame with predetermined number is generated according to predetermined condition Corresponding tracking result, the tracking result include each video frame of tracking and the face spy of each video frame extraction from tracking Levy point, expression parameter and head pose parameter;B) tracking result based on generation updates working model, wherein if more When new working model, the difference of the appearance parameter before determining the appearance parameter of the working model updated and updating is not less than scheduled limit Value, and the video frame after the video frame of the predetermined number is not the last frame in the multiple video frames continuously inputted, then Using first video frame after the video frame of the predetermined number as specified start frame, step a) is executed;And c) output services Model.
The three-dimensional face model may be expressed as
Wherein, S is 3D shape, and a is appearance component, and e is expression component, and q is head pose, and T (S, q) expression will be three-dimensional Shape S carries out function rotationally and/or translationally according to head pose q.
The standard three-dimensional faceform can include: average shape S0, appearance componentExpression componentAnd standard header Portion posture q0.Wherein, i=1:N,A kind of variation of expression face appearance, j=1:M,Indicate a kind of change of human face expression Change.
Lower online in advance the standard three-dimensional faceform can be trained using a series of three-dimensional face datas.
Step a) and step b) can be performed in parallel.
Tracking result described in step b) based on generation updates the processing of working model can include: in the tracking knot of generation In fruit, neutral expression's frame will be selected as closest to the video frame of neutral expression;According to the human face characteristic point in neutral expression's frame Face outline is extracted from neutral expression's frame;Face outline based on characteristic point and extraction in neutral expression's frame updates Working mould Type.
In selecting in the processing of property expression frame, can be in the corresponding tracking result of the video frame of T from predetermined number, It selects neutral expression's frame as follows: calculating expression parameter for the video frame of each trackingWherein, K For the number of species of expression parameter;By the highest expression parameter value of the frequency of occurrences in every kind of expression parameterAs neutral expression Value;The deviation of whole K expression parameters and corresponding neutral expression's value is selected to be respectively less than the video frame of scheduled each threshold value As neutral expression's frame.
Active contour model algorithm can be used to extract face outline from neutral expression's frame.
The face outline of the characteristic point and extraction based in neutral expression's frame updates the processing of working model can include: The head pose q of working model is updated to the head pose of neutral expression's frame;0 is set by the expression component e of working model; By the way that working model S (a, e, q) is matched with the characteristic point position of neutral expression's frame, and by working model S (a, e, q) The face outline of calculating matches the appearance component a for carrying out correction work model with the face outline extracted from neutral expression's frame.
Multiple modeling units can be used to be alternately carried out the processing of the step b) of different iteration, and by the place of different iteration Reason is merged.
In step a), also the tracking result generated can be exported by input/output interface.
In step b), when completing the update of working model, also the Working mould updated can be exported by input/output interface Type.
In step a), can be wanted according to the input rate of the multiple video frames continuously inputted, noise quality or the precision of tracking It asks and determines the predetermined number, and the predetermined number can be constant or variable.
In step a), one of following methods can be used to obtain human face characteristic point, expression parameter and head based on working model Attitude parameter: active appearance models method (AAM), Active Shape Model Method (ASM) and compound constant active appearance models (Composite Constraint AAM)。
If in step a), the face tracking processing returns to failure as a result, if can will reset the outer of working model Looks component a is 0, and returns to the beginning of step a).
According to another aspect of the present invention, a kind of face tracking method based on three-dimensional face model, the three-dimensional are provided Faceform includes the information of 3D shape, appearance parameter, expression parameter and head pose, and the face tracking method includes Following steps: using preset standard three-dimensional faceform as working model, 1 is set by modeling instruction, and will specify Beginning frame is set as first frame, performs the following operations: a) based on working model from the specified start frame of the multiple video frames continuously inputted Start to carry out face tracking, from the video frame extraction human face characteristic point of tracking, expression parameter and head pose parameter, according to pre- Fixed condition generates corresponding with the video frame of predetermined number tracking result, the tracking result include each video frame tracked and From the human face characteristic point of each video frame extraction of tracking, expression parameter and head pose parameter, and export the tracking knot Fruit;If b) modeling is designated as 1, the tracking result based on generation updates working model, wherein when updating working model, If it is determined that the difference of the appearance parameter before new appearance parameter and update is less than scheduled limit value, then 0 is set by modeling instruction; It, will if c) video frame after the video frame of the predetermined number is not the last frame in the multiple video frames continuously inputted First video frame after the video frame of the predetermined number executes step a) as specified start frame.
The three-dimensional face model may be expressed asWherein, S is 3D shape, a are appearance component, and e is expression component, and q is head pose, and T (S, q) is indicated 3D shape S according to head appearance State q carries out function rotationally and/or translationally.
The standard three-dimensional faceform can include: average shape S0, appearance componentExpression componentAnd standard header Portion posture q0, wherein i=1:N,A kind of variation of expression face appearance, j=1:M,Indicate a kind of change of human face expression Change.
Lower online in advance the standard three-dimensional faceform can be trained using a series of three-dimensional face datas.
Step a) and step b) can be performed in parallel.
Tracking result described in step b) based on generation updates the processing of working model can include: in the tracking knot of generation In fruit, neutral expression's frame will be selected as closest to the video frame of neutral expression;According to the human face characteristic point in neutral expression's frame Face outline is extracted from neutral expression's frame;Face outline based on characteristic point and extraction in neutral expression's frame updates Working mould Type.
In selecting in the processing of property expression frame, can be in the corresponding tracking result of the video frame of T from predetermined number, It selects neutral expression's frame as follows: calculating expression parameter for the video frame of each trackingWherein, K For the number of species of expression parameter;By the highest expression parameter value of the frequency of occurrences in every kind of expression parameterAs neutral expression Value;The deviation of whole K expression parameters and corresponding neutral expression's value is selected to be respectively less than the video frame of scheduled each threshold value As neutral expression's frame.
Active contour model algorithm can be used to extract face outline from neutral expression's frame.
The face outline of the characteristic point and extraction based in neutral expression's frame updates the processing of working model can include: The head pose q of working model is updated to the head pose of neutral expression's frame;0 is set by the expression component e of working model; By the way that working model S (a, e, q) is matched with the characteristic point position of neutral expression's frame, and by working model S (a, e, q) The face outline of calculating matches the appearance component a for carrying out correction work model with the face outline extracted from neutral expression's frame.
Multiple modeling units can be used to be alternately carried out the processing of the step b) of different iteration, and by the place of different iteration Reason is merged.
After the execution for completing step c), output services model can be gone back.
In step a), can be wanted according to the input rate of the multiple video frames continuously inputted, noise quality or the precision of tracking It asks and determines the predetermined number, and the predetermined number is constant or variable.
In step a), one of following methods can be used to obtain human face characteristic point, expression parameter and head based on working model Attitude parameter: active appearance models method (AAM), Active Shape Model Method (ASM) and compound constant active appearance models (Composite Constraint AAM)。
If in step a), the face tracking processing returns to failure as a result, if can will reset the outer of working model Looks component a is 0, and returns to the beginning of step a).
According to another aspect of the present invention, a kind of modelling apparatus of three-dimensional face model, the three-dimensional face model are provided Information including 3D shape, appearance parameter, expression parameter and head pose, the modelling apparatus include: the first module, are used It is set as first frame in using preset standard three-dimensional faceform as working model, and by specified start frame, controls the second mould Block handles the multiple video frames continuously inputted;Second module, for based on working model from the multiple views continuously inputted The specified start frame of frequency frame starts to carry out face tracking, from the video frame extraction human face characteristic point of tracking, expression parameter and head Portion's attitude parameter, and tracking result corresponding with the video frame of predetermined number, the tracking result are generated according to predetermined condition Human face characteristic point, expression parameter and the head pose ginseng of each video frame including tracking and each video frame extraction from tracking Number;Third module updates working model for the tracking result based on generation, wherein if determined when updating working model The difference of appearance parameter before new appearance parameter and update is not less than scheduled limit value, and after the video frame of the predetermined number Video frame be not last frame in the multiple video frames continuously inputted, then by first after the video frame of the predetermined number A video frame continues to handle the multiple video frames continuously inputted as specified start frame, the second module of control;4th mould Block is used for output services model.
The three-dimensional face model may be expressed asWherein, S is 3D shape, a are appearance component, and e is expression component, and q is head pose, and T (S, q) is indicated 3D shape S according to head appearance State q carries out function rotationally and/or translationally.
The standard three-dimensional faceform can include: average shape S0, appearance componentExpression componentAnd standard header Portion posture q0, wherein i=1:N,A kind of variation of expression face appearance, j=1:M,Indicate a kind of change of human face expression Change.
The modelling apparatus may also include that training module, be instructed for lower online in advance using a series of three-dimensional face datas Practise the standard three-dimensional faceform.
Second module and third module can be performed in parallel operation.
Third module can include: the 5th module will be closest to the view of neutral expression in the tracking result of generation Frequency frame is selected as neutral expression's frame;6th module, for being mentioned according to the human face characteristic point in neutral expression's frame from neutral expression's frame Take face outline;7th module updates working model for the face outline based on characteristic point and extraction in neutral expression's frame.
5th module can select as follows neutral expression's frame: calculate expression parameter for the video frame of each trackingWherein, K is the number of species of expression parameter;By frequency of occurrences highest in every kind of expression parameter Expression parameter valueAs neutral expression's value;Select the deviation of whole K expression parameters and corresponding neutral expression's value small In scheduled each threshold value video frame as neutral expression's frame.
Active contour model algorithm can be used to extract face outline from neutral expression's frame for 6th module.
7th module can based on the processing that the face outline of characteristic point and extraction in neutral expression's frame updates working model It include: the head pose that the head pose q of working model is updated to neutral expression's frame;The expression component e of working model is set It is set to 0;By the way that working model S (a, e, q) is matched with the characteristic point position of neutral expression's frame, and by working model S The face outline that (a, e, q) is calculated matches the appearance component for carrying out correction work model with the face outline extracted from neutral expression's frame a。
Third module may include multiple modeling units, and the multiple modeling unit is alternately carried out at the modeling of different iteration Reason, and third module merges the modeling processing of different iteration.
The tracking result of the also exportable generation of second module.
When third module completes the update of working model, third module can also be exported by input/output interface to be updated Working model.
Second module can be according to the input rate of the multiple video frames continuously inputted, noise quality or the required precision of tracking Determine the predetermined number, and the predetermined number can be constant or variable.
One of following methods can be used to obtain human face characteristic point, expression parameter and head appearance based on working model for second module State parameter: active appearance models method (AAM), Active Shape Model Method (ASM) and compound constant active appearance models (Composite Constraint AAM)。
If the second module carry out the face tracking processing returns to failure as a result, if can will reset working model Appearance component a is 0, and restarts to carry out face tracking from the specified start frame of the multiple video frames continuously inputted.
According to another aspect of the present invention, a kind of face tracking equipment based on three-dimensional face model, the three-dimensional are provided Faceform includes the information of 3D shape, appearance parameter, expression parameter and head pose, and the face tracking equipment includes Following steps: the first module, for setting 1 for modeling instruction using preset standard three-dimensional faceform as working model, And specified start frame is set as first frame, is performed the following operations: second module, for based on working model from continuously inputting The specified start frame of multiple video frames starts to carry out face tracking, video frame extraction human face characteristic point, expression parameter from tracking And head pose parameter, tracking result corresponding with the video frame of predetermined number, the tracking knot are generated according to predetermined condition Fruit includes each video frame of tracking and human face characteristic point, expression parameter and the head pose of each video frame extraction from tracking Parameter, and export the tracking result;Third module, if being configured as modeling is designated as 1, the tracking knot based on generation Fruit updates working model, wherein when updating working model, if it is determined that new appearance parameter and the appearance parameter before update it Difference is less than scheduled limit value, then sets 0 for modeling instruction;4th module, if being configured as the video of the predetermined number Video frame after frame is not the last frame in the multiple video frames continuously inputted, then will be after the video frame of the predetermined number First video frame then controls the second module and continues to handle the multiple video frames continuously inputted as specified start frame.
The three-dimensional face model may be expressed asWherein, S is 3D shape, a are appearance component, and e is expression component, and q is head pose, and T (S, q) is indicated 3D shape S according to head appearance State q carries out function rotationally and/or translationally.
The standard three-dimensional faceform can include: average shape S0, appearance componentExpression componentAnd standard header Portion posture q0, wherein i=1:N,A kind of variation of expression face appearance, j=1:M,Indicate a kind of change of human face expression Change.
The face tracking equipment can further include: training module uses a series of three-dimensional face numbers for lower online in advance According to training the standard three-dimensional faceform.
Second module and third module can be performed in parallel operation.
Third module can include: the 5th module will be closest to the view of neutral expression in the tracking result of generation Frequency frame is selected as neutral expression's frame;6th module, for being mentioned according to the human face characteristic point in neutral expression's frame from neutral expression's frame Take face outline;7th module updates working model for the face outline based on characteristic point and extraction in neutral expression's frame.
5th module can select as follows neutral expression's frame: calculate expression parameter for the video frame of each trackingWherein, K is the number of species of expression parameter;By frequency of occurrences highest in every kind of expression parameter Expression parameter valueAs neutral expression's value;Select the deviation of whole K expression parameters and corresponding neutral expression's value small In scheduled each threshold value video frame as neutral expression's frame.
Active contour model algorithm can be used to extract face outline from neutral expression's frame for 6th module.
7th module can based on the processing that the face outline of characteristic point and extraction in neutral expression's frame updates working model It include: the head pose that the head pose q of working model is updated to neutral expression's frame;The expression component e of working model is set It is set to 0;By the way that working model S (a, e, q) is matched with the characteristic point position of neutral expression's frame, and by working model S The face outline that (a, e, q) is calculated matches the appearance component for carrying out correction work model with the face outline extracted from neutral expression's frame a。
Third module may include multiple modeling units, and the multiple modeling unit is alternately carried out at the modeling of different iteration Reason, and third module merges the modeling processing of different iteration.
When third module completes the update of working model, third module, which can be exported also by input/output interface, to be updated Working model.
Second module can be according to the input rate of the multiple video frames continuously inputted, noise quality or the required precision of tracking Determine the predetermined number, and the predetermined number is constant or variable.
One of following methods can be used to obtain human face characteristic point, expression parameter and head appearance based on working model for second module State parameter: active appearance models method (AAM), Active Shape Model Method (ASM) and compound constant active appearance models (Composite Constraint AAM)。
If the second module carry out the face tracking processing returns to failure as a result, if face tracking equipment can will weight The appearance component a for setting working model is 0, and can restart to carry out from the specified start frame of the multiple video frames continuously inputted Face tracking.
Detailed description of the invention
By the description carried out with reference to the accompanying drawing, above and other purpose of the invention and feature will become more clear Chu, in which:
Figure 1A is the flow chart for showing the modeling method of three-dimensional face model of an exemplary embodiment of the present invention;
Figure 1B is the processing for showing the update working model in the modeling method of an exemplary embodiment of the present invention Flow chart;
Fig. 1 C is the flow chart shown in the face tracking method of an exemplary embodiment of the present invention;
Fig. 2 schematically shows the three-dimensional face model based on the building of general face;
Fig. 3 is the schematic diagram being exemplarily illustrated using characteristic point from video frame extraction face outline;
Fig. 4 is to be exemplarily illustrated progress Feature Points Matching and outline to match between three-dimensional face model and face outline Schematic diagram;
Fig. 5 A and Fig. 5 B are the modelling apparatus/face tracking equipment signals for showing an exemplary embodiment of the present invention Figure.
Specific embodiment
Detailed description of the present invention exemplary embodiment that hereinafter reference will be made to the drawings.
The modeling method and face tracking method of three-dimensional face model of the invention can be in general purpose computers or dedicated It is realized in processor, the method is to include the video frame (a period of time, within a few minutes) of face continuously inputted as defeated Enter, using preset high-precision standard three-dimensional faceform as working model, based on the working model to the video of input Frame carries out face tracking, and the tracking result tracked to the face of predetermined quantity is hereafter used to carry out more the working model Newly/amendment.After update/amendment working model, it is further continued for carrying out face tracking to video frame later, until obtaining Reach the three-dimensional face model of preset limit value or the face tracking and working model are completed to all videos frame of input Until update/amendment.According to demand, in the process, output has the face tracking knot of fine expression and head pose information Fruit, and/or after this process, the personalized three-dimensional face model exported.
The video frame continuously inputted can be to be mentioned in the digital video frequency flow by single general digital camera shooting The video frame of many integral photographs for taking and handling, is also possible to the video for many photos being continuously shot using digital camera Frame.The video frame continuously inputted can input to three-dimensional face model for carrying out the present invention by input/output interface Modeling method and face tracking method general purpose computer or application specific processor.
Fig. 4, which is shown, schematically shows the three-dimensional face model based on the building of any face.The three-dimensional face model Include, but are not limited to the information of the 3D shape of face, appearance parameter, expression parameter and head pose.In the present invention, The three-dimensional face model is expressed as Wherein, S is 3D shape, and a is Appearance component, e are expression component, and q is head pose, and T (S, q) expression rotates 3D shape S according to head pose q And/or the function of translation.
An exemplary embodiment of the present invention, the lower high density face surface different using expression, posture online in advance Data train the three-dimensional face model of standard.According to other embodiments of the invention, it is possible to use existing other methods obtain To the three-dimensional face model of the standard, or as needed, the three-dimensional face model with standard faces feature is determined as The three-dimensional face model of the standard.
The standard three-dimensional faceform includes:
Average shape S0: the average value of all training samples.
Appearance componentA kind of variation of each component statement face in terms of appearance.
Expression componentA kind of variation of each component statement face in terms of expression.
Head pose q0: standard three-dimensional position and the rotation angle of face are described.
Figure 1A is the flow chart for showing the modeling method of three-dimensional face model of an exemplary embodiment of the present invention.
It referring to Fig.1,, will in behaviour step S110 when starting the operation of modeling method of three-dimensional face model of the invention Preset standard three-dimensional faceform is as working model, and specified start frame is set as in the video frame continuously inputted One frame.The standard three-dimensional faceform is a series of three-dimensional trained in advance using various expressions, the human face data of posture Faceform.
In step S120, people is carried out since the specified start frame in the multiple video frames continuously inputted based on working model Face tracking, from the video frame extraction human face characteristic point of tracking, expression parameter and head pose parameter, and according to predetermined condition Generate corresponding with the video frame of predetermined number tracking result, the tracking result includes each video frame tracked and from tracking Each video frame extraction human face characteristic point, expression parameter and head pose parameter.An exemplary embodiment of the present invention, root The predetermined number is determined according to the input rate of the multiple video frames continuously inputted, noise quality or the required precision of tracking.This Outside, the predetermined number be can be with constant or variable.
An exemplary embodiment of the present invention also passes through the tracking of input/output interface output generation in step S120 As a result.
An exemplary embodiment of the present invention is obtained using one of following methods based on working model in step S120 Human face characteristic point, expression parameter and head pose parameter: active appearance models method (AAM), Active Shape Model Method (ASM) With compound constant active appearance models (Composite Constraint AAM).
Hereafter, in step S130, working model is updated based on the tracking result generated in step S120.Later with reference to figure 1B detailed description updates the specific processing of working model.
An exemplary embodiment of the present invention, in step S130, when completing the update of working model, also by defeated Enter/output interface output update working model.
In addition, in step S140, if when updating working model, determine the appearance parameter of the working model updated with more The difference of the appearance parameter of working model before new is not less than scheduled limit value, and the video after the video frame of the predetermined number Frame is not the last frame in the multiple video frames continuously inputted, then in step S150, after the video frame of the predetermined number First video frame as specified start frame, return step S120, the working model based on update, from the specified start frame Continue face tracking.
If the appearance parameter for determining the working model updated when updating working model and the working model before update The difference of appearance parameter be less than the video frame after the video frame of scheduled limit value or the predetermined number be continuously input it is multiple Last frame in video frame, thens follow the steps S150.That is, when determination has constructed the best three-dimensional to conform to a predetermined condition When faceform, or when the processing to all videos frame is completed, the update to working model can be stopped.
In step S160, exported the working model of update as personalized three-dimensional face model.
Figure 1B is exemplarily illustrated the processing carried out in the step S130 of Figure 1A.
B referring to Fig.1, in step S132, by the view in the tracking result that step S120 is generated closest to neutral expression Frequency frame is selected as neutral expression's frame.Preferred embodiment in accordance with the present invention, in the processing of step S132, it is assumed that tracking result with The video frame that reservation number is T is corresponding, calculates expression parameter for each video frame of tracking Wherein, K is the number of species of expression parameter;By the highest expression parameter value of the frequency of occurrences in every kind of expression parameterAs neutrality Expression value;Then, the deviation of whole K expression parameters and corresponding neutral expression's value is selected to be respectively less than scheduled each threshold value Video frame as neutral expression's frame.
In step S135, face outline is extracted from neutral expression's frame according to the human face characteristic point in neutral expression's frame.From step Rapid S120 includes the letter such as human face characteristic point, expression parameter and head pose parameter for each video frame extraction of tracking Breath.An exemplary embodiment of the present invention, the neutral expression's frame selected using active contour model algorithm from step S132 Extract face outline.
Fig. 3 A~Fig. 3 C is exemplarily illustrated the processing using human face characteristic point from video frame extraction face outline.According to this The exemplary embodiment of invention is referring to (figure with the characteristic point of the video frame when from the video frame extraction outline of Fig. 3 A 3B), face outline as shown in Figure 3 C is extracted from video frame extraction face outline using active contour model algorithm.In this hair In bright its face outline similarly can be extracted from neutral expression's frame.
Hereafter, in step S138, the face outline based on characteristic point and extraction in neutral expression's frame updates working model. Specifically, the head pose q of working model to be updated to the head pose of neutral expression's frame;By the expression component e of working model It is set as 0;By the way that working model S (a, e, q) is matched with the characteristic point position of neutral expression's frame, and by working model The face outline that S (a, e, q) is calculated matches the appearance point for carrying out correction work model with the face outline extracted from neutral expression's frame Measure a.Fig. 4 B shows the work for making its characteristic point be overlapped locking with the characteristic point of neutral expression's frame in Fig. 4 A adjustment working model Make model;Fig. 4 D, which is shown, is overlapped working model as far as possible to adjust working model, correction with the face outline (Fig. 4 D) of extraction The processing of its appearance parameter.
During appearance parameter of the correction with appearance representation in components, the value and correction of the appearance parameter before record correction The value of appearance parameter afterwards, thus the basis compared as step S140.
Step S120~S150 in Figure 1A iteratively carries out face tracking and model modification to the video frame continuously inputted. An exemplary embodiment of the present invention is performed in parallel the processing of step S120 and S130.By using current Working mould Type carries out face tracking, extracts human face characteristic point, expression parameter and head pose parameter, and by using including extraction Human face characteristic point and head pose parameter and corresponding video frame further update working model, can get with user more For close personalized human face model.At the same time, the face tracking of also exportable each input video frame is as a result, include expression Parameter, appearance parameter and head pose etc..
Fig. 1 C shows the processing in the face tracking method of an exemplary embodiment of the present invention.
Since the face tracking method is to export face tracking result as main purpose, in fig. 1 c, once The best model to conform to a predetermined condition is obtained, then no longer carries out the update of working model, but continues subsequent video frame Face tracking processing.
C referring to Fig.1 is grasping step S110 ' when starting the operation of modeling method of three-dimensional face model of the invention, Using preset standard three-dimensional faceform as working model, specified start frame is set as first in the video frame continuously inputted Frame, and it is 1 (or "Yes") that the variable (for example, modeling instruction) for indicating whether to continue working model and updating, which is arranged,.It is described Standard three-dimensional faceform is a series of three-dimensional face model trained in advance using various expressions, the human face data of posture.
Processing in the processing and Figure 1A of step S120 in Fig. 1 C is essentially identical, but in fig. 1 c, it completes to make a reservation for After the face tracking of the video frame of number, step S125 and S128 are executed.In step S125, the video frame of each tracking is exported Tracking result, including expression parameter, appearance parameter and head pose etc..
In step S128, it is determined whether continue the update of working model (modeling indicates whether to be 1).If modeling refers to It is shown as 1, thens follow the steps S130.The processing of S130 described in the processing of the step S130 and Figure 1A is identical.
In step S140 ', if when updating working model, before determining the appearance parameter of the working model updated and updating Working model appearance parameter difference be not less than scheduled limit value, then in step S150, by the video frame of the predetermined number First video frame afterwards is as specified start frame, and then return step S120, the working model based on update are specified from described Start frame continues face tracking.
On the other hand, if it is determined that the appearance parameter of the working model of update and the appearance parameter of the working model before update Difference be less than or equal to scheduled limit value, then in step S145, will indicate whether that the modeling for continuing working model update refers to Show and is set as 0 (or "No").That is, at this time according to the present invention, determining the three-dimensional face mould for having constructed close enough user Type.In order to reduce operand, indicate that equipment of the invention no longer updates the working model.
Hereafter, in step S148, it is more whether the video frame after determining the video frame of the predetermined number continuously inputs Last frame in a video frame.If it is determined that the video frame after the video frame of the predetermined number be not continuously input it is multiple Last frame in video frame, thens follow the steps S150, using first video frame after the video frame of the predetermined number as Start frame is specified, then return step S120.
If it is determined that the video frame after the video frame of the predetermined number is last in the multiple video frames continuously inputted One frame then terminates the processing of face tracking method of the invention.
An exemplary embodiment of the present invention, before the processing for terminating face tracking method of the invention, output is last The working model of update.
It can be seen that carry out face tracking by using current working model, extract human face characteristic point, expression parameter with And head pose parameter, and by using the human face characteristic point and head pose parameter that include extraction and corresponding video frame Further update working model, can be made iteratively based on the personalized human face model more close with user face with Track, and exportable each more accurate face tracking is as a result, include expression parameter, appearance parameter and head pose etc..
Fig. 5 A show the realization three-dimensional face model of an exemplary embodiment of the present invention modeling method and/or The schematic structure of the equipment of face tracking method.
The equipment includes tracking cell and modeling unit, the tracking cell execute the step S110 in Figure 1A~ S110 '~S125 in S120 or Fig. 1 C, the modeling unit execute step S130~S150 in Figure 1A or Fig. 1 C.
Referring to Fig. 5 A, tracking cell (is initially the standard three-dimensional faceform M using working model0) to video frame 0 To video frame t2- 1 carries out face tracking, and will be including video frame 0 to video frame t2- 1 and from the people of each video frame extraction Tracking result (i.e. 0~result of result t in Fig. 5 A of face characteristic point, expression parameter and head pose parameter2- 1) it exports.Institute It states tracking result and is provided to modeling unit, and as needed, can be exported by input/output interface to user.
Tracking result (0~result of result t that the modeling unit is exported based on tracking cell2- 1) working model is updated (hereinbefore the update of working model is described in detail with reference to Figure 1A and Figure 1B).It for convenience of explanation, will herein more New working model is known as M1
Hereafter, tracking cell is according to scheduled rule, the working model M based on update1To video frame t2- 1 to t3 carries out people Face tracks (Figure 1A description as described previously with reference to), and output tracking result (result t2- 1 to result t3).Modeling unit is based on result t2- 1~result t3Update working model M1.It is so made iteratively face tracking and updates the processing of working model, until really Surely it obtains qualified best model or completes to fully enter the processing of video frame.The tracking cell and modeling unit can Concurrently operated.
Preferred embodiment in accordance with the present invention, the equipment may include training unit, use a system for lower online in advance Column three-dimensional face data trains standard three-dimensional faceform as working model M0
Fig. 5 B shows the modeling method of the realization three-dimensional face model in accordance with an alternative illustrative embodiment of the present invention And/or the schematic structure of the equipment of face tracking method.
Unlike the equipment in Fig. 5 A, the equipment in Fig. 5 B includes multiple modeling units (such as modeling unit A and modeling Unit B), multiple modeling units are alternately carried out the processing of the step S130 of different iteration, and the processing of different iteration is melted It closes.
Although show and describing the present invention with reference to preferred embodiment, it will be understood by those skilled in the art that not In the case where being detached from the spirit and scope of the present invention that are defined by the claims, these embodiments can be carry out various modifications and Transformation.

Claims (15)

1. a kind of modeling method of three-dimensional face model, the three-dimensional face model includes 3D shape, appearance parameter, expression ginseng Several and head pose information, the modeling method the following steps are included:
Using a preset standard three-dimensional faceform as working model, and specified start frame is set as first frame, executed Following operation:
A) face tracking is carried out since the specified start frame of the multiple video frames continuously inputted based on working model, from tracking Video frame extraction human face characteristic point, expression parameter and head pose parameter, and according to predetermined condition generation and predetermined number The corresponding tracking result of video frame, each video frame and each video frame extraction from tracking that the tracking result includes tracking Human face characteristic point, expression parameter and head pose parameter;
B) tracking result based on generation updates working model, wherein if determining the work updated when updating working model The appearance parameter of model and the difference of the appearance parameter before update are not less than scheduled limit value, and the video frame of the predetermined number Video frame afterwards is not the last frame in the multiple video frames continuously inputted, then by after the video frame of the predetermined number One video frame executes step a) as specified start frame;
C) output services model.
2. modeling method as described in claim 1, wherein the three-dimensional face model is represented as
Wherein, S is 3D shape, and a is appearance component, and e is expression component, and q is head pose, and T (S, q) is indicated 3D shape S carries out function rotationally and/or translationally according to head pose q.
3. modeling method as claimed in claim 2, wherein the standard three-dimensional faceform includes: average shape S0, appearance ComponentExpression componentAnd standard head posture q0,
Wherein, i=1:N,A kind of variation of expression face appearance, j=1:M,Indicate a kind of variation of human face expression.
4. modeling method as claimed in claim 3, further includes:
The standard three-dimensional faceform is trained using a series of three-dimensional face datas under online in advance.
5. modeling method as claimed in claim 2, wherein be performed in parallel step a) and step b).
6. modeling method as claimed in claim 3, wherein based on the tracking result of generation described in step b), update work The processing of model includes:
In the tracking result of generation, neutral expression's frame will be selected as closest to the video frame of neutral expression;
Face outline is extracted from neutral expression's frame according to the human face characteristic point in neutral expression's frame;
Face outline based on characteristic point and extraction in neutral expression's frame updates working model.
7. modeling method as claimed in claim 6, wherein in selecting in the processing of property expression frame, from predetermined number be T The corresponding tracking result of video frame in, it is following to select neutral expression's frame:
Expression parameter is calculated for the video frame of each trackingWherein, K is the type of expression parameter Quantity;
By the highest expression parameter value of the frequency of occurrences in every kind of expression parameterAs neutral expression's value;
The deviation of whole K expression parameters and corresponding neutral expression's value is selected to be respectively less than the video frame of scheduled each threshold value As neutral expression's frame.
8. modeling method as claimed in claim 7, wherein extract face from neutral expression's frame using active contour model algorithm Outline.
9. modeling method as claimed in claim 8, wherein the face of the characteristic point and extraction based in neutral expression's frame Outline update working model processing include:
The head pose q of working model is updated to the head pose of neutral expression's frame;
0 is set by the expression component e of working model;
By the way that working model S (a, e, q) is matched with the characteristic point position of neutral expression's frame, and by working model S (a, E, q) calculate face outline the appearance component a for carrying out correction work model is matched with the face outline extracted from neutral expression's frame.
10. modeling method as claimed in claim 2, wherein be alternately carried out the step of different iteration using multiple modeling units Rapid processing b), and the processing of different iteration is merged.
11. modeling method as described in claim 1, wherein in step a), also exported and generated by input/output interface Tracking result.
12. modeling method as described in claim 1, wherein also pass through in step b) when completing the update of working model The working model that input/output interface output updates.
13. modeling method as claimed in claim 2, wherein in step a), according to the input of the multiple video frames continuously inputted Rate, noise quality or the required precision of tracking determine the predetermined number, and the predetermined number is constant or variable.
14. modeling method as claimed in claim 2, wherein in step a), obtained using one of following methods based on working model Take human face characteristic point, expression parameter and head pose parameter: active appearance models method (AAM), Active Shape Model Method (ASM) and compound constant active appearance models (Composite Constraint AAM).
15. modeling method as claimed in claim 13, wherein if in step a), the face tracking processing returns to mistakes It is losing as a result, then will reset working model appearance component a be 0, and return to step a) beginning.
CN201210231897.XA 2012-07-05 2012-07-05 Modeling method, face tracking method and the equipment of three-dimensional face model Expired - Fee Related CN103530900B (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN201210231897.XA CN103530900B (en) 2012-07-05 2012-07-05 Modeling method, face tracking method and the equipment of three-dimensional face model
KR1020130043463A KR102024872B1 (en) 2012-07-05 2013-04-19 Method and apparatus for modeling 3d face, method and apparatus for tracking face
US13/936,001 US20140009465A1 (en) 2012-07-05 2013-07-05 Method and apparatus for modeling three-dimensional (3d) face, and method and apparatus for tracking face

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210231897.XA CN103530900B (en) 2012-07-05 2012-07-05 Modeling method, face tracking method and the equipment of three-dimensional face model

Publications (2)

Publication Number Publication Date
CN103530900A CN103530900A (en) 2014-01-22
CN103530900B true CN103530900B (en) 2019-03-19

Family

ID=49932878

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210231897.XA Expired - Fee Related CN103530900B (en) 2012-07-05 2012-07-05 Modeling method, face tracking method and the equipment of three-dimensional face model

Country Status (2)

Country Link
KR (1) KR102024872B1 (en)
CN (1) CN103530900B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105608710B (en) * 2015-12-14 2018-10-19 四川长虹电器股份有限公司 A kind of non-rigid Face datection and tracking positioning method
CN108229246A (en) * 2016-12-14 2018-06-29 上海交通大学 Real-time three-dimensional human face posture method for tracing based on vehicle computing machine platform
CN108345821B (en) * 2017-01-24 2022-03-08 成都理想境界科技有限公司 Face tracking method and device
CN108734735B (en) * 2017-04-17 2022-05-31 佳能株式会社 Object shape tracking device and method, and image processing system
US10586368B2 (en) * 2017-10-26 2020-03-10 Snap Inc. Joint audio-video facial animation system
CN107958479A (en) * 2017-12-26 2018-04-24 南京开为网络科技有限公司 A kind of mobile terminal 3D faces augmented reality implementation method
CN108765464A (en) * 2018-05-31 2018-11-06 山东工商学院 Low-rank re-detection context long time-tracking method and system based on residual compensation
CN110796083B (en) * 2019-10-29 2023-07-04 腾讯科技(深圳)有限公司 Image display method, device, terminal and storage medium
CN112017230A (en) * 2020-09-07 2020-12-01 浙江光珀智能科技有限公司 Three-dimensional face model modeling method and system
CN113112596B (en) * 2021-05-12 2023-10-24 北京深尚科技有限公司 Face geometric model extraction and 3D face reconstruction method, equipment and storage medium
CN116152900B (en) * 2023-04-17 2023-07-18 腾讯科技(深圳)有限公司 Expression information acquisition method and device, computer equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101499128A (en) * 2008-01-30 2009-08-05 中国科学院自动化研究所 Three-dimensional human face action detecting and tracing method based on video stream
CN101777116A (en) * 2009-12-23 2010-07-14 中国科学院自动化研究所 Method for analyzing facial expressions on basis of motion tracking
CN102479388A (en) * 2010-11-22 2012-05-30 北京盛开互动科技有限公司 Expression interaction method based on face tracking and analysis

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100292810B1 (en) * 1999-03-19 2001-06-15 윤덕용 A Real time face tracking technique using face's color model and ellipsoid approximation model
KR20010089664A (en) * 1999-10-21 2001-10-08 요트.게.아. 롤페즈 System and method for three-dimensional modeling
US6845171B2 (en) * 2001-11-19 2005-01-18 Microsoft Corporation Automatic sketch generation
US7936902B2 (en) * 2004-11-12 2011-05-03 Omron Corporation Face feature point detection apparatus and feature point detection apparatus
US7755619B2 (en) * 2005-10-13 2010-07-13 Microsoft Corporation Automatic 3D face-modeling from video
US7536030B2 (en) * 2005-11-30 2009-05-19 Microsoft Corporation Real-time Bayesian 3D pose tracking
JP4999570B2 (en) * 2007-06-18 2012-08-15 キヤノン株式会社 Facial expression recognition apparatus and method, and imaging apparatus
KR100896643B1 (en) * 2007-06-18 2009-05-08 에스케이 텔레콤주식회사 Method and system for modeling face in three dimension by means of aam, and apparatus applied to the same
CN101159064B (en) * 2007-11-29 2010-09-01 腾讯科技(深圳)有限公司 Image generation system and method for generating image
KR100940862B1 (en) * 2007-12-17 2010-02-09 한국전자통신연구원 Head motion tracking method for 3d facial model animation from a video stream
KR20110021330A (en) * 2009-08-26 2011-03-04 삼성전자주식회사 Apparatus and method for creating third dimension abatar
US8339459B2 (en) * 2009-09-16 2012-12-25 Microsoft Corporation Multi-camera head pose tracking
KR101615719B1 (en) * 2009-09-18 2016-04-27 삼성전자주식회사 Apparatus and method for extracting user's third dimension facial expression

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101499128A (en) * 2008-01-30 2009-08-05 中国科学院自动化研究所 Three-dimensional human face action detecting and tracing method based on video stream
CN101777116A (en) * 2009-12-23 2010-07-14 中国科学院自动化研究所 Method for analyzing facial expressions on basis of motion tracking
CN102479388A (en) * 2010-11-22 2012-05-30 北京盛开互动科技有限公司 Expression interaction method based on face tracking and analysis

Also Published As

Publication number Publication date
KR102024872B1 (en) 2019-09-24
CN103530900A (en) 2014-01-22
KR20140009013A (en) 2014-01-22

Similar Documents

Publication Publication Date Title
CN103530900B (en) Modeling method, face tracking method and the equipment of three-dimensional face model
US9939888B2 (en) Correlating movement information received from different sources
CN108161882B (en) Robot teaching reproduction method and device based on augmented reality
EP2369553B1 (en) Three-dimensional template transformation method and apparatus
CN112541963A (en) Three-dimensional virtual image generation method and device, electronic equipment and storage medium
CN108305283A (en) Human bodys' response method and device based on depth camera and basic form
CN108664122A (en) A kind of attitude prediction method and apparatus
US11557391B2 (en) Systems and methods for human pose and shape recovery
EP3381017A1 (en) Face reconstruction from a learned embedding
US20090153569A1 (en) Method for tracking head motion for 3D facial model animation from video stream
CN104978764A (en) Three-dimensional face mesh model processing method and three-dimensional face mesh model processing equipment
JP7378642B2 (en) Generative nonlinear human shape model
CN110414546A (en) Use intermediate loss function training image signal processor
JP2010512581A5 (en)
US11963741B2 (en) Systems and methods for human pose and shape recovery
CN111127668B (en) Character model generation method and device, electronic equipment and storage medium
CN110533639A (en) A kind of key independent positioning method and device
US20150187114A1 (en) Method and apparatus for editing 3d character motion
CN110473291A (en) Container modeling method and device for container acquisition
CN114119912A (en) Rapid fitting method and device for human body model and storage medium
CN113408443B (en) Gesture posture prediction method and system based on multi-view images
CN109325994A (en) A method of enhanced based on three-dimensional face data
CN113821108B (en) Robot remote control system and control method based on multi-mode interaction technology
CN103839280B (en) A kind of human body attitude tracking of view-based access control model information
KR102147930B1 (en) Method and apparatus for recognizing pose

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
EXSB Decision made by sipo to initiate substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20190319