CN110276289A - Generate the method and human face characteristic point method for tracing of Matching Model - Google Patents

Generate the method and human face characteristic point method for tracing of Matching Model Download PDF

Info

Publication number
CN110276289A
CN110276289A CN201910523891.1A CN201910523891A CN110276289A CN 110276289 A CN110276289 A CN 110276289A CN 201910523891 A CN201910523891 A CN 201910523891A CN 110276289 A CN110276289 A CN 110276289A
Authority
CN
China
Prior art keywords
characteristic point
shape
image
matching model
matching
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910523891.1A
Other languages
Chinese (zh)
Other versions
CN110276289B (en
Inventor
王喆
许清泉
张伟
洪炜冬
曾志勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen Meitu Technology Co Ltd
Original Assignee
Xiamen Meitu Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen Meitu Technology Co Ltd filed Critical Xiamen Meitu Technology Co Ltd
Priority to CN201910523891.1A priority Critical patent/CN110276289B/en
Publication of CN110276289A publication Critical patent/CN110276289A/en
Application granted granted Critical
Publication of CN110276289B publication Critical patent/CN110276289B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Abstract

The invention discloses the method and human face characteristic point method for tracing that generate Matching Model, Matching Model is suitable for carrying out Feature Points Matching to the facial image for having marked characteristic point.Wherein, the method for generating Matching Model includes: to generate the image block centered on characteristic point based on the facial image for having marked characteristic point;Manual tag figure is generated based on described image block;Manual tag figure is inputted in the convolutional neural networks of pre-training, to export prediction label figure;And based on the penalty values between manual tag figure and prediction label figure, the training convolutional neural networks, using the network after training as the Matching Model generated.The program can be improved the accuracy of Feature Points Matching, to improve the stability and accuracy of face characteristic point tracking.

Description

Generate the method and human face characteristic point method for tracing of Matching Model
Technical field
The present invention relates to depth learning technology fields, the especially method of generation Matching Model and face characteristic point tracking side Method.
Background technique
Facial modeling is also known as face alignment, and target, which is oriented, can describe face feature in face picture Point, such as canthus, nose, the corners of the mouth, chin.Currently, facial modeling has had the method for many maturations in static images, Such as supervise gradient descent method, local binarization method.But the tracking for human face characteristic point in dynamic picture such as video flowing Positioning, due to not accounting for the timing of before and after frames in video flowing, it will usually serious characteristic point shake occur.
With the real-time filter of U.S. face camera, makeups function and augmented reality application, the real-time tracing of human face characteristic point Technology is particularly important.Face tracking algorithm or accurate but unstable at present, unstable rule leads to the shake of Real-time Special Effect;? Stablize but inaccurate, position that is inaccurate then leading to special efficacy is deviateed.Limited partial model is a kind of facial modeling algorithm, It is made the characteristic point on each average face scan for matching on its neighborhood position and has been come by the position of initialization average face It is detected at face point.Wherein, local detectors scanned picture is used first, generates a response for everyone face characteristic point Figure is then based on response diagram and optimizes a Global Face shape, to obtain the primary update of human face characteristic point.Limited office Portion's model is made of local expert and global shape model, wherein local expert instructs usually using feature and support vector machines Practice.Since feature and learning method are weaker, part unconspicuous for feature, such as the key point on lip and facial contour, Generally occur within trace error.
In consideration of it, the method for needing a kind of pair of characteristic point to position in real time, can be not only stable but also accurately carry out characteristic point and chase after Track.
Summary of the invention
For this purpose, the present invention provides generate Matching Model method and human face characteristic point method for tracing, with try hard to solve or Person at least alleviates at least one existing problem above.
According to an aspect of the invention, there is provided it is a kind of generate Matching Model method, wherein Matching Model be suitable for pair The facial image for having marked characteristic point carries out Feature Points Matching.This method can execute in calculating equipment, be primarily based on and marked The facial image of characteristic point is infused, the image block centered on characteristic point is generated.It is then based on image block, generates manual tag figure. Then, manual tag figure is inputted in the convolutional neural networks of pre-training, to export prediction label figure.Finally, based on artificial The penalty values of loss function between label figure and prediction label figure, training convolutional neural networks, using the network after training as The Matching Model of generation
Optionally, in the above-mentioned methods, manual tag figure includes positive sample region, negative sample region and inactive area.
Optionally, in the above-mentioned methods, convolutional neural networks may include convolutional neural networks include process of convolution layer and Warp lamination, wherein process of convolution layer can carry out convolution, activation, pondization processing to the image block of input, special with output prediction Levy the probability of point.Warp lamination can classify to the probability of predicted characteristics point, to export prediction label figure.
It optionally, in the above-mentioned methods, can be based on the power of area pixel point each in manual tag figure and prediction label figure Weight, calculates the penalty values of loss function.Penalty values based on loss function adjust convolutional neural networks using back-propagation algorithm Parameter, until penalty values be less than predetermined threshold.
Optionally, in the above-mentioned methods, loss function is binary cross entropy loss function, calculates loss based on following formula The penalty values of function:
Wherein, i is the index of pixel, predictiIndicate prediction label figure, lableiIndicate the picture in manual tag figure Vegetarian refreshments, wiIndicate the weight of pixel.
Optionally, in the above-mentioned methods, the weight of positive sample area pixel point and be 1, the power of negative sample area pixel point Weight and be 1, the weight of inactive area is 0.
According to another aspect of the present invention, a kind of human face characteristic point method for tracing is provided, can be held in calculating equipment Row, in the method, can be primarily based on shape and initialize to facial image to be detected, obtain face characteristic point diagram Picture.Then, multiple image blocks comprising characteristic point are obtained from face characteristic point image.It then, will be after image block input training Matching Model, with export prediction matching characteristic point.Wherein Matching Model is generated using the method for above-mentioned generation Matching Model.
Optionally, in the method, shape can be constructed based on the facial video image frame for having marked characteristic point.
It optionally, in the method, can be based on the characteristic point vector sum present frame facial image of previous frame facial image Characteristic point vector, obtain change in shape matrix.Then, principal component analysis is carried out to change in shape matrix, obtains change in shape Weight vectors.Finally, the weight vectors based on change in shape matrix and change in shape, complete the building of shape.
According to another aspect of the invention, a kind of calculating equipment is provided, comprising: one or more processors;And storage Device;One or more programs, wherein one or more programs store in memory and are configured as being handled by one or more Device executes, and one or more programs include the instruction for either executing in method as described above method.
In accordance with a further aspect of the present invention, a kind of computer-readable storage medium for storing one or more programs is provided Matter, one or more programs include instruction, and instruction is when calculating equipment execution, so that calculating equipment executes method as described above In either method.
According to the solution of the present invention, by being fused to local restricted model for convolutional neural networks as local expert model Frame in, can be improved the accuracy of Feature Points Matching.
Detailed description of the invention
To the accomplishment of the foregoing and related purposes, certain illustrative sides are described herein in conjunction with following description and drawings Face, these aspects indicate the various modes that can practice principles disclosed herein, and all aspects and its equivalent aspect It is intended to fall in the range of theme claimed.Read following detailed description in conjunction with the accompanying drawings, the disclosure it is above-mentioned And other purposes, feature and advantage will be apparent.Throughout the disclosure, identical appended drawing reference generally refers to identical Component or element.
Fig. 1 shows the schematic diagram of the limited partial model of tradition;
Fig. 2 shows the organigrams according to an embodiment of the invention for calculating equipment 100;
Fig. 3 shows the flow diagram of the method 300 according to an embodiment of the invention for generating Matching Model;
Fig. 4 shows the schematic diagram of positive sample according to an embodiment of the invention and negative sample;
Fig. 5 shows the schematic diagram of manual tag figure according to an embodiment of the invention;
Fig. 6 shows the schematic diagram of prediction label figure according to an embodiment of the invention;
Fig. 7 shows the schematic flow chart of human face characteristic point method for tracing 700 according to an embodiment of the invention.
Specific embodiment
Exemplary embodiments of the present disclosure are described in more detail below with reference to accompanying drawings.Although showing the disclosure in attached drawing Exemplary embodiment, it being understood, however, that may be realized in various forms the disclosure without should be by embodiments set forth here It is limited.On the contrary, these embodiments are provided to facilitate a more thoroughly understanding of the present invention, and can be by the scope of the present disclosure It is fully disclosed to those skilled in the art.
Traditional limited partial model is made of local expert and global shape model.Whole process is divided into model construction rank Section and point fitting stage.The model construction stage is divided into global shape model construction and Partial matched model building again.Shape Building is to generate a point spread function using active shape model ASM, will indicate several key feature points of body form Coordinate is connected into a vector, this object is indicated with the vector.And Partial matched model is to adjacent around each characteristic point Domain is modeled, and a Feature Points Matching criterion is established, to determine the best match of characteristic point.Fig. 1 shows the limited office of tradition The building schematic diagram of portion's model.As shown in Figure 1, being based on points distribution models, a people can be initialized on the face detected Then face shape model allows each point to find optimal match point in its contiguous range.For example, can along edge direction into Various rudimentary matchings, the matching error rates such as row block matching, point matching are higher.
This programme improves the human face characteristic point method for tracing based on limited partial model, and convolutional neural networks are made It is fused in limited partial model for local expert, replaces pixel characteristic and support vector machines usually used in local expert, Weak characteristic point is accurately tracked.
Fig. 2 is the block diagram of Example Computing Device 100.In basic configuration 102, calculating equipment 100, which typically comprises, is System memory 106 and one or more processor 104.Memory bus 108 can be used for storing in processor 104 and system Communication between device 106.
Depending on desired configuration, processor 104 can be any kind of processor, including but not limited to: micro process Device (μ P), microcontroller (μ C), digital information processor (DSP) or any combination of them.Processor 104 may include all Cache, processor core such as one or more rank of on-chip cache 110 and second level cache 112 etc 114 and register 116.Exemplary processor core 114 may include arithmetic and logical unit (ALU), floating-point unit (FPU), Digital signal processing core (DSP core) or any combination of them.Exemplary Memory Controller 118 can be with processor 104 are used together, or in some implementations, and Memory Controller 118 can be an interior section of processor 104.
Depending on desired configuration, system storage 106 can be any type of memory, including but not limited to: easily The property lost memory (RAM), nonvolatile memory (ROM, flash memory etc.) or any combination of them.System storage Device 106 may include operating system 120, one or more is using 122 and program data 124.In some embodiments, It may be arranged to be operated using program data 124 on an operating system using 122.In some embodiments, equipment is calculated 100 are configured as executing the method 300 or human face characteristic point method for tracing 700 for generating Matching Model, just wrap in program data 124 The instruction for executing above-mentioned each method is contained.
Calculating equipment 100 can also include facilitating from various interface equipments (for example, output equipment 142, Peripheral Interface 144 and communication equipment 146) to basic configuration 102 via the communication of bus/interface controller 130 interface bus 140.Example Output equipment 142 include graphics processing unit 148 and audio treatment unit 150.They can be configured as facilitate via One or more port A/V 152 is communicated with the various external equipments of such as display or loudspeaker etc.Outside example If interface 144 may include serial interface controller 154 and parallel interface controller 156, they, which can be configured as, facilitates Via one or more port I/O 158 and such as input equipment (for example, keyboard, mouse, pen, voice-input device, image Input equipment) or the external equipment of other peripheral hardwares (such as printer, scanner etc.) etc communicated.Exemplary communication is set Standby 146 may include network controller 160, can be arranged to convenient for via one or more communication port 164 and one A or multiple other calculate communication of the equipment 162 by network communication link.
Network communication link can be an example of communication media.Communication media can be usually presented as in such as carrier wave Or computer readable instructions, data structure, program module in the modulated data signal of other transmission mechanisms etc, and can To include any information delivery media." modulated data signal " can be such signal, one in its data set or Multiple or its change can be carried out in a manner of encoded information in the signal.As unrestricted example, communication media It may include the wired medium of such as cable network or private line network etc, and such as sound, radio frequency (RF), microwave, red Various wireless mediums including (IR) or other wireless mediums outside.Term computer-readable medium used herein may include Both storage medium and communication media.In some embodiments, one or more programs are stored in computer-readable medium, this It include the instruction for executing certain methods in a or multiple programs.
Calculating equipment 100 can be implemented as a part of portable (or mobile) electronic equipment of small size, these electronics are set It is standby to can be such as cellular phone, digital camera, personal digital assistant (PDA), personal media player device, wireless network Browsing apparatus, personal helmet, application specific equipment or may include any of the above function mixing apparatus.Certainly, it counts Calculate equipment 100 and also can be implemented as include desktop computer and notebook computer configuration personal computer, or have The server of above-mentioned configuration.Embodiments of the present invention to this with no restriction.
Fig. 3 shows the flow diagram of the method 300 according to an embodiment of the invention for generating Matching Model.Matching Model can be used for carrying out Feature Points Matching to the facial image for having marked characteristic point.Since Feature Points Matching is selected from image It takes certain characteristic points and partial analysis, and non-viewing whole image is carried out to image.Therefore facial image can be located in advance Reason, obtains the image block centered on each characteristic point.Then the image block used is as the training image of model.Below The process of the method 300 according to an embodiment of the present invention for generating Matching Model will be elaborated in conjunction with Fig. 3.
As shown in figure 3, method 300 starts from step S310.In step s310, based on the face figure for having marked characteristic point Picture generates the image block centered on characteristic point.
Training set for training convolutional neural networks includes positive sample and negative sample, such as in recognition of face, positive sample Originally it can be the picture of face, the selection of negative sample is just related to problem scenes.According to one embodiment of present invention, it can adopt Characteristic point is used to be in the image block at center as positive sample, the image block using characteristic point far from central point is as negative sample.Fig. 4 Show the schematic diagram of positive sample according to an embodiment of the invention and negative sample.As shown in figure 4, wherein A) be positive sample This, B) it is negative sample.But it since adjacent sample has substantial portion of overlapping region, can be made in training convolutional neural networks At repeating for overlapping region, operational efficiency is damaged.According to one embodiment of present invention, it can be used during characteristic point is in The larger sized image block of the heart is as training set.
Then in step s 320, it is based on image block, generates manual tag figure.
Manual tag figure includes positive sample region, negative sample region and inactive area.Fig. 5 shows according to the present invention one The schematic diagram of the manual tag figure of a embodiment.As shown in figure 5, black portions indicate that negative sample, white portion indicate positive sample, Several points at the center Fig. 5 are all white, and indicating that these pixels can be treated as is the characteristic point to be predicted.This is equivalent to do The data enhancing of positive sample.A circle grey is filled out among white area and black region, indicating that these pixels are not known is not It is the characteristic point to be predicted, these samples can be ignored, be not involved in training.
Then in step S330, by the convolutional neural networks of the manual tag figure generated in step S320 input pre-training In handled, to export prediction label figure.
According to one embodiment of present invention, convolutional neural networks may include process of convolution layer and warp lamination.Convolution Process layer can carry out convolution, activation, pondization processing to the image of input, to obtain predicting the probability of each characteristic point.Instead Convolutional layer can classify to the probability of predicted characteristics point, to export prediction label figure.
Table 1 shows the subnetwork structural parameters of Matching Model according to an embodiment of the invention.
Subnetwork structural parameters in 1 Matching Model of table
As shown in table 1, Matching Model includes process of convolution layer and warp lamination.Process of convolution layer include convolution, activation and Chi Hua.The activation primitives such as ReLU, tanh, sigmoid can be used in active coating, the continuous real value of input can be transformed to 0 and 1 Between output.Network parameter can be obtained by debugging, be not filled when process of convolution wherein, so Matching Model Size can determine the relationship output and input, so that the center of the corresponding output prediction label figure in the center of input picture.For example, The image of 20x20x3 can be input in the Matching Model of pre-training, the prediction label figure output of 14x14x1 will be obtained. Warp lamination can up-sample the probability of the predicted characteristics point of output, to carry out on the characteristic pattern of up-sampling by picture Element classification, final output prediction label figure.The Matching Model that this programme provides can receive the image block input of arbitrary dimension, lead to It crosses and up-sampling is carried out to the probability value of output and is classified pixel-by-pixel, to export prediction label figure.
Finally in step S340, based on the penalty values of loss function between manual tag figure and prediction label figure, training Convolutional neural networks, using the network after training as the Matching Model generated.
It according to one embodiment of present invention, can be based on area pixel point each in manual tag figure and prediction label figure Weight calculates the penalty values of loss function.Then, penalty values are based on, back-propagation algorithm adjustment convolutional neural networks are used Parameter, until penalty values are less than predetermined threshold.
The process of training convolutional neural networks is divided into two stages.First stage is from low level to high-level propagation Stage, i.e. propagated forward stage.Another stage is, when the result that propagated forward obtains with it is expected be not consistent when, by error from The high-level stage for propagate to bottom training, i.e. back-propagation phase.The image block array of input passes through process of convolution Output prediction label figure is obtained after layer, the processing of pond layer, warp lamination;Then it calculates between prediction label figure and manual tag figure The penalty values of the loss function of building;When penalty values are greater than desired value, full articulamentum, down-sampling are adjusted based on gradient descent method The parameter of layer, convolutional layer.Until terminating training when penalty values are equal to or less than desired value.This is to those skilled in the art For belong to contents known, it will not go into details herein.
Matconvnet or pytorch frame can be used to convolutional Neural net in a kind of implementation according to the present invention Network is trained, and wherein Matconvnet is the convolutional neural networks realized with Matlab.The model of pre-training is loaded first, so The loss function in frame is improved afterwards:
Wherein, i is the index of pixel, predictiIndicate prediction label figure, lableiIndicate the picture in manual tag figure Vegetarian refreshments, wiIndicate the weight of pixel, binary cross entropy loss function can be used in loss function loss, indicates negative sample using -1 This, 0 indicates to ignore, and 1 indicates positive sample.If indicating negative sample using the cross entropy loss function of pytorch, 0,1 is indicated Positive sample, -100 indicate to ignore.Cross entropy is also smaller, and the output of model and desired output are closer.
In order to guarantee the data balancing of positive negative sample, increase weight coefficient in loss function.Fig. 6 is shown according to this hair The schematic diagram of the prediction label figure of bright one embodiment.As shown in fig. 6, entire characteristic pattern is 13x13, gray area is 7x7, White is 5 pixels.The number of so negative sample is 13x13-7x7=120, and the number of positive sample is 5.If do not weighted, meeting There is the Loss of Loss and 120 negative sample of 5 positive samples.In this way training when, network more can pay attention to allowing negative sample not Want misclassification.As a result the prediction for being exactly network is essentially all black (being negative sample entirely).Therefore, in addition weight seeks to positive sample Total Loss size and negative sample total Loss size it is almost the same.
It can make the weight of positive sample area pixel point and for 1, the weight of negative sample area pixel point and be 1.With positive sample For this, 0.2*Loss_ positive sample 2+0.2*Loss_ positive sample 3+0.2*Loss_ positive sample 4+0.2*Loss_ positive sample 5= The positive sample of 0.2* (Loss_ positive sample 2+Loss_ positive sample 3+Loss_ positive sample 4+Loss_ positive sample 5)=0.2*5 The Loss of Loss==1 positive sample.The weight of negative sample is 1/120==0.0083.If the training frame used is supported Indicate to ignore the sample using 0 (such as matconvnet) or -100 (such as pytorch) in label figure, then weight The weight of figure grey area is filled out any value and is ok.If it does not, the region of grey must fill out 0 in weight map, to ignore It calculates.
An implementation according to the present invention, using minimum convolutional neural networks, such as input 17x17x3- > 13x13x8- > 11x11x4- > 9x9x1 output, and the optimization by carrying out single-instruction multiple-data stream (SIMD) to matconvnet, effectively Ground improves operational efficiency.
So far, the process of method 300 terminates.300 feature is carried out using the method for deep learning according to the method for the present invention Point matching, the method compared to tradition based on feature and support vector machines can be improved the accuracy and efficiency of Feature Points Matching. And based on more mature convolutional neural networks training framework, the accuracy of Feature Points Matching can be further increased.
It should be noted that the Matching Model that generates above and be limited to human face characteristic point matching, be also equally applicable to cat face, The application scenarios of the Feature Points Matchings such as dog face, it is not limited here.
A kind of method that human face characteristic point is tracked using above-mentioned matching result introduced below.Fig. 7 shows basis The schematic flow chart of the human face characteristic point method for tracing 700 of one embodiment of the invention.
With method 300, the finger for executing method 700 is stored in the program data of calculating equipment 100 as shown in Figure 2 It enables, so that calculating equipment 100 executes this method 700.As shown in fig. 7, method 700 starts from step S710, it is based on shape pair Facial image initialization to be detected, obtains face characteristic point image.
According to one embodiment of present invention, shape can be constructed based on the facial video image frame for having marked characteristic point Model.Wherein, firstly, the characteristic point vector of the characteristic point vector sum present frame facial image based on previous frame facial image, is obtained Obtain change in shape matrix.Then, principal component analysis is carried out to change in shape matrix, obtains the weight vectors of change in shape.Finally, Weight vectors based on change in shape matrix and change in shape, complete the building of shape.
Assuming that there is M picture, every picture has N number of characteristic point, and the coordinate of each characteristic point is assumed to be (xi, yi), a figure Vector x=[the x of the coordinate composition of the N number of characteristic point of on piece1, y1, x2, y2... xn, yn] indicate shape can use it is as follows Formula one indicates:
X=x '+PB
Wherein, x indicates that present frame characteristic point vector, x ' expression previous frame characteristic point vector, P are change in shape matrix.With The characteristic point vector of each frame image subtracts the characteristic point vector of previous frame image, just obtains a change in shape matrix X.Then To XX ' carry out principal component analysis, the decisive ingredient of change in shape is obtained, i.e. feature vector Pj and corresponding characteristic value enter J, K feature vector formed shape transformation matrices P in a manner of arranging discharge before selecting.These feature vectors are exactly all samples in fact The base of this transformation can state any variation in sample.There are change in shape matrix P, the B in above-mentioned formula one that can lead to Following formula two are crossed to obtain:
B=PT(x-x′)。
B is the weight vectors of change in shape, determines which characteristic point plays key effect.So far shape is just completed Building, when giving a weight B a, so that it may face shape is initialized with above-mentioned formula two, so as to the face to acquisition Feature point image carries out local feature region matching.This belongs to contents known to one skilled in the art, refuses herein It repeats.
Then in step S720, multiple image blocks comprising characteristic point are obtained from face characteristic point image.Such as it can To choose the input of the image block using centered on each characteristic point as Matching Model.
Finally in step S730, by the Matching Model after the image block input training of selection, to export the matching of prediction Characteristic point.Wherein Matching Model is generated based on method 300.
So far, the process of method 700 terminates.According to the method for the present invention 700, by using the Matching Model after training as Local expert is fused in limited partial model, replaces pixel characteristic and support vector machines usually used in local expert, makes Obtaining weak characteristic point can accurately be tracked.Particularly with the characteristic point real-time tracing in video flowing, positioning feature point can be improved Anti-jitter ability, improve the stability and accuracy of tracing characteristic points.
To sum up, according to the solution of the present invention, initialization face characteristic point image is obtained by building shape first, so Matching Model is trained afterwards and obtains preferable matching result, so that the Matching Model based on generation carries out human face characteristic point Real-time location tracking.The Stability and veracity of positioning feature point can be improved in the program, improves the efficiency of tracing characteristic points.It is right In the application scenarios of some Real-time Special Effects, it is able to ascend user experience.
A4, the method as described in A3, wherein the step of training convolutional neural networks includes:
Based on the weight of area pixel point each in the manual tag figure and prediction label figure, the loss of loss function is calculated Value;
Based on the penalty values, the parameter of the convolutional neural networks is adjusted using back-propagation algorithm, until penalty values Less than predetermined threshold.
It should be appreciated that in order to simplify the disclosure and help to understand one or more of the various inventive aspects, it is right above In the description of exemplary embodiment of the present invention, each feature of the invention be grouped together into sometimes single embodiment, figure or In person's descriptions thereof.However, the disclosed method should not be interpreted as reflecting the following intention: i.e. claimed hair Bright requirement is than feature more features expressly recited in each claim.More precisely, as the following claims As book reflects, inventive aspect is all features less than single embodiment disclosed above.Therefore, it then follows specific real Thus the claims for applying mode are expressly incorporated in the specific embodiment, wherein each claim itself is used as this hair Bright separate embodiments.
Those skilled in the art should understand that the module of the equipment in example disclosed herein or unit or groups Part can be arranged in equipment as depicted in this embodiment, or alternatively can be positioned at and the equipment in the example In different one or more equipment.Module in aforementioned exemplary can be combined into a module or furthermore be segmented into multiple Submodule.
Those skilled in the art will understand that can be carried out adaptively to the module in the equipment in embodiment Change and they are arranged in one or more devices different from this embodiment.It can be the module or list in embodiment Member or component are combined into a module or unit or component, and furthermore they can be divided into multiple submodule or subelement or Sub-component.Other than such feature and/or at least some of process or unit exclude each other, it can use any Combination is to all features disclosed in this specification (including adjoint claim, abstract and attached drawing) and so disclosed All process or units of what method or apparatus are combined.Unless expressly stated otherwise, this specification is (including adjoint power Benefit require, abstract and attached drawing) disclosed in each feature can carry out generation with an alternative feature that provides the same, equivalent, or similar purpose It replaces.
In addition, it will be appreciated by those of skill in the art that although some embodiments described herein include other embodiments In included certain features rather than other feature, but the combination of the feature of different embodiments mean it is of the invention Within the scope of and form different embodiments.For example, in the following claims, embodiment claimed is appointed Meaning one of can in any combination mode come using.
Various technologies described herein are realized together in combination with hardware or software or their combination.To the present invention Method and apparatus or the process and apparatus of the present invention some aspects or part can take insertion tangible media, such as it is soft The form of program code (instructing) in disk, CD-ROM, hard disk drive or other any machine readable storage mediums, Wherein when program is loaded into the machine of such as computer etc, and is executed by the machine, the machine becomes to practice this hair Bright equipment.
In the case where program code executes on programmable computers, calculates equipment and generally comprise processor, processor Readable storage medium (including volatile and non-volatile memory and or memory element), at least one input unit, and extremely A few output device.Wherein, memory is configured for storage program code;Processor is configured for according to the memory Instruction in the said program code of middle storage executes method of the present invention.
By way of example and not limitation, computer-readable medium includes computer storage media and communication media.It calculates Machine readable medium includes computer storage media and communication media.Computer storage medium storage such as computer-readable instruction, The information such as data structure, program module or other data.Communication media is generally modulated with carrier wave or other transmission mechanisms etc. Data-signal processed passes to embody computer readable instructions, data structure, program module or other data including any information Pass medium.Above any combination is also included within the scope of computer-readable medium.
In addition, be described as herein can be by the processor of computer system or by executing by some in the embodiment The combination of method or method element that other devices of the function are implemented.Therefore, have for implementing the method or method The processor of the necessary instruction of element forms the device for implementing this method or method element.In addition, Installation practice Element described in this is the example of following device: the device be used for implement as in order to implement the purpose of the invention element performed by Function.
As used in this, unless specifically stated, come using ordinal number " first ", " second ", " third " etc. Description plain objects, which are merely representative of, is related to the different instances of similar object, and is not intended to imply that the object being described in this way must Must have the time it is upper, spatially, sequence aspect or given sequence in any other manner.
Although the embodiment according to limited quantity describes the present invention, above description, the art are benefited from It is interior it is clear for the skilled person that in the scope of the present invention thus described, it can be envisaged that other embodiments.Additionally, it should be noted that Language used in this specification primarily to readable and introduction purpose and select, rather than in order to explain or limit Determine subject of the present invention and selects.Therefore, without departing from the scope and spirit of the appended claims, for this Many modifications and changes are obvious for the those of ordinary skill of technical field.For the scope of the present invention, to this Invent done disclosure be it is illustrative and not restrictive, it is intended that the scope of the present invention be defined by the claims appended hereto.

Claims (10)

1. a kind of method for generating Matching Model, the Matching Model is suitable for carrying out feature to the facial image for having marked characteristic point Point matching, the method are suitable for executing in calculating equipment, comprising steps of
Based on the facial image for having marked characteristic point, the image block centered on characteristic point is generated;
Based on described image block, manual tag figure is generated;
It will be handled in the convolutional neural networks of manual tag figure input pre-training, to export prediction label figure;With And
Based on the penalty values of loss function between the manual tag figure and prediction label figure, the convolutional neural networks are trained, Using the Matching Model by the network after training as generation.
2. the method for claim 1, wherein the convolutional neural networks include process of convolution layer and warp lamination,
The process of convolution layer is suitable for carrying out the image block of input convolution, activation, pondization processing, to export predicted characteristics point Probability;
The warp lamination is suitable for classifying to the probability of the predicted characteristics point, to export prediction label figure.
3. the method for claim 1, wherein the manual tag figure includes positive sample region, negative sample region and nothing Imitate region.
4. method as claimed in claim 3, wherein the loss function is binary cross entropy loss function, is based on following public affairs Formula calculates the penalty values of the loss function:
Wherein, i is the index of pixel, predictiIndicate prediction label figure, lableiIndicate the pixel in manual tag figure Point, wiIndicate the weight of pixel.
5. method as claimed in claim 4, wherein the weight of the positive sample area pixel point and be 1, the negative sample area The weight of domain pixel and be 1, the weight of the inactive area is 0.
6. a kind of human face characteristic point method for tracing, suitable for being executed in calculating equipment, wherein the described method includes:
Facial image to be detected is initialized based on shape, obtains face characteristic point image;
Multiple image blocks comprising characteristic point are obtained from the face characteristic point image;
By the Matching Model after the input training of described image block, to export the matching characteristic point of prediction.
Wherein the Matching Model is generated using method according to any one of claims 1 to 5.
7. method as claimed in claim 6, wherein the described method includes:
Based on the facial video image frame for having marked characteristic point, the shape is constructed.
8. the method for claim 7, wherein the step of building shape includes:
The characteristic point vector of characteristic point vector sum present frame facial image based on previous frame facial image obtains change in shape square Battle array;
Principal component analysis is carried out to the change in shape matrix, obtains the weight vectors of change in shape;And
Based on the weight vectors of the change in shape matrix and change in shape, the building of the shape is completed.
9. a kind of calculating equipment, comprising:
One or more processors;With
Memory;
One or more programs, wherein one or more of programs are stored in the memory and are configured as by described one A or multiple processors execute, and one or more of programs include for executing in -8 the methods according to claim 1 The instruction of either method.
10. a kind of computer readable storage medium for storing one or more programs, one or more of programs include instruction, Described instruction is when calculating equipment execution, so that the equipment that calculates executes appointing in method described in -8 according to claim 1 The instruction of one method.
CN201910523891.1A 2019-06-17 2019-06-17 Method for generating matching model and face characteristic point tracking method Active CN110276289B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910523891.1A CN110276289B (en) 2019-06-17 2019-06-17 Method for generating matching model and face characteristic point tracking method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910523891.1A CN110276289B (en) 2019-06-17 2019-06-17 Method for generating matching model and face characteristic point tracking method

Publications (2)

Publication Number Publication Date
CN110276289A true CN110276289A (en) 2019-09-24
CN110276289B CN110276289B (en) 2021-09-07

Family

ID=67960892

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910523891.1A Active CN110276289B (en) 2019-06-17 2019-06-17 Method for generating matching model and face characteristic point tracking method

Country Status (1)

Country Link
CN (1) CN110276289B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111967577A (en) * 2020-07-29 2020-11-20 华北电力大学 Energy internet scene generation method based on variational self-encoder
CN111968028A (en) * 2020-08-14 2020-11-20 北京字节跳动网络技术有限公司 Image generation method, device, equipment and computer readable medium
CN112434965A (en) * 2020-12-04 2021-03-02 广东电力信息科技有限公司 Expert label generation method, device and terminal based on word frequency
CN113111698A (en) * 2020-12-30 2021-07-13 无锡乐骐科技有限公司 Semantic perception loss-based face mark point detection method
CN116631019A (en) * 2022-03-24 2023-08-22 清华大学 Mask suitability detection method and device based on facial image

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104182974A (en) * 2014-08-12 2014-12-03 大连理工大学 A speeded up method of executing image matching based on feature points
CN106056546A (en) * 2016-05-25 2016-10-26 广东工业大学 Image restoration method and device based on sample block matching
CN106373109A (en) * 2016-08-31 2017-02-01 南方医科大学 Medical image modal synthesis method
CN106599830A (en) * 2016-12-09 2017-04-26 中国科学院自动化研究所 Method and apparatus for positioning face key points
CN107563323A (en) * 2017-08-30 2018-01-09 华中科技大学 A kind of video human face characteristic point positioning method
CN107808147A (en) * 2017-11-17 2018-03-16 厦门美图之家科技有限公司 A kind of face Confidence method based on the tracking of real-time face point
CN108090904A (en) * 2018-01-03 2018-05-29 深圳北航新兴产业技术研究院 A kind of medical image example dividing method and device
CN108121950A (en) * 2017-12-05 2018-06-05 长沙学院 A kind of big posture face alignment method and system based on 3D models
KR20180064863A (en) * 2016-12-06 2018-06-15 부산대학교 산학협력단 SMI automatic analysis method of hand-wrist radiation images using deep learning
CN108241854A (en) * 2018-01-02 2018-07-03 天津大学 A kind of deep video conspicuousness detection method based on movement and recall info
CN108805917A (en) * 2018-05-25 2018-11-13 网易(杭州)网络有限公司 Sterically defined method, medium, device and computing device
CN109146845A (en) * 2018-07-16 2019-01-04 中南大学 Head image sign point detecting method based on convolutional neural networks
CN109241982A (en) * 2018-09-06 2019-01-18 广西师范大学 Object detection method based on depth layer convolutional neural networks
CN109376596A (en) * 2018-09-14 2019-02-22 广州杰赛科技股份有限公司 Face matching process, device, equipment and storage medium
CN109558904A (en) * 2018-11-21 2019-04-02 咪咕文化科技有限公司 Classification method, device and the storage medium of image local feature

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104182974A (en) * 2014-08-12 2014-12-03 大连理工大学 A speeded up method of executing image matching based on feature points
CN106056546A (en) * 2016-05-25 2016-10-26 广东工业大学 Image restoration method and device based on sample block matching
CN106373109A (en) * 2016-08-31 2017-02-01 南方医科大学 Medical image modal synthesis method
KR20180064863A (en) * 2016-12-06 2018-06-15 부산대학교 산학협력단 SMI automatic analysis method of hand-wrist radiation images using deep learning
CN106599830A (en) * 2016-12-09 2017-04-26 中国科学院自动化研究所 Method and apparatus for positioning face key points
CN107563323A (en) * 2017-08-30 2018-01-09 华中科技大学 A kind of video human face characteristic point positioning method
CN107808147A (en) * 2017-11-17 2018-03-16 厦门美图之家科技有限公司 A kind of face Confidence method based on the tracking of real-time face point
CN108121950A (en) * 2017-12-05 2018-06-05 长沙学院 A kind of big posture face alignment method and system based on 3D models
CN108241854A (en) * 2018-01-02 2018-07-03 天津大学 A kind of deep video conspicuousness detection method based on movement and recall info
CN108090904A (en) * 2018-01-03 2018-05-29 深圳北航新兴产业技术研究院 A kind of medical image example dividing method and device
CN108805917A (en) * 2018-05-25 2018-11-13 网易(杭州)网络有限公司 Sterically defined method, medium, device and computing device
CN109146845A (en) * 2018-07-16 2019-01-04 中南大学 Head image sign point detecting method based on convolutional neural networks
CN109241982A (en) * 2018-09-06 2019-01-18 广西师范大学 Object detection method based on depth layer convolutional neural networks
CN109376596A (en) * 2018-09-14 2019-02-22 广州杰赛科技股份有限公司 Face matching process, device, equipment and storage medium
CN109558904A (en) * 2018-11-21 2019-04-02 咪咕文化科技有限公司 Classification method, device and the storage medium of image local feature

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
WENHAO WANG 等,: "Improved CT algorithm based on target block division and feature points matching", 《EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING VOLUME 2018》 *
张丽丽 等,: "基于特征点的图像匹配算法研究", 《万方数据知识服务平台》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111967577A (en) * 2020-07-29 2020-11-20 华北电力大学 Energy internet scene generation method based on variational self-encoder
CN111967577B (en) * 2020-07-29 2024-04-05 华北电力大学 Energy Internet scene generation method based on variation self-encoder
CN111968028A (en) * 2020-08-14 2020-11-20 北京字节跳动网络技术有限公司 Image generation method, device, equipment and computer readable medium
CN112434965A (en) * 2020-12-04 2021-03-02 广东电力信息科技有限公司 Expert label generation method, device and terminal based on word frequency
CN113111698A (en) * 2020-12-30 2021-07-13 无锡乐骐科技有限公司 Semantic perception loss-based face mark point detection method
CN113111698B (en) * 2020-12-30 2022-04-01 无锡乐骐科技股份有限公司 Semantic perception loss-based face mark point detection method
CN116631019A (en) * 2022-03-24 2023-08-22 清华大学 Mask suitability detection method and device based on facial image
CN116631019B (en) * 2022-03-24 2024-02-27 清华大学 Mask suitability detection method and device based on facial image

Also Published As

Publication number Publication date
CN110276289B (en) 2021-09-07

Similar Documents

Publication Publication Date Title
CN110276289A (en) Generate the method and human face characteristic point method for tracing of Matching Model
Fang et al. Bottom-up saliency detection model based on human visual sensitivity and amplitude spectrum
US20210343074A1 (en) Generation and presentation of predicted personalized three-dimensional body models
CN110062934A (en) The structure and movement in image are determined using neural network
CN109255769A (en) The training method and training pattern and image enchancing method of image enhancement network
CN110084313A (en) A method of generating object detection model
CN110517278A (en) Image segmentation and the training method of image segmentation network, device and computer equipment
KR101453630B1 (en) Robust object recognition by dynamic modeling in augmented reality
CN109816009A (en) Multi-tag image classification method, device and equipment based on picture scroll product
CN110070072A (en) A method of generating object detection model
CN110163267A (en) A kind of method that image generates the training method of model and generates image
CN110135406A (en) Image-recognizing method, device, computer equipment and storage medium
CN109978792A (en) A method of generating image enhancement model
CN110059605A (en) A kind of neural network training method calculates equipment and storage medium
CN109242961A (en) A kind of face modeling method, apparatus, electronic equipment and computer-readable medium
CN110096964A (en) A method of generating image recognition model
CN110020600A (en) Generate the method for training the data set of face alignment model
CN109978063A (en) A method of generating the alignment model of target object
CN110084253A (en) A method of generating object detection model
CN110414574A (en) A kind of object detection method calculates equipment and storage medium
CN108038823A (en) Image-type becomes the training method of network model, image-type becomes method and computing device
CN111292262B (en) Image processing method, device, electronic equipment and storage medium
CN109902716A (en) A kind of training method and image classification method being aligned disaggregated model
CA3137297C (en) Adaptive convolutions in neural networks
CN112419326B (en) Image segmentation data processing method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant