CN110414378A - A kind of face identification method based on heterogeneous facial image fusion feature - Google Patents

A kind of face identification method based on heterogeneous facial image fusion feature Download PDF

Info

Publication number
CN110414378A
CN110414378A CN201910619187.6A CN201910619187A CN110414378A CN 110414378 A CN110414378 A CN 110414378A CN 201910619187 A CN201910619187 A CN 201910619187A CN 110414378 A CN110414378 A CN 110414378A
Authority
CN
China
Prior art keywords
face
image
facial image
sketch
caricature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910619187.6A
Other languages
Chinese (zh)
Inventor
刘青山
孙玉宝
马强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Information Science and Technology
Original Assignee
Nanjing University of Information Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Information Science and Technology filed Critical Nanjing University of Information Science and Technology
Priority to CN201910619187.6A priority Critical patent/CN110414378A/en
Publication of CN110414378A publication Critical patent/CN110414378A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • G06F18/2193Validation; Performance evaluation; Active pattern learning techniques based on specific statistical tests
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of face identification methods based on heterogeneous facial image fusion feature.Facial image in face database is pre-processed, the image of the fixed size comprising face is cut out;Face of the training based on separation characteristic ajusts model, ajusts to face is carried out by pretreated facial image;Training generates the sketch of confrontation network based on circulation and caricature generates model, and the generation of sketch and caricature facial image is carried out to the facial image after ajusting;The sketch and the progress feature extraction of caricature facial image and fusion that model generates are generated to the true man's image ajusted by pretreatment and face and by sketch and caricature using residual error network, recognition of face is carried out according to fusion feature.Face recognition accuracy rate of the invention is higher, there is preferable identification and robustness.

Description

A kind of face identification method based on heterogeneous facial image fusion feature
Technical field
The invention belongs to technical field of image processing, in particular to a kind of face identification method.
Background technique
After face recognition technology comes into vogue, energy is studied in large quantities and has also been put in recognition of face.Compared with Early face recognition technology, such as Eigenfaces are mainly identified using the geometry feature of face, and face is analyzed Topological relation between the position and component of component is identified.Although this method is simple, final identification is accurate Rate is highly susceptible to the influence of posture and expression shape change.2012 or so, neural network occurred gradually in the visual field of people.Just For current stage, two classes can be mainly divided into for the research of recognition of face: first is that the proposition or improvement of network structure, two It is different the improvement of loss function.
Network structure: after entering into the visual field of people from neural network in 2012, more and more researchers start The research of recognition of face is carried out using deep learning.In 2014, face is become by the DeepFace that Facebook is proposed The bellwether of identification research, and 97.35% accuracy rate is achieved on LFW face database.Subsequent Google is released Then FaceNet using the mapping put on e-learning image to theorem in Euclid space compares two images in theorem in Euclid space The distance of point is compared.The accuracy rate that the method obtains on LFW is 99.63%.
With the continuous intensification of network, DeepID1 carries out recognition of face using the feature of higher-dimension, has extensive well Ability.DeepID2 improves loss function on the basis of DeepID1, reduce in class away from, expand class spacing, So that it is higher to train the accuracy rate that the model come obtains.2014, since VGGNet achieves overnight success on ImageNet, just have Researcher has been applied in recognition of face, is equally also achieved good results.The same period, the proposition of GoogleNet Extraordinary result is achieved in certain tasks in the same contest.Two networks of VGGNet and GoogleNet have one Common feature is exactly that the depth of network is constantly being deepened.Unlike VGGNet, GoogleNet has attempted noveler Network structure, rather than the characteristics of directly succession AlexNet, and model size ratio VGGNet is much smaller, and performance is also opposite For comparative superiority.2015, residual error network achieved champion in the tasks such as the ImageNet classification competed, and achievement is more than VGGNet.The it is proposed of residual error module can greatly reduce the loss in the training process of characteristic information, so that final mark sheet Show that retained information is more abundant.Then, the deformation of a large amount of model has been emerged in large numbers on the basis of residual error network.
Loss function: after deep learning starts the concern by everybody, various loss functions gradually by It proposes.Earliest and the most famous is Softmax loss function, and the later period, much various loss functions were all at it On the basis of deformed.Although sample can correctly be classified using Softmax loss function, it is the absence of in class Away from the constraint with class spacing.So CenterLoss had also been proposed in 2016.The it is proposed of CenterLoss primarily to The inter- object distance of every a kind of sample is constrained, so that every one kind sample is drawn close to center, to improve the accuracy rate of identification.Though Right CenterLoss reduces inter- object distance, and the accuracy rate of identification also has a certain upgrade, but there is no carry out to class spacing Constraint.In order to it is effective while realize two constraint, in 2017, Liu et al. people proposed SphereFace. The it is proposed of the SphereFace modification mainly to loss function in fact, that is, propose A-Softmax loss.In view of When Softmax calculates loss, feature vector and multiplied by weight when, contains the information of angle, sharp in other words It can make the feature learnt that there is the distribution character in angle with Softmax loss.So the proposition of A-Softmax is exactly It makes these angular distribution characteristics more obvious, to realize reduction inter- object distance, expands the effect of class spacing.
In 2018, Wang proposed CosFace, was easier to realize compared to SphereFace, CosFace, reduced The calculation amount of parameter, and the supervision constraint of Softmax loss is eliminated, training process is more succinct and can be realized receipts It holds back.Most importantly, CosFace is significantly improved the performance of model.Although cosine can go to angle, two Person still has apparent difference.In fact, realizing the maximization of classification boundaries in angular region for cosine space With more reasonable explanatory.In angular region, the gap on edge is equivalent to hyperspherical arc distance.Then, Deng proposes ArcFace, especially its loss function: the arm of angle loses (Angular Margin Loss).The loss function is Angled edge is placed in the inside of cosine function, allows the network to the characteristic that more angles are arrived in study.
In addition to other than the research in terms of network structure and loss function two to carry out recognition of face, there are also parts to study The it is proposed that the building and different face characteristics for putting forth effort on face database indicate.The face of different scales and different difficulty is known Other database is organized publication gradually and comes out.But the prior art still around use true man's image carry out recognition of face into Row research.
Summary of the invention
In order to solve the technical issues of above-mentioned background technique is mentioned, the invention proposes one kind to be melted based on heterogeneous facial image The face identification method for closing feature, the research that will use only the progress recognition of face of true man's image have been extended to true man and heterogeneous people The face identification method that face image combines.
In order to achieve the above technical purposes, the technical solution of the present invention is as follows:
A kind of face identification method based on heterogeneous facial image fusion feature, comprising the following steps:
(1) facial image in face database is pre-processed, cuts out the image of the fixed size comprising face;
(2) face of the training based on separation characteristic ajusts model, to the pendulum for carrying out face by pretreated facial image Just;
(3) train based on circulation generate confrontation network sketch and caricature generate model, to the facial image after ajusting into The generation of row sketch and caricature facial image;
(4) mould is generated using residual error network to the true man's image ajusted by pretreatment and face and by sketch and caricature The sketch and caricature facial image that type generates carry out feature extraction and fusion, carry out recognition of face according to fusion feature.
Further, in step (1), Face datection and crucial point location is carried out to facial image first, utilize key Dot position information is cut out face.
Further, in step (1), the method cut out is as follows:
Face is ajusted first with the eyes position detected, so that eyes are located in same horizontal line;Then Fixed eyes position, selects the image of predetermined size.
Further, in step (2), the face based on separation characteristic ajusts model using encoder and decoder The effect for undertaking generator judges the true and false of image using arbiter and identifies the identity and posture information of image;By mentioning The Attitude estimation in the posture code and arbiter of decoder is supplied, face characteristic is separated with posture.
Further, in step (2), the majorized function that the face based on separation characteristic ajusts model is as follows:
In above formula, VD(D, G) and VG(D, G) is objective function, and D indicates arbiter, G presentation code decoder, E () table Showing mathematic expectaion, x indicates true picture, and y indicates identity label and posture label, and c indicates posture, and z indicates noise, x, y~ Pdata(x,y)Indicate x and y from given data, z~Pz(z) noise of Gaussian distributed, c~P are indicatedz(c) it indicates to giving Fixed posture is within the scope of preset, function G presentation code decoder,WithRespectively indicate differentiation image True and false property, posture information and identity information.
Further, in step (3), the circulation generates the total losses function L (G of confrontation networkAB,GBA,DA,DB) such as Under:
L(GAB,GBA,DA,DB)=LGAN(GAB,DB,A,B)
+LGAN(GBA,DA,B,A)
+λLcyc(GAB,GBA)
In above formula, A, B indicate both modalities which, GABIndicate that the image under mode A is converted to the generator of image under mode B, GBAIndicate that the image under mode B is converted to the generator of image under mode A, DAIt indicates to differentiate whether image is figure under mode A The arbiter of picture, DBIndicate to differentiate image whether be image under mode B arbiter;LGAN(GAB,DB, A, B) and indicate arbiter DBLoss function:
LGAN(GAB,DB, A, B) and=EB~Pdata (b)(logDB(b))
+EA~Pdata (a)[log(1-DB(GAB(a)))]
In above formula, a is the image under mode A, and b is the image under mode B, and E () indicates mathematic expectaion, a~Pdata (a), b~Pdata (b) respectively indicates the distribution that two kinds of pictures under both modalities which meet data set;LGAN(GBA,DA, B, A) and table Show arbiter DALoss function:
LGAN(GBA,DA, B, A) and=EA~Pdata (a)(logDA(a))
+EB~Pdata (b)[log(1-DA(GBA(b)))]
Lcyc(GAB,GBA) indicate circulation loss function:
In above formula,Indicate GBA(GAB(a)),Indicate GAB(GBA(b));
λ is Lcyc(GAB,GBA) weight coefficient.
Further, in step (3), sketch face database is selected, by the sketch image in sketch face database Input circulation generates confrontation network and is trained, and obtains sketch and generates model, generates the sketch image that model obtains using sketch The overall situation and profile information of image can be retained;Caricature face database is selected, by the cartoon image in caricature face database Input circulation generates confrontation network and is trained, and obtains caricature and generates model, generates the cartoon image that model obtains using caricature The local message of image can be retained.
By adopting the above technical scheme bring the utility model has the advantages that
The present invention ajusts method using the face based on separation characteristic and generates the heterogeneous of confrontation network generation based on circulation Facial image carries out recognition of face, the experiment proved that, under identical experiment condition, method proposed by the present invention is in VGGFace2 Recognition accuracy on face database is higher than other recognition methods on the database.
Detailed description of the invention
Fig. 1 is basic flow chart of the invention;
Fig. 2 is that the face based on separation characteristic ajusts flow chart;
Fig. 3 makes a living into confrontation network and the face based on separation characteristic ajusts network diagram;
Fig. 4 is that the face based on separation characteristic ajusts the result figure on Multi-PIE and VGGFace2;
Fig. 5 is that circulation generates confrontation Principles of Network schematic diagram;
Fig. 6 is the heterogeneous Face image synthesis frame diagram that confrontation network is generated based on circulation;
Fig. 7 is that the sketch that confrontation network is generated based on circulation and caricature generate result figure;
Fig. 8 is the recognition of face flow chart based on heterogeneous facial image fusion feature.
Specific embodiment
Below with reference to attached drawing, technical solution of the present invention is described in detail.
As shown in Figure 1, a kind of face identification method based on heterogeneous facial image fusion feature that the present invention designs, step It is as follows:
1, the facial image in face database is pre-processed, cuts out the image of the fixed size comprising face;
2, face of the training based on separation characteristic ajusts model, to the pendulum for carrying out face by pretreated facial image Just;
3, the sketch and caricature generation model for generating confrontation network based on circulation are trained, the facial image after ajusting is carried out The generation of sketch and caricature facial image;
4, mould is generated using residual error network to the true man's image ajusted by pretreatment and face and by sketch and caricature The sketch and caricature facial image that type generates carry out feature extraction and fusion, carry out recognition of face according to fusion feature.
This implementation uses VGGFace2 and LFW face database.Due to the image in VGGFace2 and LFW face database Complex, most image all contains more human body and complex background information, therefore the face in later period for convenience The research of identification needs to carry out all pictures certain pretreatment operation.Due in VGGFace2 face database Given the coordinate information of the face key point of each picture, so can directly be carried out for VGGFace2 face database The cutting of facial image.Since LFW face database does not provide corresponding key point information, the present invention utilizes MTCNN Face datection and crucial point location are carried out to the facial image in LFW, using the key point location information detected to face into Row is cut.
Due to the method that difference has more different face to cut at present, the invention firstly uses the eyes positions detected Face is ajusted, so that eyes are on the same horizontal line.Then the position for fixing eyes, selects 256 × 256 sizes Image.The information of face is only contained in the facial image for obtain after pretreatment operation to all images, it can Largely improve the accuracy rate of identification.
Since there are more facial images to include posture information, and posture in VGGFace2 and LFW face database Information is one kind most commonly seen in the significant challenge of current recognition of face.Most people's face recognition method is due to joined difference The accuracy rate of posture, identification can reduce by 10% or so, but for people, only reduce little by little.Common solution face There are two ways to pose problem occurred in identification: the study of face ajusted with specific characteristic.The present invention utilizes two kinds of sides The combination of method proposes the face based on separation characteristic and ajusts method, and the process of method is as shown in Figure 2.
What the DRGAN in Fig. 2 was indicated is that the face indicated based on separation characteristic ajusts method.Compared to GAN network, it is based on The face of separation characteristic is ajusted there are two innovative point: first is that the effect of generator is undertaken using the structure of coding and decoding, second is that logical The Attitude estimation being supplied in the posture code and discriminator of decoder is crossed, face characteristic is separated with posture.GAN network Shown in (a) in structure such as Fig. 3, shown in (b) in network structure such as Fig. 3 that the face indicated based on separation characteristic is ajusted.
From figure 3, it can be seen that the face based on separation characteristic ajusts network encoder and decoder instead of in GAN Generator part, and arbiter needs the work completed in addition to judging the true and false of image, also to identify the body of image Part and posture information.The majorized function that face based on separation characteristic is ajusted can be repaired on the basis of the loss function of GAN Change:
In above formula, E () indicates mathematic expectaion, and x indicates that true picture, y indicate that identity label and posture label, c indicate Posture, z indicate noise, x, y~Pdata(x,y)Indicate x and y from given data, z~Pz(z) Gaussian distributed is indicated Noise, c~Pz(c) it indicates to given posture within the scope of preset, function G presentation code decoder, WithIt respectively indicates for differentiating the true and false property of image, posture information and identity information.Face based on separation information ajusts method In used encoder, decoder and arbiter network parameter etc. as shown in table 1.
Table 1
It is Multi-PIE for trained face database.The database contains posture, illumination and expression, is current One of more common face database, and the same person contains 15 kinds of different postures in the database.It randomly selects 9 kind postures of 337 people under natural expression and good illumination condition, image of the side face of left and right within ± 60 °.It is selecting In 337 people taken, the image of 200 people is as training sample, and the image of remaining 137 people is as test sample.It is instructing When white silk, the normal view that all true man's facial images are 100 × 100 all in accordance with size is aligned first, then The region of grab sample 96 × 96 in image after the alignment, to realize that data enhance.Using batch training, by 64 people Face image is as a batch.All convolutional layer weights are 0 all in accordance with mean value in network, the normal distribution that standard deviation is 0.02 It is initialized.The method choice Adam of optimization, and initial learning rate is 0.0002, momentum 0.5.By very long instruction After white silk, shown in (a) in result such as Fig. 4 that the part face obtained on test set is ajusted.
From Fig. 4 (a) as can be seen that needing the human face posture ajusted larger, difficulty is also higher, thus the people after ajusting There are the distortion of partial region and discontinuously in face.Appoint since the present embodiment will carry out final recognition of face on VGGFace2 Business, it is therefore desirable to which the biggish face of attitudes vibration in VGGFace2 is ajusted.It does so not just to extract more rationally Face characteristic, and be the generation of corresponding heterogeneous facial image for convenience.It is ajusted pair using the face based on separation characteristic Attitudes vibration biggish facial image progress face in part is ajusted in VGGFace2, the result after partially ajusting such as (b) institute in Fig. 4 Show.
From (b) in Fig. 4 as can be seen that face is by effect after ajusting or pretty good, but part picture occurs The case where distortion, there are also the diplopias for occurring blue in the picture of part.Although facial image after ajusting there are a little problem, It is that can preferably be remained in the profile and module information of face, and component all positions of the face after ajusting are several It is identical.
After ajusting to facial image, next need to complete is to carry out sketch and caricature figure to facial image The generation of picture.Since the image of GAN generates feature, more and more researchs have appeared in study and the transfer neck of the style of image Domain, wherein most commonly seen is that circulation generates confrontation network.Image under different modalities can be carried out mutual turn by the network Change, for example, can by an image horse and zebra carry out mutual conversion.Mutually turn between image under different modalities Change is exactly a kind of mapping relations in study between mode in fact.It is corresponding the object of the present invention is to be generated using true man's image Heterogeneous image, i.e. study true man's image to sketch and true man's image to the mapping function between caricature.Common GAN is unidirectional life At, although can recycle generation confrontation network by the true and false of the input picture that arbiter differentiates and then generate mutually. Just because of the characteristics of mutually generating, but optimize the mapping space of picture under two mode, to prevent different input To the result of identical output.The principle signal that circulation generates confrontation network is as shown in Figure 5.
Circulation, which generates confrontation network, can be converted to the image under A mode the image under B mode.In order to realize the two The process of generation then needs two generator GABAnd GBA, while also containing two arbiter DAAnd DBTo differentiate the true of image Vacation, as shown in (a) in Fig. 5.The two generators are mutually converted the image of A mode and B mode respectively.Under A mode Image a pass through GABObtain the image G of a generationAB(a), then by the image G of generationAB(a) pass through GBAObtain an A mould Image G under stateBA(GAB(a)) it, is expressed asAs shown in (b) in Fig. 5.Same b passes through the image that two generators obtain Referred to asAs shown in (c) in Fig. 5.Due to there is the process mutually generated in model, so needing two arbiters to carry out Differentiate the true and false property for generating image.Other than calculating primary generation and obtaining the true and false property of image, it is also necessary to secondary by two Differentiated at obtained image.In other words, in the circulation generated at one, the loss of required calculating includes input and life At circulation loss caused by the loss of the true and false property of image and secondary generation.
It can be regarded as the superposition at two networks GAN since circulation generates confrontation network, so for first arbiter DBIt is primarily used to differentiate GAB(a) the true and false property of the true picture b and under mode B, it may be assumed that
LGAN(GAB,DB, A, B) and=EB~Pdata (b)(logDB(b))
+EA~Pdata (a)[log(1-DB(GAB(a)))]
In above formula, E () indicates mathematic expectaion, and a~Pdata (a), b~Pdata (b) are respectively indicated under both modalities which Two kinds of pictures meet the distribution of data set.GABMain purpose be to allow the image G of generationAB(a) appear more like is mode Image under B.DBPurpose be then to tell G as far as possibleAB(a) and b on earth that be only the image under mode B.Therefore, excellent Changing target can indicate are as follows:
Similarly, for second arbiter DALoss function can indicate are as follows:
LGAN(GBA,DA, B, A) and=EA~Pdata (a)(logDA(a))
+EB~Pdata (b)[log(1-DA(GBA(b)))]
GBAMain purpose be to allow the image G of generationBA(b) appearing more like is image under mode A, to make Arbiter can not resolution image it is true and false.DAPurpose be then to tell G as far as possibleBA(b) and a on earth that be only mode Image under A, i.e., judgement input picture is true and false.Corresponding optimization aim can indicate are as follows:
It but is it cannot be guaranteed that the mapping learnt is so that each independent input can just with confrontation loss A satisfied output is obtained, and is easy to appear the problem of gradient disappears during training.It is reflected to further decrease The possibility variation space of function is penetrated, invention introduces a significantly more efficient loss functions: circulation loss function.Circulation loss Purpose be, differentiate a and GBA(GAB(a)) whether it is under mode A, and determines b and GAB(GBA(b)) whether all Come from mode B.Therefore, circulation loss function can indicate are as follows:
So generating in confrontation network in circulation, total loss function selected by the present invention can be indicated are as follows:
L(GAB,GBA,DA,DB)=LGAN(GAB,DB,A,B)
+LGAN(GBA,DA,B,A)
+λLcyc(GAB,GBA)
In above formula, λ indicates the weight of circulation loss function.
The present invention will carry out the generation of heterogeneous facial image on the basis of the above, generate the heterogeneous of confrontation network based on circulation Face image synthesis frame is as shown in Figure 6.Carry out model training before, the present embodiment arranged from Hong Kong Chinese University three The training that part sketch picture generates model for sketch is selected in a sketch face database, from current biggish two caricatures The training that part caricature picture generates model for caricature is selected in face database.Selected true man when two model trainings Part true man image of the picture in sketch face database.In order to facilitate differentiation, for generating the two of sketch and caricature A model is referred to as PSGAN and PCGAN.The sketch and caricature result generated by two models is as shown in Figure 7.
First row in Fig. 7 is the true man's image chosen from VGGFace2 face database at random, second be classified as by The correspondence caricature facial image that PCGAN is generated, third are classified as the correspondence sketch facial image generated by PSGAN.It is not difficult to find out that Using train caricature that PCGAN and PSGAN are respectively obtained and sketch image can preferably to show heterogeneous image exclusive The characteristics of, sketch image can preferably retain the overall situation and profile information of image, and caricature can then reinforce local features, protect More local messages are stayed.Although sketch have from the point of view of the result of generation write it is a little intermittent, can be certain Reduce the influence of illumination and background in degree.From the sketch images in Fig. 7 just it can be clearly seen that.Compared to sketch, caricature There is fuzzy the case where even losing in profile.It can be inferred that the reason of causing caricature to be distorted by analysis, be exactly used to train Caricature facial image style be not it is fixed, there are more different-styles.
The present embodiment choose VGGFace2 face database carry out recognition of face research, therefore carry out feature extraction and Before heterogeneous Face image synthesis, need to carry out the biggish facial image of attitudes vibration in database to ajust operation.Passing through After face is ajusted, all side faces either biggish picture of attitudes vibration all becomes corresponding positive face picture.Utilize instruction Two experienced models PSGAN and PCGAN generate two kinds of heterogeneous facial images of sketch and caricature respectively.
Since mainly study is the face identification method based on heterogeneous facial image fusion feature to the present invention, for extracting The residual error network of true man, caricature and sketch feature are referred to as PResNet, CResNet and SResNet.Experiment is broadly divided into three Group: 1) the true man's image ajusted and cut only is used only;2) using trained PSGAN model in step 3) by all sanctions Then the corresponding sketch of the Face image synthesis sheared uses true man and sketch image simultaneously;3) the PCGAN mould of step 3) is utilized Then the corresponding caricature of all Face image synthesis reduced is used true man and caricature by type simultaneously.In order to facilitate differentiation, this Three groups of experiments are referred to as ResNet-Fusion (1), ResNet-Fusion (2) and ResNet-Fusion (3).Present invention choosing Use the associated losses of Softmax and CenterLoss as the loss function of model training, and the weight λ of CenterLoss To indicate.It is as shown in Figure 8 based on face and the total frame of recognition of face of corresponding heterogeneous facial image fusion feature.
There is no the links that addition face is ajusted in Fig. 8, but actually before carrying out model training, attitudes vibration is larger Face picture be all sent to the face based on separation characteristic and ajust network and carry out the operation ajusted of face, in addition, in figure It is all trained before that PSGAN and PCGAN, which respectively indicates two heterogeneous Face image synthesis models,.
The present invention chooses the extraction that residual error network carries out feature to three kinds of true man, sketch and caricature images respectively.Through remarkable Picture size after face is ajusted is 256 × 256.Since picture pixels are larger, when leading to training pattern, a batch Picture number it is less, training process is longer.In order to accelerate the training of model, the input figure that the present embodiment will be used to test The size of picture is adjusted to 112 × 112.Although picture becomes smaller, certain detailed information may be lost, this is greatly greatly The fast training and optimization process of model, can save a large amount of time.For extracting the net of feature in recognition of face frame Network parameter is as shown in table 2.
Table 2
In table 2, n indicates the number of training sample.In Pool1 and Conv2, Pool2 and Conv3, Pool3 and Conv4, There are 1,2,5,3 residual error modules between Pool4 and FC6 respectively.Convolution used in each residual error module is all 3 × 3/1/1, i.e., The convolution kernel size of residual error module is 3, and step-length 1 is filled with 1.Therefore for image after by residual error module, size can't It changes.
3,310,000 facial images are shared in VGGFace2 face database.If choosing more than 300 ten thousand face figures every time As carry out three experiment if, require a great deal of time, thus the present embodiment by the training set of the database according to classification It is divided into three parts, every portion includes nearly 1,000,000 facial images, and every portion is all used for an experiment.In a network Not shared parameter between residual error network for extracting different images feature.The weight initial value of all residual error modules meets Variance is 0, the normal distribution that standard deviation is 0.01, and the initial learning rate of model is 0.01, attenuation coefficient 0.1.Utilize instruction The model perfected is tested in test set data, and the recognition accuracy for finally obtaining three experiments is as shown in table 3.
Table 3
From table 3 it is observed that only the recognition accuracy of true man's characteristics of image is 88.2%, heterogeneous facial image has been merged The final discrimination of two groups of experiments improve 1.6% and 2.0% on the basis of only with true man's image respectively.It can be seen that After having merged heterogeneous facial image feature, the character representation that final e-learning arrives has better identification.Compared to this On database from the point of view of other experimental results, the case where not utilizing pre-training model and final characteristic dimension is all 2048 Under, it is higher that fusion feature proposed by the present invention obtains face recognition accuracy rate.In order to further illustrate differentiating for fusion feature Trained three models have also been carried out face verification by property and robustness, the present embodiment respectively on LFW face database Experiment.Feature extraction and fusion are carried out to this 6000 pairs of images respectively using trained three models.Then it calculates separately Cosine similarity between this 6000 pairs is ranked up these similarities according to sequence from big to small, before statistics filters out The error sample quantity occurred in 3000 pairs of facial images, calculates the accuracy rate of face verification.The result of final test such as 4 institute of table Show.
Table 4
Method Accuracy rate (%)
Combined Joint Bayesian 92.42
Roof-Hog 92.80
Tom-vs-Pete+Attribute 93.30
FR+FCN 96.45
DeepID 2 99.47
SphereFace 99.42
PingAn AI Lab 99.80
ResNet-Fusion(1) 95.52
ResNet-Fusion(2) 96.08
ResNet-Fusion(3) 96.15
The first seven result is the face verification accuracy rate that method some other on the database obtains respectively in table 4, from As can be seen that after the feature for having merged two kinds of heterogeneous facial images of sketch and caricature, the accuracy rate of verifying is table 4 96.08% and 96.15%, the accuracy rate than not merging heterogeneous facial image feature is distinguished high by 0.56% and 0.63%.Pass through After the test for carrying out face verification on LFW face database, illustrate merging sketch and caricature feature to a certain extent Character representation later has preferable robustness.
In conclusion the face identification method proposed by the present invention based on heterogeneous facial image fusion feature using sketch and The characteristics of caricature, so that final mark sheet is shown with preferable identification and robustness.
Embodiment is merely illustrative of the invention's technical idea, and this does not limit the scope of protection of the present invention, it is all according to Technical idea proposed by the present invention, any changes made on the basis of the technical scheme are fallen within the scope of the present invention.

Claims (7)

1. a kind of face identification method based on heterogeneous facial image fusion feature, which comprises the following steps:
(1) facial image in face database is pre-processed, cuts out the image of the fixed size comprising face;
(2) face of the training based on separation characteristic ajusts model, ajusts to face is carried out by pretreated facial image;
(3) sketch and caricature generation model for generating confrontation network based on circulation are trained, element is carried out to the facial image after ajusting Retouch the generation with caricature facial image;
(4) model life is generated using residual error network to the true man's image ajusted by pretreatment and face and by sketch and caricature At sketch and caricature facial image carry out feature extraction and fusion, recognition of face is carried out according to fusion feature.
2. the face identification method according to claim 1 based on heterogeneous facial image fusion feature, which is characterized in that in step Suddenly in (1), Face datection and crucial point location is carried out to facial image first, face is cut using key point location information It cuts out.
3. the face identification method according to claim 2 based on heterogeneous facial image fusion feature, which is characterized in that in step Suddenly in (1), the method cut out is as follows:
Face is ajusted first with the eyes position detected, so that eyes are located in same horizontal line;Then it fixes Eyes position selects the image of predetermined size.
4. the face identification method according to claim 1 based on heterogeneous facial image fusion feature, which is characterized in that in step Suddenly in (2), the face based on separation characteristic ajusts the effect that model undertakes generator using encoder and decoder, uses Arbiter judges the true and false of image and identifies the identity and posture information of image;By be supplied to decoder posture code and Attitude estimation in arbiter separates face characteristic with posture.
5. the face identification method according to claim 1 based on heterogeneous facial image fusion feature, which is characterized in that in step Suddenly in (2), the majorized function that the face based on separation characteristic ajusts model is as follows:
In above formula, VD(D, G) and VG(D, G) is objective function, and D indicates that arbiter, G presentation code decoder, E () indicate number It hopes in term, x indicates that true picture, y indicate that identity label and posture label, c indicate posture, and z indicates noise, x, y~Pdata(x,y) Indicate x and y from given data, z~Pz(z) noise of Gaussian distributed, c~P are indicatedz(c) it indicates to given posture Within the scope of preset, function G presentation code decoder,WithIt respectively indicates and differentiates the true and false property of image, Posture information and identity information.
6. the face identification method according to claim 1 based on heterogeneous facial image fusion feature, which is characterized in that in step Suddenly in (3), the circulation generates the total losses function L (G of confrontation networkAB,GBA,DA,DB) it is as follows:
L(GAB,GBA,DA,DB)=LGAN(GAB,DB,A,B)+LGAN(GBA,DA,B,A)+λLcyc(GAB,GBA)
In above formula, A, B indicate both modalities which, GABIndicate that the image under mode A is converted to the generator of image under mode B, GBATable Show that the image under mode B is converted to the generator of image under mode A, DAIt indicates to differentiate whether image is image under mode A Arbiter, DBIndicate to differentiate image whether be image under mode B arbiter;LGAN(GAB,DB, A, B) and indicate arbiter DB's Loss function:
LGAN(GAB,DB, A, B) and=EB~Pdata (b)(logDB(b))+EA~Pdata (a)[log(1-DB(GAB(a)))]
In above formula, a is the image under mode A, and b is the image under mode B, and E () indicates mathematic expectaion, a~Pdata (a), b ~Pdata (b) respectively indicates the distribution that two kinds of pictures under both modalities which meet data set;LGAN(GBA,DA, B, A) and it indicates to differentiate Device DALoss function:
LGAN(GBA,DA, B, A) and=EA~Pdata (a)(logDA(a))+EB~Pdata (b)[log(1-DA(GBA(b)))]
Lcyc(GAB,GBA) indicate circulation loss function:
In above formula,Indicate GBA(GAB(a)),Indicate GAB(GBA(b));
λ is Lcyc(GAB,GBA) weight coefficient.
7. the face identification method based on heterogeneous facial image fusion feature described in any one of -6 according to claim 1, It is characterized in that, in step (3), selects sketch face database, the sketch image in sketch face database is inputted into circulation It generates confrontation network to be trained, obtains sketch and generate model, generating the sketch image that model obtains using sketch can retain The overall situation and profile information of image;Caricature face database is selected, the cartoon image in caricature face database is inputted into circulation It generates confrontation network to be trained, obtains caricature and generate model, generating the cartoon image that model obtains using caricature can retain The local message of image.
CN201910619187.6A 2019-07-10 2019-07-10 A kind of face identification method based on heterogeneous facial image fusion feature Pending CN110414378A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910619187.6A CN110414378A (en) 2019-07-10 2019-07-10 A kind of face identification method based on heterogeneous facial image fusion feature

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910619187.6A CN110414378A (en) 2019-07-10 2019-07-10 A kind of face identification method based on heterogeneous facial image fusion feature

Publications (1)

Publication Number Publication Date
CN110414378A true CN110414378A (en) 2019-11-05

Family

ID=68360876

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910619187.6A Pending CN110414378A (en) 2019-07-10 2019-07-10 A kind of face identification method based on heterogeneous facial image fusion feature

Country Status (1)

Country Link
CN (1) CN110414378A (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110879985A (en) * 2019-11-18 2020-03-13 西南交通大学 Anti-noise data face recognition model training method
CN111079549A (en) * 2019-11-22 2020-04-28 杭州电子科技大学 Method for recognizing cartoon face by using gating fusion discrimination features
CN111093029A (en) * 2019-12-31 2020-05-01 深圳云天励飞技术有限公司 Image processing method and related device
CN111160264A (en) * 2019-12-30 2020-05-15 中山大学 Cartoon figure identity recognition method based on generation of confrontation network
CN111243051A (en) * 2020-01-08 2020-06-05 浙江省北大信息技术高等研究院 Portrait photo-based stroke generating method, system and storage medium
CN111275005A (en) * 2020-02-21 2020-06-12 腾讯科技(深圳)有限公司 Drawn face image recognition method, computer-readable storage medium and related device
CN111275778A (en) * 2020-01-08 2020-06-12 浙江省北大信息技术高等研究院 Face sketch generating method and device
CN111274994A (en) * 2020-02-13 2020-06-12 腾讯科技(深圳)有限公司 Cartoon face detection method and device, electronic equipment and computer readable medium
CN111401193A (en) * 2020-03-10 2020-07-10 海尔优家智能科技(北京)有限公司 Method and device for obtaining expression recognition model and expression recognition method and device
CN111539287A (en) * 2020-04-16 2020-08-14 北京百度网讯科技有限公司 Method and device for training face image generation model
CN111797891A (en) * 2020-05-21 2020-10-20 南京大学 Unpaired heterogeneous face image generation method and device based on generation countermeasure network
CN111832622A (en) * 2020-06-11 2020-10-27 国家计算机网络与信息安全管理中心 Method and system for identifying ugly pictures of specific figures
CN112446317A (en) * 2020-11-23 2021-03-05 四川大学 Heterogeneous face recognition method and device based on feature decoupling
CN112489173A (en) * 2020-12-11 2021-03-12 杭州格像科技有限公司 Method and system for generating portrait photo cartoon
CN112633154A (en) * 2020-12-22 2021-04-09 云南翼飞视科技有限公司 Method and system for converting heterogeneous face feature vectors
CN112907692A (en) * 2021-04-09 2021-06-04 吉林大学 SFRC-GAN-based sketch-to-face reconstruction method
CN114120391A (en) * 2021-10-19 2022-03-01 哈尔滨理工大学 Multi-pose face recognition system and method thereof
WO2022068451A1 (en) * 2020-09-30 2022-04-07 北京字节跳动网络技术有限公司 Style image generation method and apparatus, model training method and apparatus, device, and medium
CN117894083A (en) * 2024-03-14 2024-04-16 中电科大数据研究院有限公司 Image recognition method and system based on deep learning

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107239766A (en) * 2017-06-08 2017-10-10 深圳市唯特视科技有限公司 A kind of utilization resists network and the significantly face of three-dimensional configuration model ajusts method
CN107437077A (en) * 2017-08-04 2017-12-05 深圳市唯特视科技有限公司 A kind of method that rotation face based on generation confrontation network represents study
CN107577985A (en) * 2017-07-18 2018-01-12 南京邮电大学 The implementation method of the face head portrait cartooning of confrontation network is generated based on circulation
WO2019056000A1 (en) * 2017-09-18 2019-03-21 Board Of Trustees Of Michigan State University Disentangled representation learning generative adversarial network for pose-invariant face recognition
CN109978792A (en) * 2019-03-28 2019-07-05 厦门美图之家科技有限公司 A method of generating image enhancement model

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107239766A (en) * 2017-06-08 2017-10-10 深圳市唯特视科技有限公司 A kind of utilization resists network and the significantly face of three-dimensional configuration model ajusts method
CN107577985A (en) * 2017-07-18 2018-01-12 南京邮电大学 The implementation method of the face head portrait cartooning of confrontation network is generated based on circulation
CN107437077A (en) * 2017-08-04 2017-12-05 深圳市唯特视科技有限公司 A kind of method that rotation face based on generation confrontation network represents study
WO2019056000A1 (en) * 2017-09-18 2019-03-21 Board Of Trustees Of Michigan State University Disentangled representation learning generative adversarial network for pose-invariant face recognition
CN109978792A (en) * 2019-03-28 2019-07-05 厦门美图之家科技有限公司 A method of generating image enhancement model

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
LUAN TRAN ET AL.: "Disentangled Representation Learning GAN for Pose-Invariant Face Recognition", 《2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR)》 *
QIANG MA ET AL.: "Facial Components-based Representation for Caricature Face Recognition", 《INTERNATIONAL JOURNAL OF PERFORMABILITY ENGINEERING》 *
查文锦: "资源受限条件下的人脸识别研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110879985B (en) * 2019-11-18 2022-11-11 西南交通大学 Anti-noise data face recognition model training method
CN110879985A (en) * 2019-11-18 2020-03-13 西南交通大学 Anti-noise data face recognition model training method
CN111079549A (en) * 2019-11-22 2020-04-28 杭州电子科技大学 Method for recognizing cartoon face by using gating fusion discrimination features
CN111079549B (en) * 2019-11-22 2023-09-22 杭州电子科技大学 Method for carrying out cartoon face recognition by utilizing gating fusion discrimination characteristics
CN111160264A (en) * 2019-12-30 2020-05-15 中山大学 Cartoon figure identity recognition method based on generation of confrontation network
CN111160264B (en) * 2019-12-30 2023-05-12 中山大学 Cartoon character identity recognition method based on generation countermeasure network
CN111093029A (en) * 2019-12-31 2020-05-01 深圳云天励飞技术有限公司 Image processing method and related device
CN111243051A (en) * 2020-01-08 2020-06-05 浙江省北大信息技术高等研究院 Portrait photo-based stroke generating method, system and storage medium
CN111275778B (en) * 2020-01-08 2023-11-21 杭州未名信科科技有限公司 Face simple drawing generation method and device
CN111275778A (en) * 2020-01-08 2020-06-12 浙江省北大信息技术高等研究院 Face sketch generating method and device
CN111243051B (en) * 2020-01-08 2023-08-18 杭州未名信科科技有限公司 Portrait photo-based simple drawing generation method, system and storage medium
CN111274994A (en) * 2020-02-13 2020-06-12 腾讯科技(深圳)有限公司 Cartoon face detection method and device, electronic equipment and computer readable medium
CN111274994B (en) * 2020-02-13 2022-08-23 腾讯科技(深圳)有限公司 Cartoon face detection method and device, electronic equipment and computer readable medium
CN111275005A (en) * 2020-02-21 2020-06-12 腾讯科技(深圳)有限公司 Drawn face image recognition method, computer-readable storage medium and related device
CN111401193B (en) * 2020-03-10 2023-11-28 海尔优家智能科技(北京)有限公司 Method and device for acquiring expression recognition model, and expression recognition method and device
CN111401193A (en) * 2020-03-10 2020-07-10 海尔优家智能科技(北京)有限公司 Method and device for obtaining expression recognition model and expression recognition method and device
CN111539287B (en) * 2020-04-16 2023-04-07 北京百度网讯科技有限公司 Method and device for training face image generation model
CN111539287A (en) * 2020-04-16 2020-08-14 北京百度网讯科技有限公司 Method and device for training face image generation model
CN111797891A (en) * 2020-05-21 2020-10-20 南京大学 Unpaired heterogeneous face image generation method and device based on generation countermeasure network
CN111832622A (en) * 2020-06-11 2020-10-27 国家计算机网络与信息安全管理中心 Method and system for identifying ugly pictures of specific figures
WO2022068451A1 (en) * 2020-09-30 2022-04-07 北京字节跳动网络技术有限公司 Style image generation method and apparatus, model training method and apparatus, device, and medium
CN112446317A (en) * 2020-11-23 2021-03-05 四川大学 Heterogeneous face recognition method and device based on feature decoupling
CN112489173A (en) * 2020-12-11 2021-03-12 杭州格像科技有限公司 Method and system for generating portrait photo cartoon
CN112633154A (en) * 2020-12-22 2021-04-09 云南翼飞视科技有限公司 Method and system for converting heterogeneous face feature vectors
CN112907692A (en) * 2021-04-09 2021-06-04 吉林大学 SFRC-GAN-based sketch-to-face reconstruction method
CN114120391A (en) * 2021-10-19 2022-03-01 哈尔滨理工大学 Multi-pose face recognition system and method thereof
CN114120391B (en) * 2021-10-19 2024-07-12 哈尔滨理工大学 Multi-pose face recognition system and method thereof
CN117894083A (en) * 2024-03-14 2024-04-16 中电科大数据研究院有限公司 Image recognition method and system based on deep learning

Similar Documents

Publication Publication Date Title
CN110414378A (en) A kind of face identification method based on heterogeneous facial image fusion feature
CN110427867B (en) Facial expression recognition method and system based on residual attention mechanism
CN109492529A (en) A kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion
Lin et al. Face recognition/detection by probabilistic decision-based neural network
Geetha et al. A vision based dynamic gesture recognition of indian sign language on kinect based depth images
CN108182409A (en) Biopsy method, device, equipment and storage medium
CN104850825A (en) Facial image face score calculating method based on convolutional neural network
CN107844760A (en) Three-dimensional face identification method based on curved surface normal direction component map Neural Networks Representation
CN104915658B (en) A kind of emotion component analyzing method and its system based on emotion Distributed learning
Ferrer et al. Static and dynamic synthesis of Bengali and Devanagari signatures
Das et al. Sign language recognition using facial expression
CN108960171B (en) Method for converting gesture recognition into identity recognition based on feature transfer learning
CN114241458B (en) Driver behavior recognition method based on attitude estimation feature fusion
Kamgar-Parsi et al. Aircraft detection: A case study in using human similarity measure
Kwaghe et al. A deep learning approach for detecting face mask using an improved Yolo-V2 with Squeezenet
Shurid et al. Bangla sign language recognition and sentence building using deep learning
CN102819730A (en) Method for extracting and recognizing facial features
Cerezo et al. Real-time facial expression recognition for natural interaction
CN110826459A (en) Migratable campus violent behavior video identification method based on attitude estimation
Luqin A survey of facial expression recognition based on convolutional neural network
Thakar et al. Sign Language to Text Conversion in Real Time using Transfer Learning
Rehman et al. A Real-Time Approach for Finger Spelling Interpretation Based on American Sign Language Using Neural Networks
Zhao et al. Fully automatic face detection and facial feature points extraction using local Gabor filter bank and PCA
Ch et al. Automatic system for driver drowsiness detection system using deep learning
CN111178141A (en) LSTM human body behavior identification method based on attention mechanism

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20191105