CN107194376A - Mask fraud convolutional neural networks training method and human face in-vivo detection method - Google Patents
Mask fraud convolutional neural networks training method and human face in-vivo detection method Download PDFInfo
- Publication number
- CN107194376A CN107194376A CN201710478892.XA CN201710478892A CN107194376A CN 107194376 A CN107194376 A CN 107194376A CN 201710478892 A CN201710478892 A CN 201710478892A CN 107194376 A CN107194376 A CN 107194376A
- Authority
- CN
- China
- Prior art keywords
- layers
- convolutional layer
- image
- max
- convolutional
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/40—Spoof detection, e.g. liveness detection
- G06V40/45—Detection of the body part being alive
Landscapes
- Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Image Analysis (AREA)
Abstract
The present invention relates to face In vivo detection field, in particular to a kind of mask fraud convolutional neural networks training method and human face in-vivo detection method.The mask fraud convolutional neural networks training method that the present invention is provided includes:Build training set;Obtain the image in training set;Face in detection image;Be normalized after face is cut feeding convolutional neural networks in, convolutional neural networks include input layer, multiple convolutional layers, ReLU layers, pooling layers of max, full articulamentum, Dropout layers and SoftmaxWithLoss layers;Convolutional neural networks are trained.By training the convolutional neural networks based on fraud mask, fraud mask is learnt, mask fraud attack is resisted during for face In vivo detection protection is provided, protect the safety of recognition of face.
Description
Technical field
The present invention relates to face In vivo detection field, trained in particular to a kind of mask fraud convolutional neural networks
Method and human face in-vivo detection method.
Background technology
Easily obtained due to face, it is contactless, easy to use the advantages of, recognition of face is liked by people, extensively
Apply in gate inhibition's entry-exit management system, login system etc..With the popularization of smart machine, face is easily obtained by other people, is passed through
Mask, photo or video are attacked face identification system.Therefore, in face identification system, face In vivo detection becomes
Obtain extremely important, be the indispensable means for ensureing user security, be also the premise that recognition of face is widely popularized.
The content of the invention
In view of this, the purpose of the embodiment of the present invention be to provide a kind of mask fraud convolutional neural networks training method and
Human face in-vivo detection method, to provide a kind of human face in-vivo detection method for tackling mask fraud.
To achieve the above object, the technical scheme that the embodiment of the present invention is used is as described below:
In a first aspect, the embodiments of the invention provide a kind of mask fraud convolutional neural networks training method, methods described
Including:Training set is built, the training set includes mask fraud video and true man's video;Obtain mask fraud video and true man regard
Image in frequency;Face in described image is detected by Face datection algorithm;The face is cut;To going out after cutting
Facial image is normalized;Facial image after the normalization is inputted to convolutional neural networks, the convolution god
Through network include input layer, multiple convolutional layers, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and
SoftmaxWithLoss layers;Facial image after the normalization is carried out random cropping by the input layer, by the people after cutting
Face image flip horizontal, obtains the image of specific pixel, first that the image of the specific pixel is sent into multiple convolutional layers
Individual convolutional layer;First convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max-
Pooling layers, the image of the specific pixel obtains the volume with first convolutional layer after first convolutional layer
The characteristic image of product core respective amount, described ReLU layers makes partial nerve member be output as 0, causes openness, the max-
Pooling layers are compressed to the characteristic image, extract principal character, and the characteristic image enters second convolutional layer;Institute
Stating second convolutional layer includes multiple convolution kernels, and second convolutional layer connects one ReLU layers and max-pooling layers, institute
State characteristic image and feature with the convolution kernel respective amount of second convolutional layer is obtained after second convolutional layer
Image, described ReLU layers makes partial nerve member be output as 0, causes openness, and the characteristic image is by second max-
After pooling layers and then enter the 3rd convolutional layer;3rd convolutional layer includes multiple convolution kernels, the 3rd convolution
Layer connection one ReLU layer and max-pooling layers, the characteristic image obtained after the 3rd convolutional layer with it is described
The characteristic image of the convolution kernel respective amount of 3rd convolutional layer, described ReLU layers makes partial nerve member be output as 0, causes sparse
Property, the characteristic image is after the 3rd max-pooling layers and then enters the 4th convolutional layer;4th convolutional layer
Including multiple convolution kernels, the 4th convolutional layer connects one ReLU layers and max-pooling layers, and then connection two connects entirely
Layer is connect, each full articulamentum is sequentially connected one ReLU layers and dropout layers, and described ReLU layers is output as partial nerve member
0, cause openness, the characteristic image obtains the convolution kernel with the 4th convolutional layer after the 4th convolutional layer
The characteristic image of respective amount, after the 4th max-pooling layers so that enter two full articulamentums and with it is described entirely
Articulamentum connection ReLU layers and dropout layers, dropout layers by training when, allow at random convolutional neural networks it is some imply
The weight of node layer does not work, and a full articulamentum and SoftmaxWithLoss layers is sequentially connected after dropout layers, finally
One full articulamentum exports the scoring event of the facial image, and score is input into last layer of SoftmaxWithLoss layers,
Obtain belonging to probability and the loss of this iteration that mask is faked with true man.
Further, one BatchNorm layers and one are sequentially connected between each described convolutional layer and ReLU layer
Scale layers, the described BatchNorm layers output standardization by neuron is 0 to average, and variance is 1, the Scale layers of reverse
The BatchNorm layers of destruction to feature.
Further, the step of structure training set includes:Training set is built according to predetermined ratio and checking collects, it is described
Training set and checking collection include mask fraud video and true man's video.
Further, going out the step of facial image is normalized after described pair of cutting includes:To going out people after cutting
Face image normalizes to 256 × 256 pixels.
Further, the input layer includes tri- passages of RGB, and size is 256 × 256, and training set is divided into 1000
Batch, each batch sizes are 128, and the pixel of the image of the specific pixel is 227 × 227.
Further, first convolutional layer includes 96 convolution kernels, and the size of each convolution kernel is 7 × 7, convolution behaviour
The step-length of work is 2, the characteristic image that 96 pixels are 111 × 111 is obtained after first convolutional layer, by first max-
After pooling layers, the characteristic image that 96 pixels are 55 × 55 is obtained, described first max-pooling layer of size is 3 ×
3, step-length is 2.
Further, second convolutional layer includes 256 convolution kernels, and the size of each convolution kernel is 5 × 5, convolution
The step-length of operation is 2, the characteristic image that 256 pixels are 26 × 26 is obtained after second convolutional layer, by second
After max-pooling layers, the characteristic image that 256 pixels are 26 × 26, described second max-pooling layers of size are obtained
For 3 × 3, step-length is 1, and Boundary filling is 1.
Further, the 3rd convolutional layer includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border
Supplement 1, the step-length of convolution operation is 1, and the characteristic image that 192 pixels are 26 × 26, warp are obtained after the 3rd convolutional layer
Cross after the 3rd max-pooling layers, obtain the characteristic image that 192 pixels are 26 × 26, the 3rd max-pooling
The size of layer is 3 × 3, and step-length is 1, and Boundary filling is 1.
Further, the 4th convolutional layer includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border
Supplement 1, the step-length of convolution operation is 1, and the characteristic image that 192 pixels are 26 × 26, warp are obtained after the 4th convolutional layer
Cross after the 4th max-pooling layers, obtain the characteristic image that 192 pixels are 13 × 13, the 4th max-pooling
The size of layer is 3 × 3, and step-length is 2.
Second aspect, the embodiments of the invention provide a kind of human face in-vivo detection method faked based on mask, the side
Method includes:Obtain facial image;Face in described image is detected by Face datection algorithm;The face is cut;
It is normalized to going out facial image after cutting;Facial image after the normalization is inputted to convolutional neural networks,
The convolutional neural networks include input layer, multiple convolutional layers, ReLU layers, max-pooling layers, full articulamentum, Dropout layers
With Softmax layers;Facial image after the normalization is carried out random cropping by the input layer, by the facial image after cutting
Flip horizontal, obtains the image of specific pixel, first convolution that the image of the specific pixel is sent into multiple convolutional layers
Layer;First convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max-pooling
Layer, the image of the specific pixel obtains corresponding with the convolution kernel of first convolutional layer after first convolutional layer
The characteristic image of quantity, described ReLU layers makes partial nerve member be output as 0, causes openness, and described max-pooling layers to institute
State characteristic image to be compressed, extract principal character, the characteristic image enters second convolutional layer;Second convolutional layer
Including multiple convolution kernels, second convolutional layer connects one ReLU layers and max-pooling layers, and the characteristic image passes through
The characteristic image with the convolution kernel respective amount of second convolutional layer is obtained after second convolutional layer, it is described ReLU layers
Partial nerve member is output as 0, cause openness, the characteristic image is after second max-pooling layers and then enters
3rd convolutional layer;3rd convolutional layer includes multiple convolution kernels, the 3rd convolutional layer connect one ReLU layer with
Max-pooling layers, the characteristic image obtains the convolution with the 3rd convolutional layer after the 3rd convolutional layer
The characteristic image of core respective amount, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image warp
Cross after the 3rd max-pooling layers and then enter the 4th convolutional layer;4th convolutional layer includes multiple convolution kernels, institute
State the 4th convolutional layer and connect one ReLU layers and max-pooling layers, then two full articulamentums of connection, each full articulamentum
Be sequentially connected one ReLU layers and dropout layers, the characteristic image obtained after the 4th convolutional layer with it is described
The characteristic image of the convolution kernel respective amount of 4th convolutional layer, described ReLU layers makes partial nerve member be output as 0, causes sparse
Property, the characteristic image after the 4th max-pooling layers so that enter two full articulamentums and with the full connection
The ReLU layers of layer connection with dropout layers, dropout layers by training when, allow some hidden layer sections of convolutional neural networks at random
The weight of point does not work, and a full articulamentum and Softmax layers is sequentially connected after dropout layers, last full articulamentum
The scoring event of the facial image is exported, score is input to last layer of Softmax layers, obtains belonging to mask fraud and true
The probability of people.
Mask fraud convolutional neural networks training method and human face in-vivo detection method provided in an embodiment of the present invention, pass through
The convolutional neural networks based on fraud mask are trained, fraud mask is learnt, mask fraud is resisted when being recognition of face and is attacked
Offer protection is provided, the safety of recognition of face is protected.
To enable the above objects, features and advantages of the present invention to become apparent, preferred embodiment cited below particularly, and coordinate
Appended accompanying drawing, is described in detail below.
Brief description of the drawings
In order to illustrate the technical solution of the embodiments of the present invention more clearly, below will be attached to what is used required in embodiment
Figure is briefly described, it will be appreciated that the following drawings illustrate only certain embodiments of the present invention, therefore is not construed as pair
The restriction of scope, for those of ordinary skill in the art, on the premise of not paying creative work, can also be according to this
A little accompanying drawings obtain other related accompanying drawings.
Fig. 1 is the schematic flow sheet of mask fraud convolutional neural networks training method provided in an embodiment of the present invention.
Fig. 2 is the training flow chart of convolutional neural networks.
Fig. 3 is the schematic diagram for the convolutional neural networks model trained in the embodiment of the present invention.
Fig. 4 is a kind of flow chart of human face in-vivo detection method faked based on mask provided in an embodiment of the present invention.
Embodiment
To make the purpose, technical scheme and advantage of the embodiment of the present invention clearer, below in conjunction with the embodiment of the present invention
In accompanying drawing, the technical scheme in the embodiment of the present invention is clearly and completely described, it is clear that described embodiment is
A part of embodiment of the present invention, rather than whole embodiments.The present invention implementation being generally described and illustrated herein in the accompanying drawings
The component of example can be arranged and designed with a variety of configurations.
Therefore, the detailed description of embodiments of the invention below to providing in the accompanying drawings is not intended to limit claimed
The scope of the present invention, but be merely representative of the present invention selected embodiment.Based on the embodiment in the present invention, this area is common
The every other embodiment that technical staff is obtained under the premise of creative work is not made, belongs to the model that the present invention is protected
Enclose.
It should be noted that:Similar label and letter represents similar terms in following accompanying drawing, therefore, once a certain Xiang Yi
It is defined in individual accompanying drawing, then it further need not be defined and explained in subsequent accompanying drawing.
Fig. 1 is refer to, is the flow signal of mask fraud convolutional neural networks training method provided in an embodiment of the present invention
Figure.This method comprises the following steps:
Step S110, builds training set, and training set includes mask fraud video and true man's video.
When training set does not include checking collection, step S101 also includes building training set according to predetermined ratio and checking collects,
The training set and checking collection include mask fraud video and true man's video.The ratio of training set and checking collection is freely configured,
For example can be 8:2 or 7:3, the embodiment of the present invention is not limited this.Mask fraud video is the video for the mask recorded, should
Mask is high with true man's human face similarity degree, and the scene that simulation is attacked with mask fraud, true man's video regards for the true man's of recording
Frequently.
Step S120, obtains the image in mask fraud video and true man's video.
Multiple images are taken out respectively from mask fraud video and true man's video, such as take out a figure every 3 frames or 5 frames
Picture, then a video take out altogether:Totalframes/3 or totalframes/5 image.
Step S130, the face in described image is detected by Face datection algorithm.
Face datection algorithm can be existing general Face datection algorithm, such as adaboost.
Step S140, cuts to face.
During being cut to face, using the algorithm of facial modeling, the precise area of face is found, is obtained
To accurate face frame, the face frame is expanded, it is general to expand 1.8 times, the area in the corresponding artwork of face frame after will be enlarged by
Domain, which is cut out, to be come, and cutting out the face come both includes face, also including certain background, and certain background there are beneficial to classification
True and false video.
Step S150, is normalized to going out facial image after cutting.
Facial image after cutting passes through normalized, normalizes to 256 × 256 pixels.
Step S160, the facial image after normalization is inputted to convolutional neural networks, and the convolutional neural networks include defeated
Enter layer, multiple convolutional layers, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and SoftmaxWithLoss layers.
The convolutional neural networks are trained by step S170.
Wherein, the facial image after normalization initially enters input layer, refer to Fig. 2, is the training of convolutional neural networks
Process, comprises the following steps:
Facial image after normalization is carried out random cropping by step S171, input layer, by the facial image water after cutting
Flat upset, obtains the image of specific pixel, first convolutional layer that the image of specific pixel is sent into multiple convolutional layers.
Wherein, the image of input layer includes tri- passages of RGB, and size is 256 × 256, and training set is divided into 1000
Batch, each batch sizes are 128.The pixel of the image of specific pixel is 227 × 227, and the image in the batch is upset
The positive negative ratio 1 of image pattern that order is sent into first convolutional layer, the batch:1, and belong to the face figure of same person
Picture, to reduce due to the difference that different macroscopic features are caused, makes e-learning to due to the substantive characteristics between true man and fraud,
Reduce difference in class.
Step S172, first convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max-
Pooling layers, the image of specific pixel obtains the convolution kernel respective amount with first convolutional layer after first convolutional layer
Characteristic image, described ReLU layers makes partial nerve member be output as 0, causes openness, and max-pooling layers are entered to characteristic image
Row compression, extracts principal character, and characteristic image enters second convolutional layer.
In the present embodiment, the number of the convolution kernel of first convolutional layer is 96, and the size of each convolution kernel is 7 × 7,
The step-length of convolution operation is 2, and the characteristic image that 96 pixels are 111 × 111, first are obtained after first convolutional layer
Max-pooling layers of size is 3 × 3, and step-length is 2, after first max-pooling layers, and it is 55 to obtain 96 pixels
× 55 characteristic image.As preferred, one BatchNorm layers and one are sequentially connected between each convolutional layer and ReLU layers
It is individual Scale layers, by BatchNorm layers, Scale layers, ReLU layers do not change characteristic image size.When depth network layer
Too much, signal and gradient are less and less, and deep layer is difficult to train, referred to as gradient disperse, it is also possible to increasing, also referred to as
Gradient is exploded, and is 0 by the BatchNorm layers of output standardization by neuron to average, variance is that 1, Scale layers of reverse are described
The BatchNorm layers of destruction to feature, to overcome deep neural network to be difficult to the disadvantage trained.Because depth network is in training
When, each layer of input data distribution is always to change, because above one layer of parameter renewal can cause next one
When the distribution of the input data of layer, BatchNorm and Scale can solve training, the data distribution in network intermediate layer changes
The problem of change, BatchNorm layers are by the way that the output standardization of neuron is arrived:Average is 0, and variance is 1, is passed through
After BatchNorm layers, all neurons all standardize to a kind of distribution, but export-restriction is in the distribution meeting again of the variance 1 of average 0
So that the ability to express of network dies down, the feature that front layer learns also is destroyed, so adding Scale layers, is learned by adding
The reconstruction parameter practised, reverses destruction to feature, the program of reverse by model training when adjust.So both neuron is done
Standardization retains the feature acquired of front layer again, because input becomes stable, promotes the convergence and to a certain extent of model
Prevent over-fitting.Pass through successively yardstick normalizing, it is to avoid gradient is disappeared and gradient is overflowed, and accelerates convergence, while conduct
A kind of Regularization Technique also improves generalization ability;Make activation primitive compared to using sigmoid tanh, it is computationally intensive, reversely pass
Broadcast when seeking error gradient, derived function amount is also very big, and sigmoid tanh functions are easily saturated, and gradient disappearance feelings occur
Condition, i.e., close to during convergence, conversion is too slow, causes information to lose.ReLU layers can make partial nerve member be output as 0, cause dilute
Property is dredged, over-fitting is not only alleviated, also overcoming gradient closer to real neuronal activation model disappears, not unsupervised
Pre-training (i.e. make by first hidden layer of training network, the network parameter values that retraining second ... is finally trained with these
For the initial value of overall network parameter) in the case of dramatically speed up convergence compared to sigmoid and tanh activation primitives.
Step S173, second convolutional layer includes multiple convolution kernels, and second convolutional layer connects one ReLU layers and max-
Pooling layers, characteristic image obtains the feature with the convolution kernel respective amount of second convolutional layer after second convolutional layer
Image, described ReLU layers makes partial nerve member be output as 0, causes openness, and the characteristic image is by second max-
After pooling layers and then enter the 3rd convolutional layer.
Second convolutional layer includes 256 convolution kernels, and the size of each convolution kernel is 5 × 5, and the step-length of convolution operation is 2,
The characteristic image that 256 pixels are 26 × 26 is obtained after second convolutional layer, ReLU layers make partial nerve member be output as 0,
Cause openness, characteristic image obtains the characteristic image that 256 pixels are 26 × 26 after second max-pooling layers,
Second max-pooling layers of size is 3 × 3, and step-length is 1, and Boundary filling is 1.
Step S174, the 3rd convolutional layer includes multiple convolution kernels, and the 3rd convolutional layer connects a max-pooling
Layer, characteristic image obtains the characteristic image with the convolution kernel respective amount of the 3rd convolutional layer after the 3rd convolutional layer,
ReLU layers make partial nerve member be output as 0, cause openness, and characteristic image is after the 3rd max-pooling layers and then enters
Enter the 4th convolutional layer.
3rd convolutional layer includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border supplement 1, convolution operation
Step-length be 1, obtained after the 3rd convolutional layer 192 pixels be 26 × 26 characteristic image, by the 3rd max-
After pooling layers, the characteristic image that 192 pixels are 26 × 26 is obtained, the 3rd max-pooling layers of size is 3 × 3,
Step-length is 1, and Boundary filling is 1.
Step S175, the 4th convolutional layer includes multiple convolution kernels, and the 4th convolutional layer connects a max-pooling
Layer, then connects two full articulamentums, and each full articulamentum is sequentially connected one ReLU layers and dropout layers, and ReLU layers make
Partial nerve member is output as 0, causes openness, and characteristic image obtains the volume with the 4th convolutional layer after the 4th convolutional layer
The characteristic image of product core respective amount, after the 4th maxpooling layers and then into two full articulamentums and with connecting entirely
Connect layer connection ReLU layers with dropout layers.
4th convolutional layer includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border supplement 1, convolution operation
Step-length be 1, obtained after the 4th convolutional layer 192 pixels be 26 × 26 characteristic image, by the 4th max-
After pooling layers, the characteristic image that 192 pixels are 13 × 13 is obtained, the 4th max-pooling layers of size is 3 × 3,
Step-length is 2.
When S176, dropout layers of step is by training, allow at random some hidden layer nodes of convolutional neural networks weight not
Work, be sequentially connected a full articulamentum and SoftmaxWithLoss layers after dropout layers, last full articulamentum is defeated
Go out the scoring event of facial image, score be input to last layer of SoftmaxWithLoss layers, obtain belonging to mask fake and
The probability of true man and the loss of current iteration.
Dropout layers eliminate the simultaneous adaptation reduced between neuron node, enhance generalization ability, dropout layers
During by training pattern, the idle mode of weight of some hidden layer nodes of network is allowed to prevent model over-fitting, canonical at random
Change means, improve generalization ability.In training process, SoftmaxWithLoss layers include Softmax layers and multidimensional Logistic
Loss layers, Softmax layers scoring event above is mapped as belonging to each class probability, be followed by a multidimensional Logistic
Loss layers, what is obtained here is the loss of current iteration.One layer is merged into Softmax layers and Loss layers of multidimensional Logistic
Stabilization numerically is ensure that, two full articulamentums are output as 512 dimensional features.
Fig. 3 is refer to, is the schematic diagram of the convolutional neural networks model in the embodiment of the present invention in training process.The present invention
The mask fraud convolutional neural networks training method that embodiment is provided is right by training the convolutional neural networks based on fraud mask
Fraud mask is learnt, and mask fraud attack is resisted during for recognition of face and provides protection, the safety of recognition of face is protected.
Fig. 4 is refer to, the embodiment of the present invention additionally provides a kind of human face in-vivo detection method faked based on mask, the party
Method comprises the following steps:
Step S210, obtains facial image.
The facial image of acquisition can be the face figure obtained by the camera of the equipment such as gate inhibition, computer or monitor
Picture.
Step S220, passes through the face in Face datection algorithm detection image.
Step S230, cutting is carried out by face
Step S240, is normalized to going out facial image after cutting.
Step S250, the facial image after normalization is inputted to convolutional neural networks, convolutional neural networks include input
Layer, multiple convolutional layers, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and Softmax layers.
Step S260, the facial image is identified.
Wherein, step S260 identification process is carried out according to the convolutional neural networks that above-described embodiment is trained, to obtain people
The true and false probability of face image.Due to making statement to detailed process and principle in the aforementioned embodiment, here is omitted.
In summary, the embodiments of the invention provide a kind of mask fraud convolutional neural networks training method and face live body
Detection method, by training the convolutional neural networks based on fraud mask, learns to fraud mask, is supported when being recognition of face
Imperial mask, which fakes to attack, provides protection, protects the safety of recognition of face.
, can also be by another way in embodiment provided herein, it should be understood that disclosed method
Realize.Embodiments described above is only schematical, for example, the flow chart and block diagram in accompanying drawing are shown according to this hair
Architectural framework in the cards, function and the operation of the method and computer program product of bright embodiment.At this point, flow
Each square frame in figure or block diagram can represent a part for a module, program segment or code, the module, program segment or generation
A part for code includes one or more executable instructions for being used to realize defined logic function.It should also be noted that at some
As in the implementation of replacement, the function of being marked in square frame can also be with different from the order marked in accompanying drawing generation.
For example, two continuous square frames can essentially be performed substantially in parallel, they can also be performed in the opposite order sometimes, this
Depending on involved function.It is also noted that each square frame and block diagram and/or flow in block diagram and/or flow chart
The combination of square frame in figure, can be realized with the special hardware based system of defined function or action is performed, or
It can be realized with the combination of specialized hardware and computer instruction.
In addition, each functional module in each embodiment of the invention can integrate to form an independent portion
Point or modules individualism, can also two or more modules be integrated to form an independent part.
If the function is realized using in the form of software function module and is used as independent production marketing or in use, can be with
It is stored in a computer read/write memory medium.Understood based on such, technical scheme is substantially in other words
The part contributed to prior art or the part of the technical scheme can be embodied in the form of software product, the meter
Calculation machine software product is stored in a storage medium, including some instructions are to cause a computer equipment (can be individual
People's computer, server, or network equipment etc.) perform all or part of step of each of the invention embodiment methods described.
And foregoing storage medium includes:USB flash disk, mobile hard disk, read-only storage (ROM, Read-Only Memory), arbitrary access are deposited
Reservoir (RAM, Random Access Memory), magnetic disc or CD etc. are various can be with the medium of store program codes.Need
Illustrate, herein, such as first and second or the like relational terms be used merely to by an entity or operation with
Another entity or operation make a distinction, and not necessarily require or imply between these entities or operation there is any this reality
The relation or order on border.Moreover, term " comprising ", "comprising" or its any other variant are intended to the bag of nonexcludability
Contain, so that process, method, article or equipment including a series of key elements are not only including those key elements, but also including
Other key elements being not expressly set out, or also include for this process, method, article or the intrinsic key element of equipment.
In the absence of more restrictions, the key element limited by sentence "including a ...", it is not excluded that including the key element
Process, method, article or equipment in also there is other identical element.
The preferred embodiments of the present invention are the foregoing is only, are not intended to limit the invention, for the skill of this area
For art personnel, the present invention can have various modifications and variations.Within the spirit and principles of the invention, that is made any repaiies
Change, equivalent substitution, improvement etc., should be included in the scope of the protection.It should be noted that:Similar label and letter exists
Similar terms is represented in following accompanying drawing, therefore, once being defined in a certain Xiang Yi accompanying drawing, is then not required in subsequent accompanying drawing
It is further defined and explained.
Claims (10)
1. a kind of mask fraud convolutional neural networks training method, it is characterised in that methods described includes:
Training set is built, the training set includes mask fraud video and true man's video;
Obtain the image in mask fraud video and true man's video;
Face in described image is detected by Face datection algorithm;
The face is cut;
It is normalized to going out facial image after cutting;
Facial image after the normalization is inputted to convolutional neural networks, the convolutional neural networks include input layer, many
Individual convolutional layer, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and SoftmaxWithLoss layers;
Facial image after the normalization is carried out random cropping by the input layer, and the facial image level after cutting is turned over
Turn, obtain the image of specific pixel, first convolutional layer that the image of the specific pixel is sent into multiple convolutional layers;
First convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max-
Pooling layers, the image of the specific pixel obtains the volume with first convolutional layer after first convolutional layer
The characteristic image of product core respective amount, described ReLU layers makes partial nerve member be output as 0, causes openness, the max-
Pooling layers are compressed to the characteristic image, extract principal character, and the characteristic image enters second convolutional layer;
Second convolutional layer includes multiple convolution kernels, and second convolutional layer connects one ReLU layers and max-
Pooling layers, the characteristic image obtains the convolution kernel pair with second convolutional layer after second convolutional layer
The characteristic image of quantity is answered, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image passes through
After two max-pooling layers and then enter the 3rd convolutional layer;
3rd convolutional layer includes multiple convolution kernels, and the 3rd convolutional layer connects one ReLU layers and max-
Pooling layers, the characteristic image obtains the convolution kernel pair with the 3rd convolutional layer after the 3rd convolutional layer
The characteristic image of quantity is answered, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image passes through
After three max-pooling layers and then enter the 4th convolutional layer;
4th convolutional layer includes multiple convolution kernels, and the 4th convolutional layer connects one ReLU layers and max-
Pooling layers, two full articulamentums are then connected, each full articulamentum is sequentially connected one ReLU layers and dropout layers, institute
State characteristic image and feature with the convolution kernel respective amount of the 4th convolutional layer is obtained after the 4th convolutional layer
Image, described ReLU layers makes partial nerve member be output as 0, causes openness, and the characteristic image is by the 4th max-
After pooling layers so enter two full articulamentums and the ReLU layers being connected with the full articulamentum and dropout layers;
Dropout layers by training when, allow the weight of some hidden layer nodes of convolutional neural networks not work at random, dropout
A full articulamentum and SoftmaxWithLoss layers are sequentially connected after layer, last full articulamentum exports the face figure
The scoring event of picture, last layer of SoftmaxWithLoss layers is input to by score, obtain belonging to mask fake and true man it is general
The loss of rate and this iteration.
2. mask fraud convolutional neural networks training method according to claim 1, it is characterised in that each described volume
It is sequentially connected one BatchNorm layers and one Scale layers between lamination and ReLU layers, described BatchNorm layers by nerve
The output standardization of member is 0 to average, and variance is 1, and described Scale layers is reversed the BatchNorm layers of destruction to feature.
3. mask fraud convolutional neural networks training method according to claim 1, it is characterised in that the structure training
The step of collection, includes:
Training set is built according to predetermined ratio and checking collects, and the training set and checking collection include mask fraud video and true man
Video.
4. mask fraud convolutional neural networks training method according to claim 1, it is characterised in that after described pair cuts
Going out the step of facial image is normalized includes:
256 × 256 pixels are normalized to going out facial image after cutting.
5. mask fraud convolutional neural networks training method according to claim 1, it is characterised in that the input layer bag
Containing tri- passages of RGB, size is 256 × 256, training set is divided into 1000 batch, each batch sizes are 128, described
The pixel of the image of specific pixel is 227 × 227.
6. the mask fraud convolutional neural networks training method according to claim any one of 1-5, it is characterised in that described
First convolutional layer includes 96 convolution kernels, and the size of each convolution kernel is 7 × 7, and the step-length of convolution operation is 2, by first
The characteristic image that 96 pixels are 111 × 111 is obtained after individual convolutional layer, after first max-pooling layers, 96 are obtained
Pixel is 55 × 55 characteristic image, and described first max-pooling layers of size is 3 × 3, and step-length is 2.
7. mask fraud convolutional neural networks training method according to claim 6, it is characterised in that second volume
Lamination includes 256 convolution kernels, and the size of each convolution kernel is 5 × 5, and the step-length of convolution operation is 2, by second convolutional layer
The characteristic image that 256 pixels are 26 × 26 is obtained afterwards, and after second max-pooling layers, it is 26 to obtain 256 pixels
× 26 characteristic image, described second max-pooling layers of size is 3 × 3, and step-length is 1, and Boundary filling is 1.
8. mask fraud convolutional neural networks training method according to claim 7, it is characterised in that the 3rd volume
Lamination includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border supplement 1, and the step-length of convolution operation is 1, by the
The characteristic image that 192 pixels are 26 × 26 is obtained after three convolutional layers, after the 3rd max-pooling layers, 192 are obtained
Individual pixel is 26 × 26 characteristic image, and described 3rd max-pooling layers of size is 3 × 3, and step-length is 1, Boundary filling
For 1.
9. mask fraud convolutional neural networks training method according to claim 8, it is characterised in that the 4th volume
Lamination includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border supplement 1, and the step-length of convolution operation is 1, by the
The characteristic image that 192 pixels are 26 × 26 is obtained after four convolutional layers, after the 4th max-pooling layers, 192 are obtained
Individual pixel is 13 × 13 characteristic image, and described 4th max-pooling layers of size is 3 × 3, and step-length is 2.
10. a kind of human face in-vivo detection method faked based on mask, it is characterised in that methods described includes:
Obtain facial image;
Face in described image is detected by Face datection algorithm;
The face is cut;
It is normalized to going out facial image after cutting;
Facial image after the normalization is inputted to convolutional neural networks, the convolutional neural networks include input layer, many
Individual convolutional layer, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and Softmax layers;
Facial image after the normalization is carried out random cropping by the input layer, and the facial image level after cutting is turned over
Turn, obtain the image of specific pixel, first convolutional layer that the image of the specific pixel is sent into multiple convolutional layers;
First convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max-
Pooling layers, the image of the specific pixel obtains the volume with first convolutional layer after first convolutional layer
The characteristic image of product core respective amount, described ReLU layers makes partial nerve member be output as 0, causes openness, the max-
Pooling layers are compressed to the characteristic image, extract principal character, and the characteristic image enters second convolutional layer;
Second convolutional layer includes multiple convolution kernels, and second convolutional layer connects one ReLU layers and max-
Pooling layers, the characteristic image obtains the convolution kernel pair with second convolutional layer after second convolutional layer
The characteristic image of quantity is answered, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image passes through
After two max-pooling layers and then enter the 3rd convolutional layer;
3rd convolutional layer includes multiple convolution kernels, and the 3rd convolutional layer connects one ReLU layers and max-
Pooling layers, the characteristic image obtains the convolution kernel pair with the 3rd convolutional layer after the 3rd convolutional layer
The characteristic image of quantity is answered, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image passes through
After three max-pooling layers and then enter the 4th convolutional layer;
4th convolutional layer includes multiple convolution kernels, and the 4th convolutional layer connects one ReLU layers and max-
Pooling layers, two full articulamentums are then connected, each full articulamentum is sequentially connected one ReLU layers and dropout layers, institute
State characteristic image and feature with the convolution kernel respective amount of the 4th convolutional layer is obtained after the 4th convolutional layer
Image, described ReLU layers makes partial nerve member be output as 0, causes openness, and the characteristic image is by the 4th max-
After pooling layers and then into two full articulamentums and the ReLU layers being connected with the full articulamentum and dropout layers,
Dropout layers by training when, allow the weight of some hidden layer nodes of convolutional neural networks not work at random, dropout layers
After be sequentially connected a full articulamentum and Softmax layers, last full articulamentum exports the scoring event of the facial image,
Score is input to last layer of Softmax layers, obtains belonging to the probability that mask is faked with true man.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710478892.XA CN107194376A (en) | 2017-06-21 | 2017-06-21 | Mask fraud convolutional neural networks training method and human face in-vivo detection method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710478892.XA CN107194376A (en) | 2017-06-21 | 2017-06-21 | Mask fraud convolutional neural networks training method and human face in-vivo detection method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107194376A true CN107194376A (en) | 2017-09-22 |
Family
ID=59879541
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710478892.XA Pending CN107194376A (en) | 2017-06-21 | 2017-06-21 | Mask fraud convolutional neural networks training method and human face in-vivo detection method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107194376A (en) |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107766820A (en) * | 2017-10-20 | 2018-03-06 | 北京小米移动软件有限公司 | Image classification method and device |
CN108038469A (en) * | 2017-12-27 | 2018-05-15 | 百度在线网络技术(北京)有限公司 | Method and apparatus for detecting human body |
CN108549854A (en) * | 2018-03-28 | 2018-09-18 | 中科博宏(北京)科技有限公司 | A kind of human face in-vivo detection method |
CN108830151A (en) * | 2018-05-07 | 2018-11-16 | 国网浙江省电力有限公司 | Mask detection method based on gauss hybrid models |
CN108921209A (en) * | 2018-06-21 | 2018-11-30 | 杭州骑轻尘信息技术有限公司 | Image identification method, device and electronic equipment |
CN108921041A (en) * | 2018-06-06 | 2018-11-30 | 深圳神目信息技术有限公司 | A kind of biopsy method and device based on RGB and IR binocular camera |
CN108985200A (en) * | 2018-07-02 | 2018-12-11 | 中国科学院半导体研究所 | A kind of In vivo detection algorithm of the non-formula based on terminal device |
CN109146000A (en) * | 2018-09-07 | 2019-01-04 | 电子科技大学 | A kind of method and device for improving convolutional neural networks based on frost weight |
CN109919010A (en) * | 2019-01-24 | 2019-06-21 | 北京三快在线科技有限公司 | Image processing method and device |
CN110674730A (en) * | 2019-09-20 | 2020-01-10 | 华南理工大学 | Monocular-based face silence living body detection method |
WO2020033979A1 (en) * | 2018-08-08 | 2020-02-13 | Rensselaer Polytechnic Institute | Enhancing contrast sensitivity and resolution in a grating interferometer by machine learning |
CN111091388A (en) * | 2020-02-18 | 2020-05-01 | 支付宝实验室(新加坡)有限公司 | Living body detection method and device, face payment method and device, and electronic equipment |
GB2582833A (en) * | 2019-04-30 | 2020-10-07 | Huawei Tech Co Ltd | Facial localisation in images |
CN112347904A (en) * | 2020-11-04 | 2021-02-09 | 杭州锐颖科技有限公司 | Living body detection method, device and medium based on binocular depth and picture structure |
CN113505722A (en) * | 2021-07-23 | 2021-10-15 | 中山大学 | In-vivo detection method, system and device based on multi-scale feature fusion |
CN113642428A (en) * | 2021-07-29 | 2021-11-12 | 北京百度网讯科技有限公司 | Face living body detection method and device, electronic equipment and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105956572A (en) * | 2016-05-15 | 2016-09-21 | 北京工业大学 | In vivo face detection method based on convolutional neural network |
CN106203305A (en) * | 2016-06-30 | 2016-12-07 | 北京旷视科技有限公司 | Human face in-vivo detection method and device |
CN106778607A (en) * | 2016-12-15 | 2017-05-31 | 国政通科技股份有限公司 | A kind of people based on recognition of face and identity card homogeneity authentication device and method |
CN106845395A (en) * | 2017-01-19 | 2017-06-13 | 北京飞搜科技有限公司 | A kind of method that In vivo detection is carried out based on recognition of face |
-
2017
- 2017-06-21 CN CN201710478892.XA patent/CN107194376A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105956572A (en) * | 2016-05-15 | 2016-09-21 | 北京工业大学 | In vivo face detection method based on convolutional neural network |
CN106203305A (en) * | 2016-06-30 | 2016-12-07 | 北京旷视科技有限公司 | Human face in-vivo detection method and device |
CN106778607A (en) * | 2016-12-15 | 2017-05-31 | 国政通科技股份有限公司 | A kind of people based on recognition of face and identity card homogeneity authentication device and method |
CN106845395A (en) * | 2017-01-19 | 2017-06-13 | 北京飞搜科技有限公司 | A kind of method that In vivo detection is carried out based on recognition of face |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107766820A (en) * | 2017-10-20 | 2018-03-06 | 北京小米移动软件有限公司 | Image classification method and device |
CN108038469A (en) * | 2017-12-27 | 2018-05-15 | 百度在线网络技术(北京)有限公司 | Method and apparatus for detecting human body |
CN108549854A (en) * | 2018-03-28 | 2018-09-18 | 中科博宏(北京)科技有限公司 | A kind of human face in-vivo detection method |
CN108549854B (en) * | 2018-03-28 | 2019-04-30 | 中科博宏(北京)科技有限公司 | A kind of human face in-vivo detection method |
CN108830151A (en) * | 2018-05-07 | 2018-11-16 | 国网浙江省电力有限公司 | Mask detection method based on gauss hybrid models |
CN108921041A (en) * | 2018-06-06 | 2018-11-30 | 深圳神目信息技术有限公司 | A kind of biopsy method and device based on RGB and IR binocular camera |
CN108921209A (en) * | 2018-06-21 | 2018-11-30 | 杭州骑轻尘信息技术有限公司 | Image identification method, device and electronic equipment |
CN108985200A (en) * | 2018-07-02 | 2018-12-11 | 中国科学院半导体研究所 | A kind of In vivo detection algorithm of the non-formula based on terminal device |
WO2020033979A1 (en) * | 2018-08-08 | 2020-02-13 | Rensselaer Polytechnic Institute | Enhancing contrast sensitivity and resolution in a grating interferometer by machine learning |
CN109146000A (en) * | 2018-09-07 | 2019-01-04 | 电子科技大学 | A kind of method and device for improving convolutional neural networks based on frost weight |
CN109146000B (en) * | 2018-09-07 | 2022-03-08 | 电子科技大学 | Method and device for improving convolutional neural network based on freezing weight |
CN109919010A (en) * | 2019-01-24 | 2019-06-21 | 北京三快在线科技有限公司 | Image processing method and device |
GB2582833B (en) * | 2019-04-30 | 2021-04-07 | Huawei Tech Co Ltd | Facial localisation in images |
GB2582833A (en) * | 2019-04-30 | 2020-10-07 | Huawei Tech Co Ltd | Facial localisation in images |
CN110674730A (en) * | 2019-09-20 | 2020-01-10 | 华南理工大学 | Monocular-based face silence living body detection method |
CN111091388A (en) * | 2020-02-18 | 2020-05-01 | 支付宝实验室(新加坡)有限公司 | Living body detection method and device, face payment method and device, and electronic equipment |
CN111091388B (en) * | 2020-02-18 | 2024-02-09 | 支付宝实验室(新加坡)有限公司 | Living body detection method and device, face payment method and device and electronic equipment |
CN112347904A (en) * | 2020-11-04 | 2021-02-09 | 杭州锐颖科技有限公司 | Living body detection method, device and medium based on binocular depth and picture structure |
CN112347904B (en) * | 2020-11-04 | 2023-08-01 | 杭州锐颖科技有限公司 | Living body detection method, device and medium based on binocular depth and picture structure |
CN113505722A (en) * | 2021-07-23 | 2021-10-15 | 中山大学 | In-vivo detection method, system and device based on multi-scale feature fusion |
CN113505722B (en) * | 2021-07-23 | 2024-01-02 | 中山大学 | Living body detection method, system and device based on multi-scale feature fusion |
CN113642428A (en) * | 2021-07-29 | 2021-11-12 | 北京百度网讯科技有限公司 | Face living body detection method and device, electronic equipment and storage medium |
CN113642428B (en) * | 2021-07-29 | 2022-09-27 | 北京百度网讯科技有限公司 | Face living body detection method and device, electronic equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107194376A (en) | Mask fraud convolutional neural networks training method and human face in-vivo detection method | |
CN107292267A (en) | Photo fraud convolutional neural networks training method and human face in-vivo detection method | |
CN107301396A (en) | Video fraud convolutional neural networks training method and human face in-vivo detection method | |
CN107220635A (en) | Human face in-vivo detection method based on many fraud modes | |
CN107273845B (en) | Facial expression recognition method based on confidence region and multi-feature weighted fusion | |
CN110569756B (en) | Face recognition model construction method, recognition method, device and storage medium | |
CN103605972B (en) | Non-restricted environment face verification method based on block depth neural network | |
CN106803055B (en) | Face identification method and device | |
CN106485214A (en) | A kind of eyes based on convolutional neural networks and mouth state identification method | |
Yan et al. | Multi-attributes gait identification by convolutional neural networks | |
CN108268859A (en) | A kind of facial expression recognizing method based on deep learning | |
Mallouh et al. | Utilizing CNNs and transfer learning of pre-trained models for age range classification from unconstrained face images | |
CN105069400A (en) | Face image gender recognition system based on stack type sparse self-coding | |
CN111241958A (en) | Video image identification method based on residual error-capsule network | |
CN109815826A (en) | The generation method and device of face character model | |
CN106529395B (en) | Signature image identification method based on depth confidence network and k mean cluster | |
Wang et al. | Topogan: A topology-aware generative adversarial network | |
CN112784929B (en) | Small sample image classification method and device based on double-element group expansion | |
CN113989890A (en) | Face expression recognition method based on multi-channel fusion and lightweight neural network | |
CN113205002B (en) | Low-definition face recognition method, device, equipment and medium for unlimited video monitoring | |
CN111178130A (en) | Face recognition method, system and readable storage medium based on deep learning | |
CN107818299A (en) | Face recognition algorithms based on fusion HOG features and depth belief network | |
Salih et al. | Study of video based facial expression and emotions recognition methods | |
Taha et al. | Emotion recognition from 2D facial expressions | |
Singh et al. | Application of complete local binary pattern method for facial expression recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20170922 |