CN107194376A - Mask fraud convolutional neural networks training method and human face in-vivo detection method - Google Patents

Mask fraud convolutional neural networks training method and human face in-vivo detection method Download PDF

Info

Publication number
CN107194376A
CN107194376A CN201710478892.XA CN201710478892A CN107194376A CN 107194376 A CN107194376 A CN 107194376A CN 201710478892 A CN201710478892 A CN 201710478892A CN 107194376 A CN107194376 A CN 107194376A
Authority
CN
China
Prior art keywords
layers
convolutional layer
image
max
convolutional
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710478892.XA
Other languages
Chinese (zh)
Inventor
张少林
李卫军
韦美丽
宁欣
董肖莉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Wave Wisdom Security And Safety Technology Co ltd
Institute of Semiconductors of CAS
Original Assignee
Beijing Wave Wisdom Security And Safety Technology Co ltd
Institute of Semiconductors of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Wave Wisdom Security And Safety Technology Co ltd, Institute of Semiconductors of CAS filed Critical Beijing Wave Wisdom Security And Safety Technology Co ltd
Priority to CN201710478892.XA priority Critical patent/CN107194376A/en
Publication of CN107194376A publication Critical patent/CN107194376A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/40Spoof detection, e.g. liveness detection
    • G06V40/45Detection of the body part being alive

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Image Analysis (AREA)

Abstract

The present invention relates to face In vivo detection field, in particular to a kind of mask fraud convolutional neural networks training method and human face in-vivo detection method.The mask fraud convolutional neural networks training method that the present invention is provided includes:Build training set;Obtain the image in training set;Face in detection image;Be normalized after face is cut feeding convolutional neural networks in, convolutional neural networks include input layer, multiple convolutional layers, ReLU layers, pooling layers of max, full articulamentum, Dropout layers and SoftmaxWithLoss layers;Convolutional neural networks are trained.By training the convolutional neural networks based on fraud mask, fraud mask is learnt, mask fraud attack is resisted during for face In vivo detection protection is provided, protect the safety of recognition of face.

Description

Mask fraud convolutional neural networks training method and human face in-vivo detection method
Technical field
The present invention relates to face In vivo detection field, trained in particular to a kind of mask fraud convolutional neural networks Method and human face in-vivo detection method.
Background technology
Easily obtained due to face, it is contactless, easy to use the advantages of, recognition of face is liked by people, extensively Apply in gate inhibition's entry-exit management system, login system etc..With the popularization of smart machine, face is easily obtained by other people, is passed through Mask, photo or video are attacked face identification system.Therefore, in face identification system, face In vivo detection becomes Obtain extremely important, be the indispensable means for ensureing user security, be also the premise that recognition of face is widely popularized.
The content of the invention
In view of this, the purpose of the embodiment of the present invention be to provide a kind of mask fraud convolutional neural networks training method and Human face in-vivo detection method, to provide a kind of human face in-vivo detection method for tackling mask fraud.
To achieve the above object, the technical scheme that the embodiment of the present invention is used is as described below:
In a first aspect, the embodiments of the invention provide a kind of mask fraud convolutional neural networks training method, methods described Including:Training set is built, the training set includes mask fraud video and true man's video;Obtain mask fraud video and true man regard Image in frequency;Face in described image is detected by Face datection algorithm;The face is cut;To going out after cutting Facial image is normalized;Facial image after the normalization is inputted to convolutional neural networks, the convolution god Through network include input layer, multiple convolutional layers, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and SoftmaxWithLoss layers;Facial image after the normalization is carried out random cropping by the input layer, by the people after cutting Face image flip horizontal, obtains the image of specific pixel, first that the image of the specific pixel is sent into multiple convolutional layers Individual convolutional layer;First convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max- Pooling layers, the image of the specific pixel obtains the volume with first convolutional layer after first convolutional layer The characteristic image of product core respective amount, described ReLU layers makes partial nerve member be output as 0, causes openness, the max- Pooling layers are compressed to the characteristic image, extract principal character, and the characteristic image enters second convolutional layer;Institute Stating second convolutional layer includes multiple convolution kernels, and second convolutional layer connects one ReLU layers and max-pooling layers, institute State characteristic image and feature with the convolution kernel respective amount of second convolutional layer is obtained after second convolutional layer Image, described ReLU layers makes partial nerve member be output as 0, causes openness, and the characteristic image is by second max- After pooling layers and then enter the 3rd convolutional layer;3rd convolutional layer includes multiple convolution kernels, the 3rd convolution Layer connection one ReLU layer and max-pooling layers, the characteristic image obtained after the 3rd convolutional layer with it is described The characteristic image of the convolution kernel respective amount of 3rd convolutional layer, described ReLU layers makes partial nerve member be output as 0, causes sparse Property, the characteristic image is after the 3rd max-pooling layers and then enters the 4th convolutional layer;4th convolutional layer Including multiple convolution kernels, the 4th convolutional layer connects one ReLU layers and max-pooling layers, and then connection two connects entirely Layer is connect, each full articulamentum is sequentially connected one ReLU layers and dropout layers, and described ReLU layers is output as partial nerve member 0, cause openness, the characteristic image obtains the convolution kernel with the 4th convolutional layer after the 4th convolutional layer The characteristic image of respective amount, after the 4th max-pooling layers so that enter two full articulamentums and with it is described entirely Articulamentum connection ReLU layers and dropout layers, dropout layers by training when, allow at random convolutional neural networks it is some imply The weight of node layer does not work, and a full articulamentum and SoftmaxWithLoss layers is sequentially connected after dropout layers, finally One full articulamentum exports the scoring event of the facial image, and score is input into last layer of SoftmaxWithLoss layers, Obtain belonging to probability and the loss of this iteration that mask is faked with true man.
Further, one BatchNorm layers and one are sequentially connected between each described convolutional layer and ReLU layer Scale layers, the described BatchNorm layers output standardization by neuron is 0 to average, and variance is 1, the Scale layers of reverse The BatchNorm layers of destruction to feature.
Further, the step of structure training set includes:Training set is built according to predetermined ratio and checking collects, it is described Training set and checking collection include mask fraud video and true man's video.
Further, going out the step of facial image is normalized after described pair of cutting includes:To going out people after cutting Face image normalizes to 256 × 256 pixels.
Further, the input layer includes tri- passages of RGB, and size is 256 × 256, and training set is divided into 1000 Batch, each batch sizes are 128, and the pixel of the image of the specific pixel is 227 × 227.
Further, first convolutional layer includes 96 convolution kernels, and the size of each convolution kernel is 7 × 7, convolution behaviour The step-length of work is 2, the characteristic image that 96 pixels are 111 × 111 is obtained after first convolutional layer, by first max- After pooling layers, the characteristic image that 96 pixels are 55 × 55 is obtained, described first max-pooling layer of size is 3 × 3, step-length is 2.
Further, second convolutional layer includes 256 convolution kernels, and the size of each convolution kernel is 5 × 5, convolution The step-length of operation is 2, the characteristic image that 256 pixels are 26 × 26 is obtained after second convolutional layer, by second After max-pooling layers, the characteristic image that 256 pixels are 26 × 26, described second max-pooling layers of size are obtained For 3 × 3, step-length is 1, and Boundary filling is 1.
Further, the 3rd convolutional layer includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border Supplement 1, the step-length of convolution operation is 1, and the characteristic image that 192 pixels are 26 × 26, warp are obtained after the 3rd convolutional layer Cross after the 3rd max-pooling layers, obtain the characteristic image that 192 pixels are 26 × 26, the 3rd max-pooling The size of layer is 3 × 3, and step-length is 1, and Boundary filling is 1.
Further, the 4th convolutional layer includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border Supplement 1, the step-length of convolution operation is 1, and the characteristic image that 192 pixels are 26 × 26, warp are obtained after the 4th convolutional layer Cross after the 4th max-pooling layers, obtain the characteristic image that 192 pixels are 13 × 13, the 4th max-pooling The size of layer is 3 × 3, and step-length is 2.
Second aspect, the embodiments of the invention provide a kind of human face in-vivo detection method faked based on mask, the side Method includes:Obtain facial image;Face in described image is detected by Face datection algorithm;The face is cut; It is normalized to going out facial image after cutting;Facial image after the normalization is inputted to convolutional neural networks, The convolutional neural networks include input layer, multiple convolutional layers, ReLU layers, max-pooling layers, full articulamentum, Dropout layers With Softmax layers;Facial image after the normalization is carried out random cropping by the input layer, by the facial image after cutting Flip horizontal, obtains the image of specific pixel, first convolution that the image of the specific pixel is sent into multiple convolutional layers Layer;First convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max-pooling Layer, the image of the specific pixel obtains corresponding with the convolution kernel of first convolutional layer after first convolutional layer The characteristic image of quantity, described ReLU layers makes partial nerve member be output as 0, causes openness, and described max-pooling layers to institute State characteristic image to be compressed, extract principal character, the characteristic image enters second convolutional layer;Second convolutional layer Including multiple convolution kernels, second convolutional layer connects one ReLU layers and max-pooling layers, and the characteristic image passes through The characteristic image with the convolution kernel respective amount of second convolutional layer is obtained after second convolutional layer, it is described ReLU layers Partial nerve member is output as 0, cause openness, the characteristic image is after second max-pooling layers and then enters 3rd convolutional layer;3rd convolutional layer includes multiple convolution kernels, the 3rd convolutional layer connect one ReLU layer with Max-pooling layers, the characteristic image obtains the convolution with the 3rd convolutional layer after the 3rd convolutional layer The characteristic image of core respective amount, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image warp Cross after the 3rd max-pooling layers and then enter the 4th convolutional layer;4th convolutional layer includes multiple convolution kernels, institute State the 4th convolutional layer and connect one ReLU layers and max-pooling layers, then two full articulamentums of connection, each full articulamentum Be sequentially connected one ReLU layers and dropout layers, the characteristic image obtained after the 4th convolutional layer with it is described The characteristic image of the convolution kernel respective amount of 4th convolutional layer, described ReLU layers makes partial nerve member be output as 0, causes sparse Property, the characteristic image after the 4th max-pooling layers so that enter two full articulamentums and with the full connection The ReLU layers of layer connection with dropout layers, dropout layers by training when, allow some hidden layer sections of convolutional neural networks at random The weight of point does not work, and a full articulamentum and Softmax layers is sequentially connected after dropout layers, last full articulamentum The scoring event of the facial image is exported, score is input to last layer of Softmax layers, obtains belonging to mask fraud and true The probability of people.
Mask fraud convolutional neural networks training method and human face in-vivo detection method provided in an embodiment of the present invention, pass through The convolutional neural networks based on fraud mask are trained, fraud mask is learnt, mask fraud is resisted when being recognition of face and is attacked Offer protection is provided, the safety of recognition of face is protected.
To enable the above objects, features and advantages of the present invention to become apparent, preferred embodiment cited below particularly, and coordinate Appended accompanying drawing, is described in detail below.
Brief description of the drawings
In order to illustrate the technical solution of the embodiments of the present invention more clearly, below will be attached to what is used required in embodiment Figure is briefly described, it will be appreciated that the following drawings illustrate only certain embodiments of the present invention, therefore is not construed as pair The restriction of scope, for those of ordinary skill in the art, on the premise of not paying creative work, can also be according to this A little accompanying drawings obtain other related accompanying drawings.
Fig. 1 is the schematic flow sheet of mask fraud convolutional neural networks training method provided in an embodiment of the present invention.
Fig. 2 is the training flow chart of convolutional neural networks.
Fig. 3 is the schematic diagram for the convolutional neural networks model trained in the embodiment of the present invention.
Fig. 4 is a kind of flow chart of human face in-vivo detection method faked based on mask provided in an embodiment of the present invention.
Embodiment
To make the purpose, technical scheme and advantage of the embodiment of the present invention clearer, below in conjunction with the embodiment of the present invention In accompanying drawing, the technical scheme in the embodiment of the present invention is clearly and completely described, it is clear that described embodiment is A part of embodiment of the present invention, rather than whole embodiments.The present invention implementation being generally described and illustrated herein in the accompanying drawings The component of example can be arranged and designed with a variety of configurations.
Therefore, the detailed description of embodiments of the invention below to providing in the accompanying drawings is not intended to limit claimed The scope of the present invention, but be merely representative of the present invention selected embodiment.Based on the embodiment in the present invention, this area is common The every other embodiment that technical staff is obtained under the premise of creative work is not made, belongs to the model that the present invention is protected Enclose.
It should be noted that:Similar label and letter represents similar terms in following accompanying drawing, therefore, once a certain Xiang Yi It is defined in individual accompanying drawing, then it further need not be defined and explained in subsequent accompanying drawing.
Fig. 1 is refer to, is the flow signal of mask fraud convolutional neural networks training method provided in an embodiment of the present invention Figure.This method comprises the following steps:
Step S110, builds training set, and training set includes mask fraud video and true man's video.
When training set does not include checking collection, step S101 also includes building training set according to predetermined ratio and checking collects, The training set and checking collection include mask fraud video and true man's video.The ratio of training set and checking collection is freely configured, For example can be 8:2 or 7:3, the embodiment of the present invention is not limited this.Mask fraud video is the video for the mask recorded, should Mask is high with true man's human face similarity degree, and the scene that simulation is attacked with mask fraud, true man's video regards for the true man's of recording Frequently.
Step S120, obtains the image in mask fraud video and true man's video.
Multiple images are taken out respectively from mask fraud video and true man's video, such as take out a figure every 3 frames or 5 frames Picture, then a video take out altogether:Totalframes/3 or totalframes/5 image.
Step S130, the face in described image is detected by Face datection algorithm.
Face datection algorithm can be existing general Face datection algorithm, such as adaboost.
Step S140, cuts to face.
During being cut to face, using the algorithm of facial modeling, the precise area of face is found, is obtained To accurate face frame, the face frame is expanded, it is general to expand 1.8 times, the area in the corresponding artwork of face frame after will be enlarged by Domain, which is cut out, to be come, and cutting out the face come both includes face, also including certain background, and certain background there are beneficial to classification True and false video.
Step S150, is normalized to going out facial image after cutting.
Facial image after cutting passes through normalized, normalizes to 256 × 256 pixels.
Step S160, the facial image after normalization is inputted to convolutional neural networks, and the convolutional neural networks include defeated Enter layer, multiple convolutional layers, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and SoftmaxWithLoss layers.
The convolutional neural networks are trained by step S170.
Wherein, the facial image after normalization initially enters input layer, refer to Fig. 2, is the training of convolutional neural networks Process, comprises the following steps:
Facial image after normalization is carried out random cropping by step S171, input layer, by the facial image water after cutting Flat upset, obtains the image of specific pixel, first convolutional layer that the image of specific pixel is sent into multiple convolutional layers.
Wherein, the image of input layer includes tri- passages of RGB, and size is 256 × 256, and training set is divided into 1000 Batch, each batch sizes are 128.The pixel of the image of specific pixel is 227 × 227, and the image in the batch is upset The positive negative ratio 1 of image pattern that order is sent into first convolutional layer, the batch:1, and belong to the face figure of same person Picture, to reduce due to the difference that different macroscopic features are caused, makes e-learning to due to the substantive characteristics between true man and fraud, Reduce difference in class.
Step S172, first convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max- Pooling layers, the image of specific pixel obtains the convolution kernel respective amount with first convolutional layer after first convolutional layer Characteristic image, described ReLU layers makes partial nerve member be output as 0, causes openness, and max-pooling layers are entered to characteristic image Row compression, extracts principal character, and characteristic image enters second convolutional layer.
In the present embodiment, the number of the convolution kernel of first convolutional layer is 96, and the size of each convolution kernel is 7 × 7, The step-length of convolution operation is 2, and the characteristic image that 96 pixels are 111 × 111, first are obtained after first convolutional layer Max-pooling layers of size is 3 × 3, and step-length is 2, after first max-pooling layers, and it is 55 to obtain 96 pixels × 55 characteristic image.As preferred, one BatchNorm layers and one are sequentially connected between each convolutional layer and ReLU layers It is individual Scale layers, by BatchNorm layers, Scale layers, ReLU layers do not change characteristic image size.When depth network layer Too much, signal and gradient are less and less, and deep layer is difficult to train, referred to as gradient disperse, it is also possible to increasing, also referred to as Gradient is exploded, and is 0 by the BatchNorm layers of output standardization by neuron to average, variance is that 1, Scale layers of reverse are described The BatchNorm layers of destruction to feature, to overcome deep neural network to be difficult to the disadvantage trained.Because depth network is in training When, each layer of input data distribution is always to change, because above one layer of parameter renewal can cause next one When the distribution of the input data of layer, BatchNorm and Scale can solve training, the data distribution in network intermediate layer changes The problem of change, BatchNorm layers are by the way that the output standardization of neuron is arrived:Average is 0, and variance is 1, is passed through After BatchNorm layers, all neurons all standardize to a kind of distribution, but export-restriction is in the distribution meeting again of the variance 1 of average 0 So that the ability to express of network dies down, the feature that front layer learns also is destroyed, so adding Scale layers, is learned by adding The reconstruction parameter practised, reverses destruction to feature, the program of reverse by model training when adjust.So both neuron is done Standardization retains the feature acquired of front layer again, because input becomes stable, promotes the convergence and to a certain extent of model Prevent over-fitting.Pass through successively yardstick normalizing, it is to avoid gradient is disappeared and gradient is overflowed, and accelerates convergence, while conduct A kind of Regularization Technique also improves generalization ability;Make activation primitive compared to using sigmoid tanh, it is computationally intensive, reversely pass Broadcast when seeking error gradient, derived function amount is also very big, and sigmoid tanh functions are easily saturated, and gradient disappearance feelings occur Condition, i.e., close to during convergence, conversion is too slow, causes information to lose.ReLU layers can make partial nerve member be output as 0, cause dilute Property is dredged, over-fitting is not only alleviated, also overcoming gradient closer to real neuronal activation model disappears, not unsupervised Pre-training (i.e. make by first hidden layer of training network, the network parameter values that retraining second ... is finally trained with these For the initial value of overall network parameter) in the case of dramatically speed up convergence compared to sigmoid and tanh activation primitives.
Step S173, second convolutional layer includes multiple convolution kernels, and second convolutional layer connects one ReLU layers and max- Pooling layers, characteristic image obtains the feature with the convolution kernel respective amount of second convolutional layer after second convolutional layer Image, described ReLU layers makes partial nerve member be output as 0, causes openness, and the characteristic image is by second max- After pooling layers and then enter the 3rd convolutional layer.
Second convolutional layer includes 256 convolution kernels, and the size of each convolution kernel is 5 × 5, and the step-length of convolution operation is 2, The characteristic image that 256 pixels are 26 × 26 is obtained after second convolutional layer, ReLU layers make partial nerve member be output as 0, Cause openness, characteristic image obtains the characteristic image that 256 pixels are 26 × 26 after second max-pooling layers, Second max-pooling layers of size is 3 × 3, and step-length is 1, and Boundary filling is 1.
Step S174, the 3rd convolutional layer includes multiple convolution kernels, and the 3rd convolutional layer connects a max-pooling Layer, characteristic image obtains the characteristic image with the convolution kernel respective amount of the 3rd convolutional layer after the 3rd convolutional layer, ReLU layers make partial nerve member be output as 0, cause openness, and characteristic image is after the 3rd max-pooling layers and then enters Enter the 4th convolutional layer.
3rd convolutional layer includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border supplement 1, convolution operation Step-length be 1, obtained after the 3rd convolutional layer 192 pixels be 26 × 26 characteristic image, by the 3rd max- After pooling layers, the characteristic image that 192 pixels are 26 × 26 is obtained, the 3rd max-pooling layers of size is 3 × 3, Step-length is 1, and Boundary filling is 1.
Step S175, the 4th convolutional layer includes multiple convolution kernels, and the 4th convolutional layer connects a max-pooling Layer, then connects two full articulamentums, and each full articulamentum is sequentially connected one ReLU layers and dropout layers, and ReLU layers make Partial nerve member is output as 0, causes openness, and characteristic image obtains the volume with the 4th convolutional layer after the 4th convolutional layer The characteristic image of product core respective amount, after the 4th maxpooling layers and then into two full articulamentums and with connecting entirely Connect layer connection ReLU layers with dropout layers.
4th convolutional layer includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border supplement 1, convolution operation Step-length be 1, obtained after the 4th convolutional layer 192 pixels be 26 × 26 characteristic image, by the 4th max- After pooling layers, the characteristic image that 192 pixels are 13 × 13 is obtained, the 4th max-pooling layers of size is 3 × 3, Step-length is 2.
When S176, dropout layers of step is by training, allow at random some hidden layer nodes of convolutional neural networks weight not Work, be sequentially connected a full articulamentum and SoftmaxWithLoss layers after dropout layers, last full articulamentum is defeated Go out the scoring event of facial image, score be input to last layer of SoftmaxWithLoss layers, obtain belonging to mask fake and The probability of true man and the loss of current iteration.
Dropout layers eliminate the simultaneous adaptation reduced between neuron node, enhance generalization ability, dropout layers During by training pattern, the idle mode of weight of some hidden layer nodes of network is allowed to prevent model over-fitting, canonical at random Change means, improve generalization ability.In training process, SoftmaxWithLoss layers include Softmax layers and multidimensional Logistic Loss layers, Softmax layers scoring event above is mapped as belonging to each class probability, be followed by a multidimensional Logistic Loss layers, what is obtained here is the loss of current iteration.One layer is merged into Softmax layers and Loss layers of multidimensional Logistic Stabilization numerically is ensure that, two full articulamentums are output as 512 dimensional features.
Fig. 3 is refer to, is the schematic diagram of the convolutional neural networks model in the embodiment of the present invention in training process.The present invention The mask fraud convolutional neural networks training method that embodiment is provided is right by training the convolutional neural networks based on fraud mask Fraud mask is learnt, and mask fraud attack is resisted during for recognition of face and provides protection, the safety of recognition of face is protected.
Fig. 4 is refer to, the embodiment of the present invention additionally provides a kind of human face in-vivo detection method faked based on mask, the party Method comprises the following steps:
Step S210, obtains facial image.
The facial image of acquisition can be the face figure obtained by the camera of the equipment such as gate inhibition, computer or monitor Picture.
Step S220, passes through the face in Face datection algorithm detection image.
Step S230, cutting is carried out by face
Step S240, is normalized to going out facial image after cutting.
Step S250, the facial image after normalization is inputted to convolutional neural networks, convolutional neural networks include input Layer, multiple convolutional layers, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and Softmax layers.
Step S260, the facial image is identified.
Wherein, step S260 identification process is carried out according to the convolutional neural networks that above-described embodiment is trained, to obtain people The true and false probability of face image.Due to making statement to detailed process and principle in the aforementioned embodiment, here is omitted.
In summary, the embodiments of the invention provide a kind of mask fraud convolutional neural networks training method and face live body Detection method, by training the convolutional neural networks based on fraud mask, learns to fraud mask, is supported when being recognition of face Imperial mask, which fakes to attack, provides protection, protects the safety of recognition of face.
, can also be by another way in embodiment provided herein, it should be understood that disclosed method Realize.Embodiments described above is only schematical, for example, the flow chart and block diagram in accompanying drawing are shown according to this hair Architectural framework in the cards, function and the operation of the method and computer program product of bright embodiment.At this point, flow Each square frame in figure or block diagram can represent a part for a module, program segment or code, the module, program segment or generation A part for code includes one or more executable instructions for being used to realize defined logic function.It should also be noted that at some As in the implementation of replacement, the function of being marked in square frame can also be with different from the order marked in accompanying drawing generation. For example, two continuous square frames can essentially be performed substantially in parallel, they can also be performed in the opposite order sometimes, this Depending on involved function.It is also noted that each square frame and block diagram and/or flow in block diagram and/or flow chart The combination of square frame in figure, can be realized with the special hardware based system of defined function or action is performed, or It can be realized with the combination of specialized hardware and computer instruction.
In addition, each functional module in each embodiment of the invention can integrate to form an independent portion Point or modules individualism, can also two or more modules be integrated to form an independent part.
If the function is realized using in the form of software function module and is used as independent production marketing or in use, can be with It is stored in a computer read/write memory medium.Understood based on such, technical scheme is substantially in other words The part contributed to prior art or the part of the technical scheme can be embodied in the form of software product, the meter Calculation machine software product is stored in a storage medium, including some instructions are to cause a computer equipment (can be individual People's computer, server, or network equipment etc.) perform all or part of step of each of the invention embodiment methods described. And foregoing storage medium includes:USB flash disk, mobile hard disk, read-only storage (ROM, Read-Only Memory), arbitrary access are deposited Reservoir (RAM, Random Access Memory), magnetic disc or CD etc. are various can be with the medium of store program codes.Need Illustrate, herein, such as first and second or the like relational terms be used merely to by an entity or operation with Another entity or operation make a distinction, and not necessarily require or imply between these entities or operation there is any this reality The relation or order on border.Moreover, term " comprising ", "comprising" or its any other variant are intended to the bag of nonexcludability Contain, so that process, method, article or equipment including a series of key elements are not only including those key elements, but also including Other key elements being not expressly set out, or also include for this process, method, article or the intrinsic key element of equipment. In the absence of more restrictions, the key element limited by sentence "including a ...", it is not excluded that including the key element Process, method, article or equipment in also there is other identical element.
The preferred embodiments of the present invention are the foregoing is only, are not intended to limit the invention, for the skill of this area For art personnel, the present invention can have various modifications and variations.Within the spirit and principles of the invention, that is made any repaiies Change, equivalent substitution, improvement etc., should be included in the scope of the protection.It should be noted that:Similar label and letter exists Similar terms is represented in following accompanying drawing, therefore, once being defined in a certain Xiang Yi accompanying drawing, is then not required in subsequent accompanying drawing It is further defined and explained.

Claims (10)

1. a kind of mask fraud convolutional neural networks training method, it is characterised in that methods described includes:
Training set is built, the training set includes mask fraud video and true man's video;
Obtain the image in mask fraud video and true man's video;
Face in described image is detected by Face datection algorithm;
The face is cut;
It is normalized to going out facial image after cutting;
Facial image after the normalization is inputted to convolutional neural networks, the convolutional neural networks include input layer, many Individual convolutional layer, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and SoftmaxWithLoss layers;
Facial image after the normalization is carried out random cropping by the input layer, and the facial image level after cutting is turned over Turn, obtain the image of specific pixel, first convolutional layer that the image of the specific pixel is sent into multiple convolutional layers;
First convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max- Pooling layers, the image of the specific pixel obtains the volume with first convolutional layer after first convolutional layer The characteristic image of product core respective amount, described ReLU layers makes partial nerve member be output as 0, causes openness, the max- Pooling layers are compressed to the characteristic image, extract principal character, and the characteristic image enters second convolutional layer;
Second convolutional layer includes multiple convolution kernels, and second convolutional layer connects one ReLU layers and max- Pooling layers, the characteristic image obtains the convolution kernel pair with second convolutional layer after second convolutional layer The characteristic image of quantity is answered, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image passes through After two max-pooling layers and then enter the 3rd convolutional layer;
3rd convolutional layer includes multiple convolution kernels, and the 3rd convolutional layer connects one ReLU layers and max- Pooling layers, the characteristic image obtains the convolution kernel pair with the 3rd convolutional layer after the 3rd convolutional layer The characteristic image of quantity is answered, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image passes through After three max-pooling layers and then enter the 4th convolutional layer;
4th convolutional layer includes multiple convolution kernels, and the 4th convolutional layer connects one ReLU layers and max- Pooling layers, two full articulamentums are then connected, each full articulamentum is sequentially connected one ReLU layers and dropout layers, institute State characteristic image and feature with the convolution kernel respective amount of the 4th convolutional layer is obtained after the 4th convolutional layer Image, described ReLU layers makes partial nerve member be output as 0, causes openness, and the characteristic image is by the 4th max- After pooling layers so enter two full articulamentums and the ReLU layers being connected with the full articulamentum and dropout layers;
Dropout layers by training when, allow the weight of some hidden layer nodes of convolutional neural networks not work at random, dropout A full articulamentum and SoftmaxWithLoss layers are sequentially connected after layer, last full articulamentum exports the face figure The scoring event of picture, last layer of SoftmaxWithLoss layers is input to by score, obtain belonging to mask fake and true man it is general The loss of rate and this iteration.
2. mask fraud convolutional neural networks training method according to claim 1, it is characterised in that each described volume It is sequentially connected one BatchNorm layers and one Scale layers between lamination and ReLU layers, described BatchNorm layers by nerve The output standardization of member is 0 to average, and variance is 1, and described Scale layers is reversed the BatchNorm layers of destruction to feature.
3. mask fraud convolutional neural networks training method according to claim 1, it is characterised in that the structure training The step of collection, includes:
Training set is built according to predetermined ratio and checking collects, and the training set and checking collection include mask fraud video and true man Video.
4. mask fraud convolutional neural networks training method according to claim 1, it is characterised in that after described pair cuts Going out the step of facial image is normalized includes:
256 × 256 pixels are normalized to going out facial image after cutting.
5. mask fraud convolutional neural networks training method according to claim 1, it is characterised in that the input layer bag Containing tri- passages of RGB, size is 256 × 256, training set is divided into 1000 batch, each batch sizes are 128, described The pixel of the image of specific pixel is 227 × 227.
6. the mask fraud convolutional neural networks training method according to claim any one of 1-5, it is characterised in that described First convolutional layer includes 96 convolution kernels, and the size of each convolution kernel is 7 × 7, and the step-length of convolution operation is 2, by first The characteristic image that 96 pixels are 111 × 111 is obtained after individual convolutional layer, after first max-pooling layers, 96 are obtained Pixel is 55 × 55 characteristic image, and described first max-pooling layers of size is 3 × 3, and step-length is 2.
7. mask fraud convolutional neural networks training method according to claim 6, it is characterised in that second volume Lamination includes 256 convolution kernels, and the size of each convolution kernel is 5 × 5, and the step-length of convolution operation is 2, by second convolutional layer The characteristic image that 256 pixels are 26 × 26 is obtained afterwards, and after second max-pooling layers, it is 26 to obtain 256 pixels × 26 characteristic image, described second max-pooling layers of size is 3 × 3, and step-length is 1, and Boundary filling is 1.
8. mask fraud convolutional neural networks training method according to claim 7, it is characterised in that the 3rd volume Lamination includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border supplement 1, and the step-length of convolution operation is 1, by the The characteristic image that 192 pixels are 26 × 26 is obtained after three convolutional layers, after the 3rd max-pooling layers, 192 are obtained Individual pixel is 26 × 26 characteristic image, and described 3rd max-pooling layers of size is 3 × 3, and step-length is 1, Boundary filling For 1.
9. mask fraud convolutional neural networks training method according to claim 8, it is characterised in that the 4th volume Lamination includes 192 convolution kernels, and the size of each convolution kernel is 3 × 3, border supplement 1, and the step-length of convolution operation is 1, by the The characteristic image that 192 pixels are 26 × 26 is obtained after four convolutional layers, after the 4th max-pooling layers, 192 are obtained Individual pixel is 13 × 13 characteristic image, and described 4th max-pooling layers of size is 3 × 3, and step-length is 2.
10. a kind of human face in-vivo detection method faked based on mask, it is characterised in that methods described includes:
Obtain facial image;
Face in described image is detected by Face datection algorithm;
The face is cut;
It is normalized to going out facial image after cutting;
Facial image after the normalization is inputted to convolutional neural networks, the convolutional neural networks include input layer, many Individual convolutional layer, ReLU layers, max-pooling layers, full articulamentum, Dropout layers and Softmax layers;
Facial image after the normalization is carried out random cropping by the input layer, and the facial image level after cutting is turned over Turn, obtain the image of specific pixel, first convolutional layer that the image of the specific pixel is sent into multiple convolutional layers;
First convolutional layer includes multiple convolution kernels, and first convolutional layer connects one ReLU layers and max- Pooling layers, the image of the specific pixel obtains the volume with first convolutional layer after first convolutional layer The characteristic image of product core respective amount, described ReLU layers makes partial nerve member be output as 0, causes openness, the max- Pooling layers are compressed to the characteristic image, extract principal character, and the characteristic image enters second convolutional layer;
Second convolutional layer includes multiple convolution kernels, and second convolutional layer connects one ReLU layers and max- Pooling layers, the characteristic image obtains the convolution kernel pair with second convolutional layer after second convolutional layer The characteristic image of quantity is answered, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image passes through After two max-pooling layers and then enter the 3rd convolutional layer;
3rd convolutional layer includes multiple convolution kernels, and the 3rd convolutional layer connects one ReLU layers and max- Pooling layers, the characteristic image obtains the convolution kernel pair with the 3rd convolutional layer after the 3rd convolutional layer The characteristic image of quantity is answered, described ReLU layers makes partial nerve member be output as 0, causes openness, the characteristic image passes through After three max-pooling layers and then enter the 4th convolutional layer;
4th convolutional layer includes multiple convolution kernels, and the 4th convolutional layer connects one ReLU layers and max- Pooling layers, two full articulamentums are then connected, each full articulamentum is sequentially connected one ReLU layers and dropout layers, institute State characteristic image and feature with the convolution kernel respective amount of the 4th convolutional layer is obtained after the 4th convolutional layer Image, described ReLU layers makes partial nerve member be output as 0, causes openness, and the characteristic image is by the 4th max- After pooling layers and then into two full articulamentums and the ReLU layers being connected with the full articulamentum and dropout layers, Dropout layers by training when, allow the weight of some hidden layer nodes of convolutional neural networks not work at random, dropout layers After be sequentially connected a full articulamentum and Softmax layers, last full articulamentum exports the scoring event of the facial image, Score is input to last layer of Softmax layers, obtains belonging to the probability that mask is faked with true man.
CN201710478892.XA 2017-06-21 2017-06-21 Mask fraud convolutional neural networks training method and human face in-vivo detection method Pending CN107194376A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710478892.XA CN107194376A (en) 2017-06-21 2017-06-21 Mask fraud convolutional neural networks training method and human face in-vivo detection method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710478892.XA CN107194376A (en) 2017-06-21 2017-06-21 Mask fraud convolutional neural networks training method and human face in-vivo detection method

Publications (1)

Publication Number Publication Date
CN107194376A true CN107194376A (en) 2017-09-22

Family

ID=59879541

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710478892.XA Pending CN107194376A (en) 2017-06-21 2017-06-21 Mask fraud convolutional neural networks training method and human face in-vivo detection method

Country Status (1)

Country Link
CN (1) CN107194376A (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107766820A (en) * 2017-10-20 2018-03-06 北京小米移动软件有限公司 Image classification method and device
CN108038469A (en) * 2017-12-27 2018-05-15 百度在线网络技术(北京)有限公司 Method and apparatus for detecting human body
CN108549854A (en) * 2018-03-28 2018-09-18 中科博宏(北京)科技有限公司 A kind of human face in-vivo detection method
CN108830151A (en) * 2018-05-07 2018-11-16 国网浙江省电力有限公司 Mask detection method based on gauss hybrid models
CN108921209A (en) * 2018-06-21 2018-11-30 杭州骑轻尘信息技术有限公司 Image identification method, device and electronic equipment
CN108921041A (en) * 2018-06-06 2018-11-30 深圳神目信息技术有限公司 A kind of biopsy method and device based on RGB and IR binocular camera
CN108985200A (en) * 2018-07-02 2018-12-11 中国科学院半导体研究所 A kind of In vivo detection algorithm of the non-formula based on terminal device
CN109146000A (en) * 2018-09-07 2019-01-04 电子科技大学 A kind of method and device for improving convolutional neural networks based on frost weight
CN109919010A (en) * 2019-01-24 2019-06-21 北京三快在线科技有限公司 Image processing method and device
CN110674730A (en) * 2019-09-20 2020-01-10 华南理工大学 Monocular-based face silence living body detection method
WO2020033979A1 (en) * 2018-08-08 2020-02-13 Rensselaer Polytechnic Institute Enhancing contrast sensitivity and resolution in a grating interferometer by machine learning
CN111091388A (en) * 2020-02-18 2020-05-01 支付宝实验室(新加坡)有限公司 Living body detection method and device, face payment method and device, and electronic equipment
GB2582833A (en) * 2019-04-30 2020-10-07 Huawei Tech Co Ltd Facial localisation in images
CN112347904A (en) * 2020-11-04 2021-02-09 杭州锐颖科技有限公司 Living body detection method, device and medium based on binocular depth and picture structure
CN113505722A (en) * 2021-07-23 2021-10-15 中山大学 In-vivo detection method, system and device based on multi-scale feature fusion
CN113642428A (en) * 2021-07-29 2021-11-12 北京百度网讯科技有限公司 Face living body detection method and device, electronic equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105956572A (en) * 2016-05-15 2016-09-21 北京工业大学 In vivo face detection method based on convolutional neural network
CN106203305A (en) * 2016-06-30 2016-12-07 北京旷视科技有限公司 Human face in-vivo detection method and device
CN106778607A (en) * 2016-12-15 2017-05-31 国政通科技股份有限公司 A kind of people based on recognition of face and identity card homogeneity authentication device and method
CN106845395A (en) * 2017-01-19 2017-06-13 北京飞搜科技有限公司 A kind of method that In vivo detection is carried out based on recognition of face

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105956572A (en) * 2016-05-15 2016-09-21 北京工业大学 In vivo face detection method based on convolutional neural network
CN106203305A (en) * 2016-06-30 2016-12-07 北京旷视科技有限公司 Human face in-vivo detection method and device
CN106778607A (en) * 2016-12-15 2017-05-31 国政通科技股份有限公司 A kind of people based on recognition of face and identity card homogeneity authentication device and method
CN106845395A (en) * 2017-01-19 2017-06-13 北京飞搜科技有限公司 A kind of method that In vivo detection is carried out based on recognition of face

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107766820A (en) * 2017-10-20 2018-03-06 北京小米移动软件有限公司 Image classification method and device
CN108038469A (en) * 2017-12-27 2018-05-15 百度在线网络技术(北京)有限公司 Method and apparatus for detecting human body
CN108549854A (en) * 2018-03-28 2018-09-18 中科博宏(北京)科技有限公司 A kind of human face in-vivo detection method
CN108549854B (en) * 2018-03-28 2019-04-30 中科博宏(北京)科技有限公司 A kind of human face in-vivo detection method
CN108830151A (en) * 2018-05-07 2018-11-16 国网浙江省电力有限公司 Mask detection method based on gauss hybrid models
CN108921041A (en) * 2018-06-06 2018-11-30 深圳神目信息技术有限公司 A kind of biopsy method and device based on RGB and IR binocular camera
CN108921209A (en) * 2018-06-21 2018-11-30 杭州骑轻尘信息技术有限公司 Image identification method, device and electronic equipment
CN108985200A (en) * 2018-07-02 2018-12-11 中国科学院半导体研究所 A kind of In vivo detection algorithm of the non-formula based on terminal device
WO2020033979A1 (en) * 2018-08-08 2020-02-13 Rensselaer Polytechnic Institute Enhancing contrast sensitivity and resolution in a grating interferometer by machine learning
CN109146000A (en) * 2018-09-07 2019-01-04 电子科技大学 A kind of method and device for improving convolutional neural networks based on frost weight
CN109146000B (en) * 2018-09-07 2022-03-08 电子科技大学 Method and device for improving convolutional neural network based on freezing weight
CN109919010A (en) * 2019-01-24 2019-06-21 北京三快在线科技有限公司 Image processing method and device
GB2582833B (en) * 2019-04-30 2021-04-07 Huawei Tech Co Ltd Facial localisation in images
GB2582833A (en) * 2019-04-30 2020-10-07 Huawei Tech Co Ltd Facial localisation in images
CN110674730A (en) * 2019-09-20 2020-01-10 华南理工大学 Monocular-based face silence living body detection method
CN111091388A (en) * 2020-02-18 2020-05-01 支付宝实验室(新加坡)有限公司 Living body detection method and device, face payment method and device, and electronic equipment
CN111091388B (en) * 2020-02-18 2024-02-09 支付宝实验室(新加坡)有限公司 Living body detection method and device, face payment method and device and electronic equipment
CN112347904A (en) * 2020-11-04 2021-02-09 杭州锐颖科技有限公司 Living body detection method, device and medium based on binocular depth and picture structure
CN112347904B (en) * 2020-11-04 2023-08-01 杭州锐颖科技有限公司 Living body detection method, device and medium based on binocular depth and picture structure
CN113505722A (en) * 2021-07-23 2021-10-15 中山大学 In-vivo detection method, system and device based on multi-scale feature fusion
CN113505722B (en) * 2021-07-23 2024-01-02 中山大学 Living body detection method, system and device based on multi-scale feature fusion
CN113642428A (en) * 2021-07-29 2021-11-12 北京百度网讯科技有限公司 Face living body detection method and device, electronic equipment and storage medium
CN113642428B (en) * 2021-07-29 2022-09-27 北京百度网讯科技有限公司 Face living body detection method and device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN107194376A (en) Mask fraud convolutional neural networks training method and human face in-vivo detection method
CN107292267A (en) Photo fraud convolutional neural networks training method and human face in-vivo detection method
CN107301396A (en) Video fraud convolutional neural networks training method and human face in-vivo detection method
CN107220635A (en) Human face in-vivo detection method based on many fraud modes
CN107273845B (en) Facial expression recognition method based on confidence region and multi-feature weighted fusion
CN110569756B (en) Face recognition model construction method, recognition method, device and storage medium
CN103605972B (en) Non-restricted environment face verification method based on block depth neural network
CN106803055B (en) Face identification method and device
CN106485214A (en) A kind of eyes based on convolutional neural networks and mouth state identification method
Yan et al. Multi-attributes gait identification by convolutional neural networks
CN108268859A (en) A kind of facial expression recognizing method based on deep learning
Mallouh et al. Utilizing CNNs and transfer learning of pre-trained models for age range classification from unconstrained face images
CN105069400A (en) Face image gender recognition system based on stack type sparse self-coding
CN111241958A (en) Video image identification method based on residual error-capsule network
CN109815826A (en) The generation method and device of face character model
CN106529395B (en) Signature image identification method based on depth confidence network and k mean cluster
Wang et al. Topogan: A topology-aware generative adversarial network
CN112784929B (en) Small sample image classification method and device based on double-element group expansion
CN113989890A (en) Face expression recognition method based on multi-channel fusion and lightweight neural network
CN113205002B (en) Low-definition face recognition method, device, equipment and medium for unlimited video monitoring
CN111178130A (en) Face recognition method, system and readable storage medium based on deep learning
CN107818299A (en) Face recognition algorithms based on fusion HOG features and depth belief network
Salih et al. Study of video based facial expression and emotions recognition methods
Taha et al. Emotion recognition from 2D facial expressions
Singh et al. Application of complete local binary pattern method for facial expression recognition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20170922