WO2023061123A1 - Facial silent living body detection method and apparatus, and storage medium and device - Google Patents

Facial silent living body detection method and apparatus, and storage medium and device Download PDF

Info

Publication number
WO2023061123A1
WO2023061123A1 PCT/CN2022/118048 CN2022118048W WO2023061123A1 WO 2023061123 A1 WO2023061123 A1 WO 2023061123A1 CN 2022118048 W CN2022118048 W CN 2022118048W WO 2023061123 A1 WO2023061123 A1 WO 2023061123A1
Authority
WO
WIPO (PCT)
Prior art keywords
face
face image
parameters
parameter
living body
Prior art date
Application number
PCT/CN2022/118048
Other languages
French (fr)
Chinese (zh)
Inventor
王洪
Original Assignee
北京眼神科技有限公司
北京眼神智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京眼神科技有限公司, 北京眼神智能科技有限公司 filed Critical 北京眼神科技有限公司
Publication of WO2023061123A1 publication Critical patent/WO2023061123A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/141Control of illumination
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/40Spoof detection, e.g. liveness detection
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02BCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO BUILDINGS, e.g. HOUSING, HOUSE APPLIANCES OR RELATED END-USER APPLICATIONS
    • Y02B20/00Energy efficient lighting technologies, e.g. halogen lamps or gas discharge lamps
    • Y02B20/40Control techniques providing energy savings, e.g. smart controller or presence detection

Definitions

  • the present application relates to the field of face recognition, in particular to a method, device, storage medium and equipment for face silent liveness detection.
  • face recognition technology has been widely used in the fields of finance and security. Because the face has the advantages of easy acquisition and non-contact, but it is also very easy to be used by others to break through the face recognition system by taking photos or remakes of videos. Therefore, face liveness detection technology is particularly important as the first threshold of face recognition technology.
  • the first way is face movement liveness detection.
  • the liveness detection system issues random head and face movement instructions, and the user completes the corresponding actions according to the instructions to determine that the person is alive;
  • the first method is to extract the features used to distinguish the real face and the prosthetic face from the RGB image, and do the binary classification of the real face and the prosthetic face;
  • the third method is the integration of the first two methods.
  • a plurality of multi-expression RGB images are provided for the second way of human face liveness detection to determine whether it is a real human face.
  • Human face motion liveness detection requires a high degree of cooperation from users, and the random head and face motion commands issued are generally a single motion command, such as nodding, turning the head, blinking, and opening the mouth. There is a high probability that it can be deceived through the liveness detection system, or videos that take multiple actions of the user can also easily pass through the liveness detection system.
  • the RGB image is extracted to judge the features of the real face and the prosthetic face.
  • the method of live detection is silent live detection. This live detection method does not require user cooperation and is more easily accepted by users.
  • deep learning methods are generally used. Driven by a large amount of live and non-living face data, automatic learning can effectively distinguish the features of real faces and prosthetic faces. The difference between real and fake face imaging. However, due to differences in lighting, resolution, different mobile phone lenses and other imaging differences in RGB images, as well as through high-definition video playback, it is easy to fool the liveness detection algorithm and cause misjudgment.
  • a method, device, storage medium, and equipment for face silent liveness detection are provided.
  • the present application provides a method for face silent living body detection, the method comprising:
  • the supplementary light parameters include RGB color parameters and brightness parameters;
  • the preprocessed face image is input into a trained convolutional neural network to extract feature vectors, which are used to distinguish real faces and fake faces, and obtain living body scores according to the feature vectors; and According to the face image after the pretreatment, the light supplement parameter is regressed to obtain the light supplement parameter after regression;
  • the face image is from a real face, otherwise , it is determined that the face image is from a prosthetic face.
  • the acquisition of the face image collected under specific supplementary light parameters includes:
  • the supplementary light parameter is randomly changed every certain time interval, and the supplementary light parameter corresponding to the acquisition time is recorded as the specific supplementary light parameter.
  • the RGB color parameter is ⁇
  • the brightness parameter is ⁇
  • the RGB color parameter is calculated by the following formula:
  • R, G, and B are the R, G, and B values of the fill light, respectively, and ⁇ [0,1].
  • the convolutional neural network is trained by the following method:
  • the training sample set is input into the convolutional neural network, and the feature vector sample is extracted, and the feature vector sample is used to distinguish between a real human face and a prosthetic human face, and returns the fill light parameter according to the training sample set;
  • the preprocessing of the face image includes:
  • the face image is aligned and scaled by binocular coordinates to obtain the preprocessed face image.
  • the supplementary light is performed through the screen of the mobile terminal.
  • the method also includes:
  • the obtaining the living body score according to the feature vector includes:
  • the key points of the human face further include key points of the nose, key points of the left corner of the mouth, and key points of the right corner of the mouth.
  • the living body score is greater than a set threshold and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range, it is determined that the person The face image is from a real face, otherwise, it is determined that the face image is from a prosthetic face, including:
  • the face image is from a prosthetic face
  • the face image is from a prosthetic face.
  • the present application provides a face silent living body detection device, the device comprising:
  • An image acquisition module configured to acquire a face image collected under specific light supplement parameters, and preprocess the face image to obtain a preprocessed face image; wherein, the light supplement parameters include RGB color parameters and brightness parameters;
  • a processing module configured to input the preprocessed face image into a trained convolutional neural network to extract a feature vector, which is used to distinguish a real face from a prosthetic face, obtained according to the feature vector In vivo score; And according to the face image after the preprocessing regression supplementary light parameter, obtain the supplementary light parameter after regression;
  • a judging module configured to determine that the face image is from If it is a real face, otherwise, it is determined that the face image is from a prosthetic face.
  • the image acquisition module is specifically used for:
  • the supplementary light parameter is randomly changed every certain time interval, and the supplementary light parameter corresponding to the acquisition time is recorded as the specific supplementary light parameter.
  • the RGB color parameter is ⁇
  • the brightness parameter is ⁇
  • the device also includes:
  • a color parameter determination module configured to calculate the RGB color parameters by the following formula:
  • R, G, and B are the R, G, and B values of the fill light, respectively, and ⁇ [0,1].
  • the device also includes:
  • the sample acquisition module is used to obtain the face images collected under a plurality of supplementary light parameters, preprocess and add labels to obtain a training sample set; wherein, the labels include RGB color parameters ⁇ and brightness parameters ⁇ of supplementary light parameters;
  • the forward processing module is used for inputting the training sample set into the convolutional neural network, extracting feature vector samples for discriminating real faces and prosthetic faces, and regressing fill light parameters according to the training sample set;
  • the backpropagation module is used to calculate the loss of the eigenvector samples through ArcFace Loss, calculate the loss of the fill light parameter regression through the Euclidean loss function, and update the parameters of the convolutional neural network through backpropagation; wherein, when calculating the fill light parameter regression
  • the image acquisition module includes:
  • a face detection and positioning unit configured to perform face detection and key point positioning on the face image to obtain binocular coordinates
  • the human face normalization unit is used to align and scale the human face image through binocular coordinates to obtain the preprocessed human face image.
  • the device also includes:
  • the calculation module is used to calculate the Euclidean distance between the regressed fill light parameter and the specific fill light parameter, and obtain the difference between the regressed fill light parameter and the specific fill light parameter.
  • the device also includes:
  • the calculation module is used to calculate the Euclidean distance between the regressed fill light parameter and the specific fill light parameter, and obtain the difference between the regressed fill light parameter and the specific fill light parameter.
  • the processing module is specifically used for:
  • the judging module is specifically used for when the living body score is greater than a set threshold, and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range , it is determined that the face image is from a real face, otherwise, it is determined that the face image is from a fake face, including:
  • the face image is from a prosthetic face
  • the face image is from a prosthetic face.
  • the present application provides a computer-readable storage medium for face silent liveness detection, including a memory for storing processor-executable instructions, and when the instructions are executed by the processor, the first aspect is implemented. The steps of the described human face silent living body detection method.
  • the present application provides a device for face silent liveness detection, including at least one processor and a memory storing computer-executable instructions.
  • the processor executes the instructions, the human body described in the first aspect is realized Steps of face silent liveness detection method.
  • Fig. 1 is the flow chart of the face silent life detection method according to one or more embodiments
  • FIG. 2 is a schematic diagram of a training process of a convolutional neural network according to one or more embodiments
  • FIG. 3 is a flowchart of a training method of a convolutional neural network according to one or more embodiments
  • Fig. 4 is a schematic diagram of a face silent liveness detection device according to one or more embodiments.
  • the ultra-high-definition capability of the lens and screen The ultra-high-definition face video or image captured by the ultra-high-definition lens is played on the ultra-high-definition screen. To a certain extent, it is difficult for the human eye to distinguish it from the real face. Algorithmic challenge.
  • the embodiment of the present application provides a face silent liveness detection method, as shown in Figure 1, the method includes:
  • S100 Obtain a face image collected under specific fill light parameters, perform preprocessing on the face image, and obtain a preprocessed face image.
  • the fill light parameters include RGB color parameters and brightness parameters.
  • the fill light parameters have four parameters.
  • the first three are fill light RGB values, namely RGB color parameters, including R color parameters, G color parameters, and B color parameters.
  • the parameter represents the color parameter of red
  • the G color parameter represents the color parameter of green
  • the B color parameter represents the color parameter of blue, that is, which color of light is to be supplemented.
  • the fourth parameter is the brightness parameter.
  • supplementary light with specific supplementary light parameters may be provided, so that a human face image may be collected under the specific supplementary light parameters. Based on this, it is possible to record the specific supplementary light parameter, and record the face image collected under the specific parameter.
  • the face image is preprocessed to obtain the preprocessed face image.
  • S200 Input the preprocessed face image into the trained convolutional neural network, extract the feature vector, and obtain the in vivo score according to the feature vector; and return the fill light parameter according to the preprocessed face image to obtain the return fill light light parameters.
  • the eigenvectors are used to discriminate real faces and prosthetic faces.
  • the preprocessed face image can be input to the pre-trained convolutional neural network to obtain the output result of the convolutional neural network, which is the feature vector of the preprocessed face image . Calculations are performed based on the eigenvectors to obtain the score of the living body. It is also possible to perform reverse reasoning based on the pre-trained convolutional neural network and the pre-processed face image to regress the fill light parameters.
  • the convolutional neural network of the present application not only obtains the living body score for living body detection, but also performs reverse reasoning based on the input face image, and returns to obtain the supplementary light parameters, that is, obtains the supplementary light parameters after regression.
  • the regressed fill light parameters are used for verification with the aforementioned specific fill light parameters.
  • This application compares the regressed supplementary light parameters with specific supplementary light parameters, and calculates the difference, which can distinguish whether the collected face video or image is played on an ultra-high-definition screen.
  • the Euclidean distance between the light parameter and the specific fill light parameter is obtained.
  • the difference between the regressed fill light parameter and the specific fill light parameter is small, and for the face played on the screen, a layer of fill light information (the light played on the screen) is added to the played screen at the same time, so that There is a large deviation between the regressed fill light parameter and the specific fill light parameter, and this deviation can accurately reject the attack of the face on the high-definition screen.
  • the human face silent living body detection method can obtain the human face image collected under specific supplementary light parameters, and input the trained convolutional neural network to extract the feature vector, and according to the extracted feature vector Get the living body score, and at the same time regress to get the fill light parameters; judge whether it is a real face according to the live score, the difference between the regressed fill light parameters and the specific fill light parameters.
  • the material and 3D information differences between the real face and the prosthetic face are highlighted through RGB fill light, which is beneficial to distinguish between real and fake faces, and effectively alleviates the influence of ambient light; through face image fill light
  • the parameters are regressed, and the regressed fill light parameters are verified with specific fill light parameters, which solves the misjudgment of the high-definition remake of the prosthetic face.
  • RGB supplementary light can effectively avoid the influence of ambient light on the face silent liveness detection method, and the high-definition remake of ordinary faces, under the assistance of RGB supplementary light in this application, adds colorful lighting factors, so it is basically correct sentenced. That is, the face image is collected under RGB supplementary light, which can effectively alleviate the influence of ambient light; and the prosthetic face of different materials has obvious difference in surface material compared with the real face skin.
  • the difference can be amplified, which is more conducive to distinguishing real and fake faces; human faces contain rich 3D information, while photos and screen prosthetic faces are basically flat (or there is no rich 3D information of real faces), RGB fill light
  • the method of irradiating the human face can enlarge the 3D information difference between the real face and the prosthetic face, which is also beneficial to distinguish the real face from the fake face.
  • the supplementary light parameters are randomly changed at regular intervals, and the supplementary light parameters corresponding to the acquisition time are recorded as specific supplementary light parameters.
  • the front camera of the mobile phone is used to collect face images.
  • the mobile phone screen randomly changes the light of different colors and brightness, collects face images, and records The corresponding fill light parameters.
  • the aforementioned RGB color parameter is ⁇ , and the brightness parameter is ⁇ ;
  • R, G, and B are the R, G, and B values of the fill light, respectively, and ⁇ [0,1].
  • the R, G, and B values of the fill light are divided by 255, and the value range is controlled at [0,1], in order to facilitate convergence during training.
  • the aforementioned convolutional neural network can be a ResNet (Residual Network, deep residual) network, which can be obtained by training as shown in Figure 3:
  • ResNet Residual Network, deep residual
  • S10 Acquire face images collected under multiple fill light parameters, preprocess the face images, and add labels to the face images to obtain a training sample set.
  • the label includes the RGB color parameter ⁇ and the brightness parameter ⁇ among the fill light parameters, and ⁇ and parameter ⁇ constitute a 1*4-dimensional vector.
  • the preprocessing in this step is the same as the preprocessing method in S100.
  • the preprocessing includes face preprocessing mainly includes face detection, face key point location, head pose estimation, and face normalization. of:
  • MTCNN Multi-task convolutional neural network, multi-task convolutional neural network detection algorithm to realize face detection and face feature point location.
  • face detection and face feature point location For example, the following five key points of face can be used, including left eye, right eye, Nose, left mouth corner, right mouth corner.
  • the 3D attitude information is estimated by the 2D coordinate information of the key points, and the algorithm used for estimation can be the SolvePnP algorithm (monocular relative pose estimation function) in OpenCV.
  • the normalization operation including face alignment and scaling is performed on the face image: for example, the eyes can be aligned to (94,108) and (129,108), the nose can be aligned to (112,128), and the corners of the mouth can be aligned to (98,148) and ( 126,148), scaled to a size of (224,224).
  • S20 Input the training sample set into the convolutional neural network, extract feature vector samples, the feature vector samples are used to distinguish real faces and prosthetic faces, and return fill light parameters according to the training sample set.
  • the method further includes: supplementing light through the screen of the mobile terminal.
  • the method provided by the embodiment of the present disclosure can be applied to mobile terminals such as mobile phones, and the RGB supplementary light of different brightnesses is performed through the screen of the mobile terminal in consideration of the actual usage scenarios of the mobile terminal. It can make full use of the advantages of mobile terminal devices without adding additional lighting hardware; and use screen lighting to provide uniform lighting on the plane, unify the light intensity to a certain extent, and effectively alleviate the impact of ambient light.
  • embodiments of the present application are not limited to mobile terminals, and can also be applied to other devices other than mobile terminals, providing RGB supplementary light with different brightnesses through additional supplementary light hardware instead of the screen supplementary light.
  • the method also includes:
  • the gap between the regressed fill light parameter and a specific fill light parameter can be accurately determined by calculating the Euclidean distance.
  • the obtaining the living body score according to the feature vector includes:
  • the key points of the human face further include key points of the nose, key points of the left corner of the mouth, and key points of the right corner of the mouth.
  • the face image is aligned and scaled through binocular coordinates to obtain a preprocessed face image, including:
  • Scaling is performed on the face image to obtain a preprocessed face image.
  • the consistency of each data sample in the sample data set can be ensured and the training effect can be improved by preprocessing the alignment and scaling of the face images.
  • steps in the flow charts shown in FIG. 1 and FIG. 3 are displayed sequentially as indicated by the arrows, these steps are not necessarily executed sequentially in the order indicated by the arrows. Unless otherwise specified herein, there is no strict order restriction on the execution of these steps, and these steps can be executed in other orders. Moreover, at least some of the steps in FIG. 1 and FIG. 3 may include a plurality of sub-steps or stages, and these sub-steps or stages are not necessarily executed at the same time, but may be executed at different times. These steps or stages The order of execution is not necessarily performed sequentially, but may be performed alternately or alternately with at least a part of other steps or sub-steps or stages in other steps.
  • the embodiment of the present application provides a face silent liveness detection device, as shown in Figure 4, the device includes:
  • the image acquisition module 401 is configured to acquire a human face image collected under specific supplementary light parameters, and perform preprocessing on the human face image to obtain a preprocessed human face image; wherein the supplementary light parameters include RGB colors parameter and brightness parameter.
  • the processing module 402 is used to input the preprocessed face image into a trained convolutional neural network to extract a feature vector, which is used to distinguish a real face from a prosthetic face, according to the feature vector Obtaining the living body score; and regressing the supplementary light parameters according to the preprocessed face image to obtain the regression supplementary light parameters.
  • Judging module 403 used to determine that the face image is from a real face when the score of the living body is greater than the set threshold and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range;
  • the face image is from a prosthetic face.
  • This application obtains the face image collected under the specific supplementary light parameters, and inputs the trained convolutional neural network, extracts the feature vector, obtains the score of the living body according to the extracted feature vector, and returns the supplementary light parameter at the same time; Judging whether it is a real face or not according to the living body score, the difference between the regressed fill light parameter and the specific fill light parameter.
  • This application highlights the difference in material and 3D information between the real face and the prosthetic face through RGB fill light, which is beneficial to distinguish between real and fake faces, and effectively alleviates the impact of ambient light; regression is performed through face image fill light parameters , and verify the regressed fill light parameters with specific fill light parameters, which solves the misjudgment of the high-definition remake of the prosthetic face.
  • the face image collected under specific supplementary light parameters including:
  • the supplementary light parameters are randomly changed at regular intervals, and the supplementary light parameters corresponding to the acquisition time are recorded as specific supplementary light parameters.
  • RGB color parameter
  • brightness parameter
  • RGB color parameter is calculated by the following formula:
  • R, G, and B are the R, G, and B values of the fill light, respectively, and ⁇ [0,1].
  • the convolutional neural network of this application is obtained through the following module training:
  • the sample acquisition module is used to acquire face images collected under multiple light supplement parameters, perform preprocessing on the face images, and add labels to the face images to obtain a training sample set.
  • the label includes an RGB color parameter ⁇ and a brightness parameter ⁇ of the fill light parameter.
  • the forward processing module is used to input the training sample set into the convolutional neural network, extract feature vector samples, and the feature vector samples are used to distinguish real faces and fake faces, and return and complement the training sample set according to the training sample set. light parameters.
  • the backpropagation module is used to calculate the loss of feature vector samples through ArcFace Loss, calculate the loss of fill light parameter regression through the Euclidean loss function, and update the parameters of the convolutional neural network through backpropagation.
  • the image acquisition module includes:
  • the face detection and positioning unit is used to perform face detection and face key point positioning on the face image to obtain binocular coordinates.
  • the human face normalization unit is used for aligning and scaling human face images through binocular coordinates.
  • This application can be used in mobile terminals, and supplementary light can be performed through the screen of the mobile terminal.
  • the device also includes:
  • the calculation module is used to calculate the Euclidean distance between the regressed fill light parameter and the specific fill light parameter, and obtain the difference between the regressed fill light parameter and the specific fill light parameter.
  • the processing module is specifically used for:
  • the judging module is specifically used for when the living body score is greater than a set threshold, and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range , it is determined that the face image is from a real face, otherwise, it is determined that the face image is from a fake face, including:
  • the face image is from a prosthetic face
  • the face image is from a prosthetic face.
  • Each module in the above-mentioned face silent living body detection device can be fully or partially realized by software, hardware and combinations thereof.
  • the network interface may be an Ethernet card or a wireless network card or the like.
  • the above-mentioned modules can be embedded in or independent of the processor in the computer device in the form of hardware, and can also be stored in the memory of the computer device in the form of software, so that the processor can invoke and execute the corresponding operations of the above-mentioned modules.
  • a component may be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer.
  • a component may be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer.
  • an application running on a server and a server can be a component.
  • One or more components can reside within a process and/or thread of execution, and a component can be localized on one computer and/or distributed between two or more computers.
  • the methods described in the above embodiments provided by this application can implement business logic through computer programs and record them on a storage medium, and the storage medium can be read and executed by a computer to implement the business logic described in the embodiments of this specification.
  • the method describes the effect of the program. Therefore, the present application also provides a computer-readable storage medium for face silent liveness detection, including a memory for storing processor-executable instructions, and when the instructions are executed by the processor, the face silent liveness detection method comprising the above-mentioned embodiments is realized A step of.
  • This application highlights the difference in material and 3D information between the real face and the prosthetic face through RGB fill light, which is beneficial to distinguish between real and fake faces, and effectively alleviates the impact of ambient light; regression is performed through face image fill light parameters , and verify the regressed fill light parameters with specific fill light parameters, which solves the misjudgment of the high-definition remake of the prosthetic face.
  • the storage medium may include a physical device for storing information, and information is usually digitized and then stored using an electrical, magnetic, or optical medium. Described storage medium can include: the device that utilizes electric energy mode to store information such as, various memory, as RAM, ROM etc.; USB stick; a device that stores information optically, such as a CD or DVD. Of course, there are other readable storage media, such as quantum memory, graphene memory and so on.
  • the present application also provides a device for face silent liveness detection, which may be a separate computer, or may include one or more of the methods or one or more of the methods described in this specification.
  • the device for face silent life detection may include at least one processor and a memory storing computer-executable instructions. When the processor executes the instructions, the face silent life detection method described in any one or more embodiments above is implemented. step.
  • This application highlights the difference in material and 3D information between the real face and the prosthetic face through RGB fill light, which is beneficial to distinguish between real and fake faces, and effectively alleviates the impact of ambient light; regression is performed through face image fill light parameters , and verify the regressed fill light parameters with specific fill light parameters, which solves the misjudgment of the high-definition remake of the prosthetic face.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Molecular Biology (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

A facial silent living body detection method and apparatus, and a storage medium and a device. The method comprises: acquiring a facial image that is collected on the basis of a specific light supplementing parameter, and performing pre-processing (S100); inputting the facial image into a trained convolutional neural network, extracting a feature vector, obtaining a living body score according to the extracted feature vector, and regressing a light supplementing parameter according to the facial image (S200); and when the living body score is greater than a set threshold value and the difference between the regressed light supplementing parameter and the specific light supplementing parameter is within a preset range, determining that the facial image is from a real face, otherwise, determining that the facial image is from a fake face (S300).

Description

人脸静默活体检测方法、装置、存储介质及设备Face silent liveness detection method, device, storage medium and equipment
本申请要求于2021年10月15日申请的,申请号为202111201839.8,名称为“人脸静默活体检测方法、装置、存储介质及设备”的中国专利申请的优先权,在此将其全文作为参考This application claims the priority of the Chinese patent application filed on October 15, 2021, with the application number 202111201839.8, and the title is "Method, device, storage medium and equipment for face silent liveness detection", the full text of which is hereby taken as a reference
技术领域technical field
本申请涉及人脸识别领域,特别是指一种人脸静默活体检测方法、装置、存储介质及设备。The present application relates to the field of face recognition, in particular to a method, device, storage medium and equipment for face silent liveness detection.
背景技术Background technique
目前,人脸识别技术已被广泛应用于金融和安全领域。由于人脸具有获取方便,非接触等优点,但也极易被他人利用,以照片或翻拍视频的方式,攻破人脸识别系统。因此,人脸活体检测技术作为人脸识别技术第一道门槛,显的尤为重要。At present, face recognition technology has been widely used in the fields of finance and security. Because the face has the advantages of easy acquisition and non-contact, but it is also very easy to be used by others to break through the face recognition system by taking photos or remakes of videos. Therefore, face liveness detection technology is particularly important as the first threshold of face recognition technology.
目前,移动端人脸活体检测主要有三种方式,第一种方式是人脸动作活体检测,活体检测系统下发随机头脸部动作指令,用户按指令完成相应的动作则判断为活体;第二种方式是对RGB图像提取用于判别真实人脸和假体人脸的特征,做真实人脸和假体人脸的二分类;第三种方式是前两种方式的整合,在利用第一种方式进行人脸动作活体检测的过程中,为第二种方式的人脸活体检测提供多张多表情的RGB图像用于判别是否为真实人脸。At present, there are mainly three methods for face liveness detection on the mobile terminal. The first way is face movement liveness detection. The liveness detection system issues random head and face movement instructions, and the user completes the corresponding actions according to the instructions to determine that the person is alive; The first method is to extract the features used to distinguish the real face and the prosthetic face from the RGB image, and do the binary classification of the real face and the prosthetic face; the third method is the integration of the first two methods. In the process of performing human face liveness detection in the first way, a plurality of multi-expression RGB images are provided for the second way of human face liveness detection to determine whether it is a real human face.
人脸动作活体检测需要用户高度配合,且下发的随机头脸部动作指令一般为单一的动作指令,比如点头、转头、眨眼和张嘴等,不法分子通过抖动、扭曲或旋转照片等动作有很大概率能欺骗通过活体检测系统,或者拍摄用户的多个动作的视频同样能较轻松通过活体检测系统。Human face motion liveness detection requires a high degree of cooperation from users, and the random head and face motion commands issued are generally a single motion command, such as nodding, turning the head, blinking, and opening the mouth. There is a high probability that it can be deceived through the liveness detection system, or videos that take multiple actions of the user can also easily pass through the liveness detection system.
第二种方式中对RGB图像提取用于判断真实人脸和假体人脸的特征 进行活体检测的方式为静默活体检测,这种活体检测方式无需用户配合,更容易被用户所接受。提取用于判断真实人脸和假体人脸的特征时一般使用深度学习方法,通过大量的活体和非活体人脸数据驱动,自动学习能够有效判别真实人脸和假体人脸的特征,区分真假人脸成像差异。但是RGB图像由于光照、分辨率、不同手机镜头等成像差异,以及通过高清视频播放方式,容易骗过活体检测算法,造成误判。In the second method, the RGB image is extracted to judge the features of the real face and the prosthetic face. The method of live detection is silent live detection. This live detection method does not require user cooperation and is more easily accepted by users. When extracting features for judging real faces and prosthetic faces, deep learning methods are generally used. Driven by a large amount of live and non-living face data, automatic learning can effectively distinguish the features of real faces and prosthetic faces. The difference between real and fake face imaging. However, due to differences in lighting, resolution, different mobile phone lenses and other imaging differences in RGB images, as well as through high-definition video playback, it is easy to fool the liveness detection algorithm and cause misjudgment.
发明内容Contents of the invention
根据本申请的各种实施例,提供一种人脸静默活体检测方法、装置、存储介质及设备。According to various embodiments of the present application, a method, device, storage medium, and equipment for face silent liveness detection are provided.
本申请提供技术方案如下:This application provides technical scheme as follows:
第一方面,本申请提供一种人脸静默活体检测方法,所述方法包括:In a first aspect, the present application provides a method for face silent living body detection, the method comprising:
获取在特定的补光参数下采集的人脸图像,对所述人脸图像进行预处理,得到预处理后的人脸图像;其中,所述补光参数包括RGB色彩参数和亮度参数;Obtaining a face image collected under specific supplementary light parameters, performing preprocessing on the human face image to obtain a preprocessed human face image; wherein, the supplementary light parameters include RGB color parameters and brightness parameters;
将所述预处理后的人脸图像输入经过训练的卷积神经网络,提取特征向量,所述特征向量用于判别真实人脸和假体人脸,根据所述特征向量得到活体分值;并根据所述预处理后的人脸图像回归补光参数,得到回归后的补光参数;The preprocessed face image is input into a trained convolutional neural network to extract feature vectors, which are used to distinguish real faces and fake faces, and obtain living body scores according to the feature vectors; and According to the face image after the pretreatment, the light supplement parameter is regressed to obtain the light supplement parameter after regression;
当所述活体分值大于设定的阈值,且所述回归后的补光参数与所述特定的补光参数的差距在预设范围内时,判定所述人脸图像来自真实人脸,否则,判定所述人脸图像来自假体人脸。When the living body score is greater than the set threshold, and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range, it is determined that the face image is from a real face, otherwise , it is determined that the face image is from a prosthetic face.
在一实施例中,所述获取在特定的补光参数下采集的人脸图像,包括:In one embodiment, the acquisition of the face image collected under specific supplementary light parameters includes:
在采集人脸图像的过程中,每间隔一定时间随机变换补光参数,并记录采集时刻对应的补光参数作为所述特定的补光参数。During the process of collecting the face image, the supplementary light parameter is randomly changed every certain time interval, and the supplementary light parameter corresponding to the acquisition time is recorded as the specific supplementary light parameter.
在一实施例中,所述RGB色彩参数为α,亮度参数为β,通过以下公式计算所述RGB色彩参数:In one embodiment, the RGB color parameter is α, the brightness parameter is β, and the RGB color parameter is calculated by the following formula:
α=(R/255,G/255,B/255),α = (R/255, G/255, B/255),
其中,R、G、B分别为补光的R、G、B值,β∈[0,1]。Among them, R, G, and B are the R, G, and B values of the fill light, respectively, and β∈[0,1].
在一实施例中,所述卷积神经网络通过如下方法训练得到:In one embodiment, the convolutional neural network is trained by the following method:
获取多个补光参数下采集的人脸图像,对所述人脸图像进行预处理,并对所述人脸图像添加标签,得到训练样本集;其中,所述标签包括补光参数中的RGB色彩参数α和亮度参数β;Acquiring face images collected under a plurality of fill light parameters, preprocessing the face images, and adding labels to the face images to obtain a training sample set; wherein the labels include RGB in the fill light parameters Color parameter α and brightness parameter β;
将所述训练样本集输入卷积神经网络,提取特征向量样本,所述特征向量样本用于判别真实人脸和假体人脸,并根据所述训练样本集回归补光参数;The training sample set is input into the convolutional neural network, and the feature vector sample is extracted, and the feature vector sample is used to distinguish between a real human face and a prosthetic human face, and returns the fill light parameter according to the training sample set;
通过ArcFace Loss计算所述特征向量样本的损失,通过欧式loss函数计算补光参数回归的损失,并通过反向传播更新所述卷积神经网络的参数;其中,在计算补光参数回归的损失时,假体训练样本的回归量包括RGB色彩参数和亮度参数,其中,α=(0,0,0),β=0。Calculate the loss of the eigenvector sample by ArcFace Loss, calculate the loss of the fill light parameter regression through the Euclidean loss function, and update the parameters of the convolutional neural network through backpropagation; wherein, when calculating the loss of the fill light parameter regression , the regressors of the prosthetic training samples include RGB color parameters and brightness parameters, where α=(0,0,0), β=0.
在一实施例中,所述对所述人脸图像进行预处理,包括:In one embodiment, the preprocessing of the face image includes:
对所述人脸图像进行人脸检测和人脸关键点定位,得到双眼坐标;Carry out face detection and face key point location to described face image, obtain binocular coordinates;
通过双眼坐标对人脸图像进行对齐和缩放,得到预处理后的人脸图像。The face image is aligned and scaled by binocular coordinates to obtain the preprocessed face image.
在一实施例中,通过移动终端的屏幕进行补光。In an embodiment, the supplementary light is performed through the screen of the mobile terminal.
在一实施例中,所述方法还包括:In one embodiment, the method also includes:
计算所述回归后的补光参数与所述特定的补光参数的欧氏距离,得到所述回归后的补光参数与所述特定的补光参数的差距。Calculate the Euclidean distance between the regressed fill light parameter and the specific fill light parameter, and obtain the difference between the regressed fill light parameter and the specific fill light parameter.
在一实施例中,所述根据所述特征向量得到活体分值,包括:In one embodiment, the obtaining the living body score according to the feature vector includes:
将所述特征向量输入至ArcFace Loss,得到所述活体分值。Input the feature vector into ArcFace Loss to get the living body score.
在一实施例中,所述人脸关键点还包括鼻子关键点、左嘴角关键点息以及右嘴角关键点。In an embodiment, the key points of the human face further include key points of the nose, key points of the left corner of the mouth, and key points of the right corner of the mouth.
在一实施例中,所述当所述活体分值大于设定的阈值,且所述回归后的补光参数与所述特定的补光参数的差距在预设范围内时,判定所述人脸图像来自真实人脸,否则,判定所述人脸图像来自假体人脸,包括:In one embodiment, when the living body score is greater than a set threshold and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range, it is determined that the person The face image is from a real face, otherwise, it is determined that the face image is from a prosthetic face, including:
在所述活体分值大于或者等于设定的阈值,计算所述回归后的补光参数与所述特定的补光参数的差距;When the living body score is greater than or equal to a set threshold, calculate the difference between the regressed fill light parameter and the specific fill light parameter;
在所述差距在预设范围内的情况下,确定所述人脸图像来自真实人脸,When the difference is within a preset range, it is determined that the face image is from a real face,
在所述差距不在预设范围内的情况下,确定所述人脸图像来自假体人脸;When the difference is not within the preset range, it is determined that the face image is from a prosthetic face;
在所述活体分值小于所述设定的阈值的情况下,确定所述人脸图像来自假体人脸。In a case where the living body score is smaller than the set threshold, it is determined that the face image is from a prosthetic face.
第二方面,本申请提供一种人脸静默活体检测装置,所述装置包括:In a second aspect, the present application provides a face silent living body detection device, the device comprising:
图像获取模块,用于获取在特定的补光参数下采集的人脸图像,对所述人脸图像进行预处理,得到预处理后的人脸图像;其中,所述补光参数包括RGB色彩参数和亮度参数;An image acquisition module, configured to acquire a face image collected under specific light supplement parameters, and preprocess the face image to obtain a preprocessed face image; wherein, the light supplement parameters include RGB color parameters and brightness parameters;
处理模块,用于将所述预处理后的人脸图像输入经过训练的卷积神经网络,提取特征向量,所述特征向量用于判别真实人脸和假体人脸,根据所述特征向量得到活体分值;并根据所述预处理后的人脸图像回归补光参数,得到回归后的补光参数;A processing module, configured to input the preprocessed face image into a trained convolutional neural network to extract a feature vector, which is used to distinguish a real face from a prosthetic face, obtained according to the feature vector In vivo score; And according to the face image after the preprocessing regression supplementary light parameter, obtain the supplementary light parameter after regression;
判断模块,用于当所述活体分值大于设定的阈值,且所述回归后的补光参数与所述特定的补光参数的差距在预设范围内时,判定所述人脸图像来自真实人脸,否则,判定所述人脸图像来自假体人脸。A judging module, configured to determine that the face image is from If it is a real face, otherwise, it is determined that the face image is from a prosthetic face.
在一实施例中,所述图像获取模块具体用于:In one embodiment, the image acquisition module is specifically used for:
在采集人脸图像的过程中,每间隔一定时间随机变换补光参数,并记录采集时刻对应的补光参数作为所述特定的补光参数。During the process of collecting the face image, the supplementary light parameter is randomly changed every certain time interval, and the supplementary light parameter corresponding to the acquisition time is recorded as the specific supplementary light parameter.
在一实施例中,所述RGB色彩参数为α,亮度参数为β;所述装置还包括:In one embodiment, the RGB color parameter is α, and the brightness parameter is β; the device also includes:
色彩参数确定模块,用于通过以下公式计算所述RGB色彩参数:A color parameter determination module, configured to calculate the RGB color parameters by the following formula:
α=(R/255,G/255,B/255),α = (R/255, G/255, B/255),
其中,R、G、B分别为补光的R、G、B值,β∈[0,1]。Among them, R, G, and B are the R, G, and B values of the fill light, respectively, and β∈[0,1].
在一实施例中,所述装置还包括:In one embodiment, the device also includes:
样本获取模块,用于获取多个补光参数下采集的人脸图像,进行预处理并添加标签,得到训练样本集;其中,所述标签包括补光参数的RGB色彩参数α和亮度参数β;The sample acquisition module is used to obtain the face images collected under a plurality of supplementary light parameters, preprocess and add labels to obtain a training sample set; wherein, the labels include RGB color parameters α and brightness parameters β of supplementary light parameters;
前向处理模块,用于将所述训练样本集输入卷积神经网络,提取用于判别真实人脸和假体人脸的特征向量样本,并根据训练样本集回归补光参数;The forward processing module is used for inputting the training sample set into the convolutional neural network, extracting feature vector samples for discriminating real faces and prosthetic faces, and regressing fill light parameters according to the training sample set;
反向传播模块,用于通过ArcFace Loss计算特征向量样本的损失,通过欧式loss函数计算补光参数回归的损失,并通过反向传播更新卷积神经网络的参数;其中,在计算补光参数回归的损失时,假体训练样本的回归量包括RGB色彩参数和亮度参数,其中,α=(0,0,0),β=0。The backpropagation module is used to calculate the loss of the eigenvector samples through ArcFace Loss, calculate the loss of the fill light parameter regression through the Euclidean loss function, and update the parameters of the convolutional neural network through backpropagation; wherein, when calculating the fill light parameter regression When the loss of , the regressor of the prosthetic training sample includes RGB color parameters and brightness parameters, where α=(0,0,0), β=0.
在一实施例中,所述图像获取模块包括:In one embodiment, the image acquisition module includes:
人脸检测和定位单元,用于对所述人脸图像进行人脸检测和人脸关键点定位,得到双眼坐标;A face detection and positioning unit, configured to perform face detection and key point positioning on the face image to obtain binocular coordinates;
人脸归一化单元,用于通过双眼坐标对人脸图像进行对齐和缩放,得到预处理后的人脸图像。The human face normalization unit is used to align and scale the human face image through binocular coordinates to obtain the preprocessed human face image.
在一实施例中,所述装置还包括:In one embodiment, the device also includes:
计算模块,用于计算所述回归后的补光参数与所述特定的补光参数的欧氏距离,得到所述回归后的补光参数与所述特定的补光参数的差距。The calculation module is used to calculate the Euclidean distance between the regressed fill light parameter and the specific fill light parameter, and obtain the difference between the regressed fill light parameter and the specific fill light parameter.
在一个示例中,所述装置还包括:In one example, the device also includes:
计算模块,用于计算所述回归后的补光参数与所述特定的补光参数的欧氏距离,得到所述回归后的补光参数与所述特定的补光参数的差距。The calculation module is used to calculate the Euclidean distance between the regressed fill light parameter and the specific fill light parameter, and obtain the difference between the regressed fill light parameter and the specific fill light parameter.
在一个示例中,所述处理模块具体用于:In one example, the processing module is specifically used for:
将所述特征向量输入至ArcFace Loss,得到所述活体分值。Input the feature vector into ArcFace Loss to get the living body score.
在一个示例中,该判断模块,具体用于所述当所述活体分值大于设定的阈值,且所述回归后的补光参数与所述特定的补光参数的差距在预设范围内时,判定所述人脸图像来自真实人脸,否则,判定所述人脸图像来自假体人脸,包括:In an example, the judging module is specifically used for when the living body score is greater than a set threshold, and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range , it is determined that the face image is from a real face, otherwise, it is determined that the face image is from a fake face, including:
在所述活体分值大于或者等于设定的阈值,计算所述回归后的补光 参数与所述特定的补光参数的差距;When the living body score is greater than or equal to the set threshold, calculate the difference between the regressed fill light parameter and the specific fill light parameter;
在所述差距在预设范围内的情况下,确定所述人脸图像来自真实人脸,When the difference is within a preset range, it is determined that the face image is from a real face,
在所述差距不在预设范围内的情况下,确定所述人脸图像来自假体人脸;When the difference is not within the preset range, it is determined that the face image is from a prosthetic face;
在所述活体分值小于所述设定的阈值的情况下,确定所述人脸图像来自假体人脸。In a case where the living body score is smaller than the set threshold, it is determined that the face image is from a prosthetic face.
第三方面,本申请提供一种用于人脸静默活体检测的计算机可读存储介质,包括用于存储处理器可执行指令的存储器,所述指令被所述处理器执行时实现包括第一方面所述的人脸静默活体检测方法的步骤。In a third aspect, the present application provides a computer-readable storage medium for face silent liveness detection, including a memory for storing processor-executable instructions, and when the instructions are executed by the processor, the first aspect is implemented. The steps of the described human face silent living body detection method.
第四方面,本申请提供一种用于人脸静默活体检测的设备,包括至少一个处理器以及存储计算机可执行指令的存储器,所述处理器执行所述指令时实现第一方面所述的人脸静默活体检测方法的步骤。In a fourth aspect, the present application provides a device for face silent liveness detection, including at least one processor and a memory storing computer-executable instructions. When the processor executes the instructions, the human body described in the first aspect is realized Steps of face silent liveness detection method.
本申请的一个或多个实施例的细节在下面的附图和描述中提出。本申请的其他特征、目的和优点将从说明书、附图以及权利要求书变得明显。The details of one or more embodiments of the application are set forth in the accompanying drawings and the description below. Other features, objects and advantages of the present application will be apparent from the description, drawings and claims.
附图说明Description of drawings
图1为根据一个或多个实施例的人脸静默活体检测方法的流程图;Fig. 1 is the flow chart of the face silent life detection method according to one or more embodiments;
图2为根据一个或多个实施例的卷积神经网络的训练过程的示意图;2 is a schematic diagram of a training process of a convolutional neural network according to one or more embodiments;
图3为根据一个或多个实施例的卷积神经网络的训练方法的流程图;FIG. 3 is a flowchart of a training method of a convolutional neural network according to one or more embodiments;
图4为根据一个或多个实施例的人脸静默活体检测装置的示意图。Fig. 4 is a schematic diagram of a face silent liveness detection device according to one or more embodiments.
为了更好地描述和说明这里公开的那些发明的实施例和/或示例,可以参考一幅或多幅附图。用于描述附图的附加细节或示例不应当被认为是对所公开的发明、目前描述的实施例和/或示例以及目前理解的这些发明的最佳模式中的任何一者的范围的限制。In order to better describe and illustrate embodiments and/or examples of the inventions disclosed herein, reference may be made to one or more of the accompanying drawings. Additional details or examples used to describe the drawings should not be considered limitations on the scope of any of the disclosed inventions, the presently described embodiments and/or examples, and the best mode of these inventions currently understood.
具体实施方式Detailed ways
为使本申请要解决的技术问题、技术方案和优点更加清楚,下面将结合附图及具体实施例对本申请的技术方案进行清楚、完整地描述。显 然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。通常在此处附图中描述和示出的本申请实施例的组件可以以各种不同的配置来布置和设计。因此,以下对在附图中提供的本申请的实施例的详细描述并非旨在限制要求保护的本申请的范围,而是仅仅表示本申请的选定实施例。基于本申请的实施例,本领域技术人员在没有做出创造性劳动的前提下所获得的所有其他实施例,都属于本申请保护的范围。In order to make the technical problems, technical solutions and advantages to be solved by the present application clearer, the technical solutions of the present application will be clearly and completely described below in conjunction with the accompanying drawings and specific embodiments. Apparently, the described embodiments are only a part of the embodiments of the present application, rather than all the embodiments. The components of the embodiments of the application generally described and illustrated in the figures herein may be arranged and designed in a variety of different configurations. Accordingly, the following detailed description of the embodiments of the application provided in the accompanying drawings is not intended to limit the scope of the claimed application, but merely represents selected embodiments of the application. Based on the embodiments of the present application, all other embodiments obtained by those skilled in the art without making creative efforts belong to the scope of protection of the present application.
在一个实施例中,目前,人脸活体检测主要有两个影响因素:一是环境光照,多变的环境光照一直是人脸各项任务的巨大挑战,环境光严重影响在人脸静默活体检测的准确性。二是镜头和屏幕的超高清能力,超高清镜头拍摄的超高清人脸视频或图像,在超高清屏幕上播放,一定程度上人眼很难将其与真实人脸区分,同样是对活体检测算法的巨大挑战。In one embodiment, at present, there are two main influencing factors for human face liveness detection: one is ambient light, which has always been a huge challenge for various tasks of human face, and ambient light seriously affects the silent liveness detection of human face. accuracy. The second is the ultra-high-definition capability of the lens and screen. The ultra-high-definition face video or image captured by the ultra-high-definition lens is played on the ultra-high-definition screen. To a certain extent, it is difficult for the human eye to distinguish it from the real face. Algorithmic challenge.
为解决上述问题,本申请实施例提供一种人脸静默活体检测方法,如图1所示,该方法包括:In order to solve the above problems, the embodiment of the present application provides a face silent liveness detection method, as shown in Figure 1, the method includes:
S100:获取在特定的补光参数下采集的人脸图像,对人脸图像进行预处理,得到预处理后的人脸图像。S100: Obtain a face image collected under specific fill light parameters, perform preprocessing on the face image, and obtain a preprocessed face image.
其中,补光参数包括RGB色彩参数和亮度参数,补光参数有四个参数,前三个是补光RGB数值,即RGB色彩参数,包括R色彩参数、G色彩参数以及B色彩参数,R色彩参数表示红色的色彩参数,G色彩参数表示绿色的色彩参数,B色彩参数表示蓝色的色彩参数,即补哪种色彩的光。第四个参数是亮度参数。Among them, the fill light parameters include RGB color parameters and brightness parameters. The fill light parameters have four parameters. The first three are fill light RGB values, namely RGB color parameters, including R color parameters, G color parameters, and B color parameters. The parameter represents the color parameter of red, the G color parameter represents the color parameter of green, and the B color parameter represents the color parameter of blue, that is, which color of light is to be supplemented. The fourth parameter is the brightness parameter.
在本申请实施例中,可以提供特定的补光参数的补光,这样,可以在特定的补光参数下,采集人脸图像。基于此,可以记录该特定的补光参数,以及记录在该特定的参数下采集到的人脸图像。对人脸图像进行预处理,得到预处理后的人脸图像。In the embodiment of the present application, supplementary light with specific supplementary light parameters may be provided, so that a human face image may be collected under the specific supplementary light parameters. Based on this, it is possible to record the specific supplementary light parameter, and record the face image collected under the specific parameter. The face image is preprocessed to obtain the preprocessed face image.
S200:将预处理后的人脸图像输入经过训练的卷积神经网络,提取特征向量,根据特征向量得到活体分值;并根据预处理后的人脸图像回归补光参数,得到回归后的补光参数。S200: Input the preprocessed face image into the trained convolutional neural network, extract the feature vector, and obtain the in vivo score according to the feature vector; and return the fill light parameter according to the preprocessed face image to obtain the return fill light light parameters.
其中,特征向量用于判别真实人脸和假体人脸。Among them, the eigenvectors are used to discriminate real faces and prosthetic faces.
在本实施例中,可以将预处理后的人脸图像输入至预先训练的卷积神经网络,得到该卷积神经网络的输出结果,该输出结果是该预处理后的人脸图像的特征向量。基于特征向量进行计算,得到活体分值。还可以基于预先训练的卷积神经网络以及预处理后的人脸图像进行反向推理,对补光参数进行回归。In this embodiment, the preprocessed face image can be input to the pre-trained convolutional neural network to obtain the output result of the convolutional neural network, which is the feature vector of the preprocessed face image . Calculations are performed based on the eigenvectors to obtain the score of the living body. It is also possible to perform reverse reasoning based on the pre-trained convolutional neural network and the pre-processed face image to regress the fill light parameters.
基于此,本申请的卷积神经网络不仅得到用于活体检测的活体分值,而且还根据输入的人脸图像进行反向推理,回归得到补光参数,即得到回归后的补光参数,该回归后的补光参数用于与前述的特定的补光参数进行校验。Based on this, the convolutional neural network of the present application not only obtains the living body score for living body detection, but also performs reverse reasoning based on the input face image, and returns to obtain the supplementary light parameters, that is, obtains the supplementary light parameters after regression. The regressed fill light parameters are used for verification with the aforementioned specific fill light parameters.
S300:当活体分值大于设定的阈值且回归后的补光参数与特定的补光参数的差距在预设范围内时,判定人脸图像来自真实人脸,否则,判定人脸图像来自假体人脸。S300: When the living body score is greater than the set threshold and the difference between the regressed fill light parameter and the specific fill light parameter is within the preset range, determine that the face image is from a real face; otherwise, determine that the face image is from a fake face. body face.
本申请将回归的补光参数与特定的补光参数进行比较,计算其差距,即可区分采集的是否是在超高清屏幕上播放的人脸视频或图像,所述差距可以通过计算回归的补光参数与特定的补光参数的欧氏距离得到。对于真实人脸,回归的补光参数与所述特定的补光参数差距较小,而对于屏幕播放的人脸,播放的屏幕同时又附加了一层补光信息(屏幕播放的光),使得回归的补光参数与所述特定的补光参数存在较大的偏差,通过这种偏差能够精准拒绝高清屏幕人脸的攻击。This application compares the regressed supplementary light parameters with specific supplementary light parameters, and calculates the difference, which can distinguish whether the collected face video or image is played on an ultra-high-definition screen. The Euclidean distance between the light parameter and the specific fill light parameter is obtained. For a real face, the difference between the regressed fill light parameter and the specific fill light parameter is small, and for the face played on the screen, a layer of fill light information (the light played on the screen) is added to the played screen at the same time, so that There is a large deviation between the regressed fill light parameter and the specific fill light parameter, and this deviation can accurately reject the attack of the face on the high-definition screen.
本申请实施例提供的人脸静默活体检测方法,可以获取在特定的补光参数下采集的人脸图像,并输入经过训练的卷积神经网络,提取特征向量,根据所提取的所述特征向量得到活体分值,同时回归得到补光参数;根据活体分值、回归后的补光参数与特定的补光参数的差距判断是否是真实人脸。通过采用本方法,通过RGB补光,凸出真实人脸与假体人脸的材质和3D信息差异,有利于区分真假人脸,并有效缓解了环境光的影响;通过人脸图像补光参数进行回归,并将回归后的补光参数与特定的补光参数进行校验,解决了对高清翻拍假体人脸的误判。The human face silent living body detection method provided in the embodiment of the present application can obtain the human face image collected under specific supplementary light parameters, and input the trained convolutional neural network to extract the feature vector, and according to the extracted feature vector Get the living body score, and at the same time regress to get the fill light parameters; judge whether it is a real face according to the live score, the difference between the regressed fill light parameters and the specific fill light parameters. By adopting this method, the material and 3D information differences between the real face and the prosthetic face are highlighted through RGB fill light, which is beneficial to distinguish between real and fake faces, and effectively alleviates the influence of ambient light; through face image fill light The parameters are regressed, and the regressed fill light parameters are verified with specific fill light parameters, which solves the misjudgment of the high-definition remake of the prosthetic face.
也就是说,RGB补光能有效避免环境光照对人脸静默活体检测方法的影响,并且普通人脸高清翻拍,在本申请的RGB补光辅助情况下,增加了多彩光照因素,因此也基本无误判。即,在RGB补光下采集人脸图像,补光可以有效缓解环境光的影响;并且不同材质的假体人脸,相对真实人脸肌肤,表面材质差异明显,利用不同强度的RGB补光,能将该差异放大,更有利于判别真假人脸;人脸含有丰富的3D信息,而照片和屏幕假体人脸基本为平面(或者没有真实人脸丰富的3D信息),以RGB补光的方式照射在人脸上,能够放大真实人脸与假体人脸的3D信息差异,同样有利于判别真假人脸。That is to say, RGB supplementary light can effectively avoid the influence of ambient light on the face silent liveness detection method, and the high-definition remake of ordinary faces, under the assistance of RGB supplementary light in this application, adds colorful lighting factors, so it is basically correct sentenced. That is, the face image is collected under RGB supplementary light, which can effectively alleviate the influence of ambient light; and the prosthetic face of different materials has obvious difference in surface material compared with the real face skin. The difference can be amplified, which is more conducive to distinguishing real and fake faces; human faces contain rich 3D information, while photos and screen prosthetic faces are basically flat (or there is no rich 3D information of real faces), RGB fill light The method of irradiating the human face can enlarge the 3D information difference between the real face and the prosthetic face, which is also beneficial to distinguish the real face from the fake face.
在一个示例中,在采集人脸图像的过程中,每间隔一定时间随机变换补光参数,并记录采集时刻对应的补光参数作为特定的补光参数。In one example, during the process of capturing the face image, the supplementary light parameters are randomly changed at regular intervals, and the supplementary light parameters corresponding to the acquisition time are recorded as specific supplementary light parameters.
以移动终端为例,使用手机前置摄像头采集人脸图像,在采集过程中,每间隔时间间隔(可记为τ),手机屏幕随机变换不同色彩和亮度的光,采集人脸图像,并记录对应的补光参数。Taking the mobile terminal as an example, the front camera of the mobile phone is used to collect face images. During the collection process, every time interval (which can be recorded as τ), the mobile phone screen randomly changes the light of different colors and brightness, collects face images, and records The corresponding fill light parameters.
由于随机生成参数补光,采集的人脸图像存在差异,通过人脸图像特征回归补光参数进行校验,从而有效增加对高清视频的防Hack能力。Due to the randomly generated parameter fill light, there are differences in the collected face images, and the fill light parameters are verified through the facial image feature regression, thereby effectively increasing the anti-hacking ability of high-definition videos.
在一个示例中,前述的RGB色彩参数为α,亮度参数为β;In an example, the aforementioned RGB color parameter is α, and the brightness parameter is β;
其中,α=(R/255,G/255,B/255),R、G、B分别为补光的R、G、B值,β∈[0,1]。Wherein, α=(R/255, G/255, B/255), R, G, and B are the R, G, and B values of the fill light, respectively, and β∈[0,1].
本申请将补光的R、G、B值分别除以255,将其取值范围控制在[0,1],为了在训练时更有利于收敛。In this application, the R, G, and B values of the fill light are divided by 255, and the value range is controlled at [0,1], in order to facilitate convergence during training.
在一个示例中,前述的卷积神经网络可以为ResNet(Residual Network,深度残差)网络,其可以通过如图3所示的方法训练得到:In an example, the aforementioned convolutional neural network can be a ResNet (Residual Network, deep residual) network, which can be obtained by training as shown in Figure 3:
S10:获取多个补光参数下采集的人脸图像,对人脸图像进行预处理,并对人脸图像添加标签,得到训练样本集。S10: Acquire face images collected under multiple fill light parameters, preprocess the face images, and add labels to the face images to obtain a training sample set.
其中,标签包括补光参数中的RGB色彩参数α和亮度参数β,α与参数β构成1*4维的向量。Among them, the label includes the RGB color parameter α and the brightness parameter β among the fill light parameters, and α and parameter β constitute a 1*4-dimensional vector.
本步骤的预处理与S100的预处理方法相同,在其中一个示例中,预处理包括人脸预处理主要包括人脸检测、人脸关键点定位、头部姿态估计和人脸归一化,具体的:The preprocessing in this step is the same as the preprocessing method in S100. In one example, the preprocessing includes face preprocessing mainly includes face detection, face key point location, head pose estimation, and face normalization. of:
使用MTCNN(Multi-task convolutional neural network,多任务卷积神经网络)检测算法,实现人脸检测和人脸特征点定位,例如,人脸关键点可采用如下5个,包括左眼、右眼、鼻子、左嘴角、右嘴角。Use MTCNN (Multi-task convolutional neural network, multi-task convolutional neural network) detection algorithm to realize face detection and face feature point location. For example, the following five key points of face can be used, including left eye, right eye, Nose, left mouth corner, right mouth corner.
通过关键点的2D坐标信息,估计3D姿态信息,进行估计所采用的算法可以为OpenCV中的SolvePnP算法(单目相对位姿估计函数)。The 3D attitude information is estimated by the 2D coordinate information of the key points, and the algorithm used for estimation can be the SolvePnP algorithm (monocular relative pose estimation function) in OpenCV.
通过双眼坐标,对人脸图像进行包括人脸对齐和缩放的归一化操作:例如可将双眼对齐到(94,108)和(129,108),鼻子对齐到(112,128),嘴角对齐到(98,148)和(126,148),缩放到大小为(224,224)。Through the coordinates of the eyes, the normalization operation including face alignment and scaling is performed on the face image: for example, the eyes can be aligned to (94,108) and (129,108), the nose can be aligned to (112,128), and the corners of the mouth can be aligned to (98,148) and ( 126,148), scaled to a size of (224,224).
S20:将训练样本集输入卷积神经网络,提取特征向量样本,特征向量样本用于判别真实人脸和假体人脸,并根据训练样本集回归补光参数。S20: Input the training sample set into the convolutional neural network, extract feature vector samples, the feature vector samples are used to distinguish real faces and prosthetic faces, and return fill light parameters according to the training sample set.
S30:通过ArcFace Loss计算特征向量样本的损失,通过欧式loss函数计算补光参数回归的损失,如图2所示,并通过反向传播更新卷积神经网络的参数。S30: Calculate the loss of the eigenvector samples through ArcFace Loss, calculate the loss of the fill light parameter regression through the Euclidean loss function, as shown in Figure 2, and update the parameters of the convolutional neural network through backpropagation.
本步骤中,采用欧式loss进行参数回归,LabelLoss对网络输出512维的特征向量,使用ArcFaceLoss(人脸识别损失函数)进行训练。这样,可以通过LabelLoss以及regression loss进行反向传播,从而更新卷积神经网络(resnet18网络)的参数。In this step, European loss is used for parameter regression, LabelLoss outputs a 512-dimensional feature vector to the network, and ArcFaceLoss (face recognition loss function) is used for training. In this way, the parameters of the convolutional neural network (resnet18 network) can be updated through Backpropagation through LabelLoss and regression loss.
本申请视真实人脸为闭集,假体人脸为开集。因为假体人脸攻击方式多样,与真实人脸相比是一个比较开放的集合,所以在计算补光参数回归的损失时,假体训练样本的回归量包括RGB色彩参数和亮度参数,其中,α=(0,0,0),β=0。This application regards the real face as a closed set, and the prosthetic face as an open set. Because the prosthetic face has various attack methods, it is a relatively open set compared with the real face, so when calculating the loss of the fill light parameter regression, the regressor of the prosthetic training sample includes RGB color parameters and brightness parameters, among them, α=(0,0,0), β=0.
基于以上方案,通过LabelLoss以及regression loss进行反向传播,可以达到更好的训练效果,提高卷积神经网络的精度。Based on the above scheme, backpropagation through LabelLoss and regression loss can achieve better training effect and improve the accuracy of convolutional neural network.
在一个示例中,该方法还包括:通过移动终端的屏幕进行补光。In an example, the method further includes: supplementing light through the screen of the mobile terminal.
在本申请实施例中,本公开实施例所提供的方法可以应用于手机等 移动终端,考虑移动终端的实际使用场景,通过移动终端的屏幕进行不同亮度的RGB补光。可以充分利用移动终端设备的优点,无需增加额外的补光硬件;并且使用屏幕补光,提供平面的均匀补光,在一定程度上统一光照强度,有效缓解环境光影响。In the embodiment of the present application, the method provided by the embodiment of the present disclosure can be applied to mobile terminals such as mobile phones, and the RGB supplementary light of different brightnesses is performed through the screen of the mobile terminal in consideration of the actual usage scenarios of the mobile terminal. It can make full use of the advantages of mobile terminal devices without adding additional lighting hardware; and use screen lighting to provide uniform lighting on the plane, unify the light intensity to a certain extent, and effectively alleviate the impact of ambient light.
当然,本申请实施例不仅限于移动终端,也可以应用于移动终端之外的其他设备,不通过屏幕补光,而通过额外的补光硬件,提供不同亮度的RGB补光。Of course, the embodiments of the present application are not limited to mobile terminals, and can also be applied to other devices other than mobile terminals, providing RGB supplementary light with different brightnesses through additional supplementary light hardware instead of the screen supplementary light.
基于以上方案,可以通过移动终端进行不同程度的补光,增加补光灵活性以及便捷性。Based on the above scheme, different degrees of supplementary light can be performed through the mobile terminal, increasing the flexibility and convenience of supplementary light.
在一个示例中,所述方法还包括:In one example, the method also includes:
计算所述回归后的补光参数与所述特定的补光参数的欧氏距离,得到所述回归后的补光参数与所述特定的补光参数的差距。Calculate the Euclidean distance between the regressed fill light parameter and the specific fill light parameter, and obtain the difference between the regressed fill light parameter and the specific fill light parameter.
基于以上方案,可以通过计算欧氏距离,准确地确定回归后的补光参数与特定的补光参数之间的差距。Based on the above solution, the gap between the regressed fill light parameter and a specific fill light parameter can be accurately determined by calculating the Euclidean distance.
在一个示例中,所述根据所述特征向量得到活体分值,包括:In an example, the obtaining the living body score according to the feature vector includes:
将所述特征向量输入至ArcFace Loss,得到所述活体分值。Input the feature vector into ArcFace Loss to get the living body score.
基于以上方案,可以通过提高人脸检测的准确度。Based on the above scheme, the accuracy of face detection can be improved.
在一个示例中,所述人脸关键点还包括鼻子关键点、左嘴角关键点息以及右嘴角关键点。In an example, the key points of the human face further include key points of the nose, key points of the left corner of the mouth, and key points of the right corner of the mouth.
在一个示例中,所述通过双眼坐标对人脸图像进行对齐和缩放,得到预处理后的人脸图像,包括:In an example, the face image is aligned and scaled through binocular coordinates to obtain a preprocessed face image, including:
对各所述人脸关键点的坐标信息进行对齐处理;Aligning the coordinate information of each of the key points of the human face;
对所述人脸图像进行缩放处理,得到预处理后的人脸图像。Scaling is performed on the face image to obtain a preprocessed face image.
基于以上方案,通过对人脸图像进行对齐以及缩放的预处理后,可以保证样本数据集中各数据样本的一致性,提高训练效果。Based on the above scheme, the consistency of each data sample in the sample data set can be ensured and the training effect can be improved by preprocessing the alignment and scaling of the face images.
应该理解的是,虽然图1以及图3所示流程图中的各个步骤按照箭头的指示依次显示,但是这些步骤并不是必然按照箭头指示的顺序依次执行。除非本文中有明确的说明,这些步骤的执行并没有严格的顺序限制, 这些步骤可以以其它的顺序执行。而且,图1以及图3中的至少一部分步骤可以包括多个子步骤或者多个阶段,这些子步骤或者阶段并不必然是在同一时刻执行完成,而是可以在不同的时刻执行,这些步骤或者阶段的执行顺序也不必然是依次进行,而是可以与其它步骤或者其它步骤中的子步骤或者阶段的至少一部分轮流或者交替地执行。It should be understood that although the steps in the flow charts shown in FIG. 1 and FIG. 3 are displayed sequentially as indicated by the arrows, these steps are not necessarily executed sequentially in the order indicated by the arrows. Unless otherwise specified herein, there is no strict order restriction on the execution of these steps, and these steps can be executed in other orders. Moreover, at least some of the steps in FIG. 1 and FIG. 3 may include a plurality of sub-steps or stages, and these sub-steps or stages are not necessarily executed at the same time, but may be executed at different times. These steps or stages The order of execution is not necessarily performed sequentially, but may be performed alternately or alternately with at least a part of other steps or sub-steps or stages in other steps.
在一个实施例中,本申请实施例提供一种人脸静默活体检测装置,如图4所示,该装置包括:In one embodiment, the embodiment of the present application provides a face silent liveness detection device, as shown in Figure 4, the device includes:
图像获取模块401,用于获取在特定的补光参数下采集的人脸图像,对所述人脸图像进行预处理,得到预处理后的人脸图像;其中,所述补光参数包括RGB色彩参数和亮度参数。The image acquisition module 401 is configured to acquire a human face image collected under specific supplementary light parameters, and perform preprocessing on the human face image to obtain a preprocessed human face image; wherein the supplementary light parameters include RGB colors parameter and brightness parameter.
处理模块402,用于将所述预处理后的人脸图像输入经过训练的卷积神经网络,提取特征向量,所述特征向量用于判别真实人脸和假体人脸,根据所述特征向量得到活体分值;并根据所述预处理后的人脸图像回归补光参数,得到回归后的补光参数。The processing module 402 is used to input the preprocessed face image into a trained convolutional neural network to extract a feature vector, which is used to distinguish a real face from a prosthetic face, according to the feature vector Obtaining the living body score; and regressing the supplementary light parameters according to the preprocessed face image to obtain the regression supplementary light parameters.
判断模块403,用于当活体分值大于设定的阈值且回归后的补光参数与特定的补光参数的差距在预设范围内时,判定人脸图像来自真实人脸,否则,判定人脸图像来自假体人脸。Judging module 403, used to determine that the face image is from a real face when the score of the living body is greater than the set threshold and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range; The face image is from a prosthetic face.
本申请获取在特定的补光参数下采集的人脸图像,并输入经过训练的卷积神经网络,提取特征向量,根据所提取的所述特征向量得到活体分值,同时回归得到补光参数;根据活体分值、回归后的补光参数与特定的补光参数的差距判断是否是真实人脸。本申请通过RGB补光,凸出真实人脸与假体人脸的材质和3D信息差异,有利于区分真假人脸,并有效缓解了环境光的影响;通过人脸图像补光参数进行回归,并将回归后的补光参数与特定的补光参数进行校验,解决了对高清翻拍假体人脸的误判。This application obtains the face image collected under the specific supplementary light parameters, and inputs the trained convolutional neural network, extracts the feature vector, obtains the score of the living body according to the extracted feature vector, and returns the supplementary light parameter at the same time; Judging whether it is a real face or not according to the living body score, the difference between the regressed fill light parameter and the specific fill light parameter. This application highlights the difference in material and 3D information between the real face and the prosthetic face through RGB fill light, which is beneficial to distinguish between real and fake faces, and effectively alleviates the impact of ambient light; regression is performed through face image fill light parameters , and verify the regressed fill light parameters with specific fill light parameters, which solves the misjudgment of the high-definition remake of the prosthetic face.
所述的图像获取模块中,获取在特定的补光参数下采集的人脸图像,包括:In the described image acquisition module, obtain the face image collected under specific supplementary light parameters, including:
在采集人脸图像的过程中,每间隔一定时间随机变换补光参数,并 记录采集时刻对应的补光参数作为特定的补光参数。In the process of collecting face images, the supplementary light parameters are randomly changed at regular intervals, and the supplementary light parameters corresponding to the acquisition time are recorded as specific supplementary light parameters.
前述的RGB色彩参数为α,亮度参数为β,通过以下公式计算所述RGB色彩参数:The aforementioned RGB color parameter is α, and the brightness parameter is β, and the RGB color parameter is calculated by the following formula:
α=(R/255,G/255,B/255),α = (R/255, G/255, B/255),
其中,R、G、B分别为补光的R、G、B值,β∈[0,1]。Among them, R, G, and B are the R, G, and B values of the fill light, respectively, and β∈[0,1].
本申请的卷积神经网络通过如下模块训练得到:The convolutional neural network of this application is obtained through the following module training:
样本获取模块,用于获取多个补光参数下采集的人脸图像,对所述人脸图像进行预处理,并对所述人脸图像添加标签,得到训练样本集。The sample acquisition module is used to acquire face images collected under multiple light supplement parameters, perform preprocessing on the face images, and add labels to the face images to obtain a training sample set.
其中,标签包括补光参数的RGB色彩参数α和亮度参数β。Wherein, the label includes an RGB color parameter α and a brightness parameter β of the fill light parameter.
前向处理模块,用于将所述训练样本集输入卷积神经网络,提取特征向量样本,所述特征向量样本用于判别真实人脸和假体人脸,并根据所述训练样本集回归补光参数。The forward processing module is used to input the training sample set into the convolutional neural network, extract feature vector samples, and the feature vector samples are used to distinguish real faces and fake faces, and return and complement the training sample set according to the training sample set. light parameters.
反向传播模块,用于通过ArcFace Loss计算特征向量样本的损失,通过欧式loss函数计算补光参数回归的损失,并通过反向传播更新卷积神经网络的参数。The backpropagation module is used to calculate the loss of feature vector samples through ArcFace Loss, calculate the loss of fill light parameter regression through the Euclidean loss function, and update the parameters of the convolutional neural network through backpropagation.
其中,在计算补光参数回归的损失时,假体训练样本的回归量包括RGB色彩参数和亮度参数,其中,α=(0,0,0),β=0。Wherein, when calculating the regression loss of the fill light parameter, the regressor of the prosthetic training sample includes RGB color parameters and brightness parameters, where α=(0,0,0), β=0.
在一个示例中,所述图像获取模块包括:In one example, the image acquisition module includes:
人脸检测和定位单元,用于对人脸图像进行人脸检测和人脸关键点定位,得到双眼坐标。The face detection and positioning unit is used to perform face detection and face key point positioning on the face image to obtain binocular coordinates.
人脸归一化单元,用于通过双眼坐标对人脸图像进行对齐和缩放。The human face normalization unit is used for aligning and scaling human face images through binocular coordinates.
本申请可以用于移动终端,通过移动终端的屏幕进行补光。This application can be used in mobile terminals, and supplementary light can be performed through the screen of the mobile terminal.
在一个示例中,所述装置还包括:In one example, the device also includes:
计算模块,用于计算所述回归后的补光参数与所述特定的补光参数的欧氏距离,得到所述回归后的补光参数与所述特定的补光参数的差距。The calculation module is used to calculate the Euclidean distance between the regressed fill light parameter and the specific fill light parameter, and obtain the difference between the regressed fill light parameter and the specific fill light parameter.
在一个示例中,所述处理模块具体用于:In one example, the processing module is specifically used for:
将所述特征向量输入至ArcFace Loss,得到所述活体分值。Input the feature vector into ArcFace Loss to get the living body score.
在一个示例中,该判断模块,具体用于所述当所述活体分值大于设 定的阈值,且所述回归后的补光参数与所述特定的补光参数的差距在预设范围内时,判定所述人脸图像来自真实人脸,否则,判定所述人脸图像来自假体人脸,包括:In an example, the judging module is specifically used for when the living body score is greater than a set threshold, and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range , it is determined that the face image is from a real face, otherwise, it is determined that the face image is from a fake face, including:
在所述活体分值大于或者等于设定的阈值,计算所述回归后的补光参数与所述特定的补光参数的差距;When the living body score is greater than or equal to a set threshold, calculate the difference between the regressed fill light parameter and the specific fill light parameter;
在所述差距在预设范围内的情况下,确定所述人脸图像来自真实人脸,When the difference is within a preset range, it is determined that the face image is from a real face,
在所述差距不在预设范围内的情况下,确定所述人脸图像来自假体人脸;When the difference is not within the preset range, it is determined that the face image is from a prosthetic face;
在所述活体分值小于所述设定的阈值的情况下,确定所述人脸图像来自假体人脸。In a case where the living body score is smaller than the set threshold, it is determined that the face image is from a prosthetic face.
本申请实施例所提供的装置,其实现原理及产生的技术效果和前述上述实施例所提供的方法相同,为简要描述,该装置实施例部分未提及之处,可参考前述上述实施例所提供的方法中相应内容。所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,前述描述的装置和单元的具体工作过程,均可以参考上述上述实施例所提供的方法中的对应过程,在此不再赘述。The implementation principle and technical effects of the device provided by the embodiment of the present application are the same as those of the method provided by the above-mentioned embodiment. The corresponding content in the method provided. Those skilled in the art can clearly understand that for the convenience and brevity of the description, the specific working process of the devices and units described above can refer to the corresponding process in the method provided by the above-mentioned embodiments, and will not be repeated here. .
上述人脸静默活体检测装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。其中,网络接口可以是以太网卡或无线网卡等。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。Each module in the above-mentioned face silent living body detection device can be fully or partially realized by software, hardware and combinations thereof. Wherein, the network interface may be an Ethernet card or a wireless network card or the like. The above-mentioned modules can be embedded in or independent of the processor in the computer device in the form of hardware, and can also be stored in the memory of the computer device in the form of software, so that the processor can invoke and execute the corresponding operations of the above-mentioned modules.
如在本申请中所使用的,术语“组件”、“模块”和“系统”等旨在表示计算机相关的实体,它可以是硬件、硬件和软件的组合、软件、或者执行中的软件。例如,组件可以但不限于是,在处理器上运行的进程、处理器、对象、可执行码、执行的线程、程序和/或计算机。作为说明,运行在服务器上的应用程序和服务器都可以是组件。一个或多个组件可以驻留在进程和/或执行的线程中,并且组件可以位于一个计算机内和/或 分布在两个或更多的计算机之间。As used in this application, the terms "component," "module," and "system" and the like are intended to mean a computer-related entity, which may be hardware, a combination of hardware and software, software, or software in execution. For example, a component may be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer. As an illustration, both an application running on a server and a server can be a component. One or more components can reside within a process and/or thread of execution, and a component can be localized on one computer and/or distributed between two or more computers.
在一个实施例中,本申请提供的上述实施例所述的方法可以通过计算机程序实现业务逻辑并记录在存储介质上,所述的存储介质可以计算机读取并执行,实现本说明书实施例所述的方法所描述方案的效果。因此,本申请还提供用于人脸静默活体检测的计算机可读存储介质,包括用于存储处理器可执行指令的存储器,指令被处理器执行时实现包括上述实施例的人脸静默活体检测方法的步骤。In one embodiment, the methods described in the above embodiments provided by this application can implement business logic through computer programs and record them on a storage medium, and the storage medium can be read and executed by a computer to implement the business logic described in the embodiments of this specification. The method describes the effect of the program. Therefore, the present application also provides a computer-readable storage medium for face silent liveness detection, including a memory for storing processor-executable instructions, and when the instructions are executed by the processor, the face silent liveness detection method comprising the above-mentioned embodiments is realized A step of.
本申请通过RGB补光,凸出真实人脸与假体人脸的材质和3D信息差异,有利于区分真假人脸,并有效缓解了环境光的影响;通过人脸图像补光参数进行回归,并将回归后的补光参数与特定的补光参数进行校验,解决了对高清翻拍假体人脸的误判。This application highlights the difference in material and 3D information between the real face and the prosthetic face through RGB fill light, which is beneficial to distinguish between real and fake faces, and effectively alleviates the impact of ambient light; regression is performed through face image fill light parameters , and verify the regressed fill light parameters with specific fill light parameters, which solves the misjudgment of the high-definition remake of the prosthetic face.
所述存储介质可以包括用于存储信息的物理装置,通常是将信息数字化后再以利用电、磁或者光学等方式的媒体加以存储。所述存储介质有可以包括:利用电能方式存储信息的装置如,各式存储器,如RAM、ROM等;利用磁能方式存储信息的装置如,硬盘、软盘、磁带、磁芯存储器、磁泡存储器、U盘;利用光学方式存储信息的装置如,CD或DVD。当然,还有其他方式的可读存储介质,例如量子存储器、石墨烯存储器等等。The storage medium may include a physical device for storing information, and information is usually digitized and then stored using an electrical, magnetic, or optical medium. Described storage medium can include: the device that utilizes electric energy mode to store information such as, various memory, as RAM, ROM etc.; USB stick; a device that stores information optically, such as a CD or DVD. Of course, there are other readable storage media, such as quantum memory, graphene memory and so on.
上述所述的存储介质根据上述实施例所提供的方法的描述还可以包括其他的实施方式,本实施例的实现原理及产生的技术效果和前述上述实施例所提供的方法相同,具体可以参照相关上述实施例所提供的方法的描述,在此不作一一赘述。The description of the above-mentioned storage medium according to the method provided by the above-mentioned embodiment may also include other implementation modes. The implementation principle and technical effect of this embodiment are the same as the method provided by the above-mentioned above-mentioned embodiment. For details, please refer to the relevant The descriptions of the methods provided in the foregoing embodiments are not repeated here.
在一个实施例中,本申请还提供一种用于人脸静默活体检测的设备,所述的设备可以为单独的计算机,也可以包括使用了本说明书的一个或多个所述方法或一个或多个实施例装置的实际操作装置等。所述人脸静默活体检测的设备可以包括至少一个处理器以及存储计算机可执行指令的存储器,处理器执行所述指令时实现上述任意一个或者多个实施例中所述人脸静默活体检测方法的步骤。In one embodiment, the present application also provides a device for face silent liveness detection, which may be a separate computer, or may include one or more of the methods or one or more of the methods described in this specification. The actual operating device of multiple embodiment devices and the like. The device for face silent life detection may include at least one processor and a memory storing computer-executable instructions. When the processor executes the instructions, the face silent life detection method described in any one or more embodiments above is implemented. step.
本申请通过RGB补光,凸出真实人脸与假体人脸的材质和3D信息差异,有利于区分真假人脸,并有效缓解了环境光的影响;通过人脸图像补光参数进行回归,并将回归后的补光参数与特定的补光参数进行校验,解决了对高清翻拍假体人脸的误判。This application highlights the difference in material and 3D information between the real face and the prosthetic face through RGB fill light, which is beneficial to distinguish between real and fake faces, and effectively alleviates the impact of ambient light; regression is performed through face image fill light parameters , and verify the regressed fill light parameters with specific fill light parameters, which solves the misjudgment of the high-definition remake of the prosthetic face.
上述所述的设备根据上述实施例所提供的方法的描述还可以包括其他的实施方式,本实施例的实现原理及产生的技术效果和前述上述实施例所提供的方法相同,具体可以参照相关上述实施例所提供的方法的描述,在此不作一一赘述。The description of the above-mentioned device according to the method provided by the above-mentioned embodiment may also include other implementations. The implementation principle and technical effect of this embodiment are the same as the method provided by the above-mentioned above-mentioned embodiment. For details, please refer to the relevant above-mentioned The descriptions of the methods provided in the embodiments are not repeated here.
最后应说明的是:以上所述实施例,仅为本申请的具体实施方式,用以说明本申请的技术方案,而非对其限制,本申请的保护范围并不局限于此,尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员理解:任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,其依然可以对前述实施例所记载的技术方案进行修改或可轻易想到变化,或者对其中部分技术特征进行等同替换;而这些修改、变化或者替换,并不使相应技术方案的本质脱离本申请实施例技术方案的精神和范围。都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以所述权利要求的保护范围为准。Finally, it should be noted that: the above-described embodiments are only specific implementations of the application, used to illustrate the technical solutions of the application, rather than limiting it, and the scope of protection of the application is not limited thereto, although referring to the aforementioned The embodiment has described this application in detail, and those of ordinary skill in the art understand that any person familiar with the art within the technical scope disclosed in this application can still modify or modify the technical solutions described in the foregoing embodiments. Changes can be easily thought of, or equivalent replacements can be made to some of the technical features; and these modifications, changes or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of the embodiments of the present application. All should be covered within the scope of protection of this application. Therefore, the protection scope of the present application should be determined by the protection scope of the claims.

Claims (15)

  1. 一种人脸静默活体检测方法,其特征在于,所述方法包括:A method for face silent living body detection, characterized in that the method comprises:
    获取在特定的补光参数下采集的人脸图像,对所述人脸图像进行预处理,得到预处理后的人脸图像;其中,所述补光参数包括RGB色彩参数和亮度参数;Obtaining a face image collected under specific supplementary light parameters, performing preprocessing on the human face image to obtain a preprocessed human face image; wherein, the supplementary light parameters include RGB color parameters and brightness parameters;
    将所述预处理后的人脸图像输入经过训练的卷积神经网络,提取特征向量,所述特征向量用于判别真实人脸和假体人脸,根据所述特征向量得到活体分值;并根据所述预处理后的人脸图像回归补光参数,得到回归后的补光参数;The preprocessed face image is input into a trained convolutional neural network to extract feature vectors, which are used to distinguish real faces and fake faces, and obtain living body scores according to the feature vectors; and According to the face image after the pretreatment, the light supplement parameter is regressed to obtain the light supplement parameter after regression;
    当所述活体分值大于设定的阈值,且所述回归后的补光参数与所述特定的补光参数的差距在预设范围内时,判定所述人脸图像来自真实人脸,否则,判定所述人脸图像来自假体人脸。When the living body score is greater than the set threshold, and the difference between the regressed fill light parameter and the specific fill light parameter is within a preset range, it is determined that the face image is from a real face, otherwise , it is determined that the face image is from a prosthetic face.
  2. 根据权利要求1所述的人脸静默活体检测方法,其特征在于,所述获取在特定的补光参数下采集的人脸图像,包括:The human face silent living body detection method according to claim 1, wherein said obtaining the human face image collected under specific supplementary light parameters comprises:
    在采集人脸图像的过程中,每间隔一定时间随机变换补光参数,并记录采集时刻对应的补光参数作为所述特定的补光参数。During the process of collecting the face image, the supplementary light parameter is randomly changed every certain time interval, and the supplementary light parameter corresponding to the acquisition time is recorded as the specific supplementary light parameter.
  3. 根据权利要求1或2所述的人脸静默活体检测方法,其特征在于,所述RGB色彩参数为α,亮度参数为β,通过以下公式计算所述RGB色彩参数:According to claim 1 or 2 described human face silent living body detection method, it is characterized in that, described RGB color parameter is α, and brightness parameter is β, calculates described RGB color parameter by following formula:
    α=(R/255,G/255,B/255),α = (R/255, G/255, B/255),
    其中,R、G、B分别为补光的R、G、B值,β∈[0,1]。Among them, R, G, and B are the R, G, and B values of the fill light, respectively, and β∈[0,1].
  4. 根据权利要求1-3任一项所述的人脸静默活体检测方法,其特征在于,所述卷积神经网络通过如下方法训练得到:According to the human face silent living detection method described in any one of claims 1-3, it is characterized in that, the convolutional neural network is obtained by training as follows:
    获取多个补光参数下采集的多张人脸图像,分别对各人脸图像进行预处理,并对各人脸图像添加标签,得到训练样本集;其中,所述标签包括补光参数中的RGB色彩参数α和亮度参数β;Obtain a plurality of face images collected under a plurality of fill light parameters, preprocess each face image respectively, and add labels to each face image to obtain a training sample set; wherein, the labels include RGB color parameter α and brightness parameter β;
    将所述训练样本集输入卷积神经网络,提取特征向量样本,所述特 征向量样本用于判别真实人脸和假体人脸,并根据所述训练样本集回归补光参数;The training sample set is input into the convolutional neural network, and feature vector samples are extracted, and the feature vector samples are used to distinguish real faces and artificial faces, and return light supplement parameters according to the training sample set;
    通过ArcFace Loss计算所述特征向量样本的损失,通过欧式loss函数计算补光参数回归的损失,并通过反向传播更新所述卷积神经网络的参数;其中,在计算补光参数回归的损失时,假体训练样本的回归量包括RGB色彩参数和亮度参数,其中,α=(0,0,0),β=0。Calculate the loss of the eigenvector sample by ArcFace Loss, calculate the loss of the fill light parameter regression through the Euclidean loss function, and update the parameters of the convolutional neural network through backpropagation; wherein, when calculating the loss of the fill light parameter regression , the regressors of the prosthetic training samples include RGB color parameters and brightness parameters, where α=(0,0,0), β=0.
  5. 根据权利要求4所述的人脸静默活体检测方法,其特征在于,所述对所述人脸图像进行预处理,包括:The human face silent living body detection method according to claim 4, wherein said preprocessing the human face image includes:
    对所述人脸图像进行人脸检测和人脸关键点定位,得到双眼坐标;Carry out face detection and face key point location to described face image, obtain binocular coordinates;
    通过双眼坐标对人脸图像进行对齐和缩放,得到预处理后的人脸图像。The face image is aligned and scaled by binocular coordinates to obtain the preprocessed face image.
  6. 根据权利要求1-5任一项所述的人脸静默活体检测方法,其特征在于,通过移动终端的屏幕进行补光。The face silent living body detection method according to any one of claims 1-5, characterized in that the supplementary light is performed through the screen of the mobile terminal.
  7. 根据权利要求1所述的人脸静默活体检测方法,其特征在于,所述方法还包括:The face silent living body detection method according to claim 1, is characterized in that, described method also comprises:
    计算所述回归后的补光参数与所述特定的补光参数的欧氏距离,得到所述回归后的补光参数与所述特定的补光参数的差距。Calculate the Euclidean distance between the regressed fill light parameter and the specific fill light parameter, and obtain the difference between the regressed fill light parameter and the specific fill light parameter.
  8. 根据权利要求1所述的人脸静默活体检测方法,其特征在于,所述根据所述特征向量得到活体分值,包括:The face silent living body detection method according to claim 1, wherein said obtaining the living body score according to said feature vector comprises:
    将所述特征向量输入至ArcFace Loss,得到所述活体分值。Input the feature vector into ArcFace Loss to get the living body score.
  9. 根据权利要求8所述的人脸静默活体检测方法,其特征在于,所述人脸关键点还包括鼻子关键点、左嘴角关键点息以及右嘴角关键点。The human face silent living body detection method according to claim 8, wherein the key points of the human face also include key points of the nose, key points of the left corner of the mouth and key points of the right corner of the mouth.
  10. 根据权利要求1所述的人脸静默活体检测方法,其特征在于,所述当所述活体分值大于设定的阈值,且所述回归后的补光参数与所述特定的补光参数的差距在预设范围内时,判定所述人脸图像来自真实人脸,否则,判定所述人脸图像来自假体人脸,包括:The face silent living body detection method according to claim 1, wherein when the living body score is greater than a set threshold, and the regressed supplementary light parameter and the specific supplementary light parameter When the difference is within the preset range, it is determined that the human face image is from a real human face, otherwise, it is determined that the human face image is from a prosthetic human face, including:
    在所述活体分值大于或者等于设定的阈值,计算所述回归后的补光参数与所述特定的补光参数的差距;When the living body score is greater than or equal to a set threshold, calculate the difference between the regressed fill light parameter and the specific fill light parameter;
    在所述差距在预设范围内的情况下,确定所述人脸图像来自真实人脸,When the difference is within a preset range, it is determined that the face image is from a real face,
    在所述差距不在预设范围内的情况下,确定所述人脸图像来自假体人脸;When the difference is not within the preset range, it is determined that the face image is from a prosthetic face;
    在所述活体分值小于所述设定的阈值的情况下,确定所述人脸图像来自假体人脸。In a case where the living body score is smaller than the set threshold, it is determined that the face image is from a prosthetic face.
  11. 一种人脸静默活体检测装置,其特征在于,所述装置包括:A human face silent living body detection device is characterized in that said device comprises:
    图像获取模块,用于获取在特定的补光参数下采集的人脸图像,对所述人脸图像进行预处理,得到预处理后的人脸图像;其中,所述补光参数包括RGB色彩参数和亮度参数;An image acquisition module, configured to acquire a face image collected under specific light supplement parameters, and preprocess the face image to obtain a preprocessed face image; wherein, the light supplement parameters include RGB color parameters and brightness parameters;
    处理模块,用于将所述预处理后的人脸图像输入经过训练的卷积神经网络,提取特征向量,所述特征向量用于判别真实人脸和假体人脸,根据所述特征向量得到活体分值;并根据所述预处理后的人脸图像回归补光参数,得到回归后的补光参数;A processing module, configured to input the preprocessed face image into a trained convolutional neural network to extract a feature vector, which is used to distinguish a real face from a prosthetic face, obtained according to the feature vector In vivo score; And according to the face image after the preprocessing regression supplementary light parameter, obtain the supplementary light parameter after regression;
    判断模块,用于当所述活体分值大于设定的阈值,且所述回归后的补光参数与所述特定的补光参数的差距在预设范围内时,判定所述人脸图像来自真实人脸,否则,判定所述人脸图像来自假体人脸。A judging module, configured to determine that the face image is from If it is a real face, otherwise, it is determined that the face image is from a prosthetic face.
  12. 根据权利要求11所述的人脸静默活体检测装置,其特征在于,所述装置还包括:The human face silent living body detection device according to claim 11, wherein the device further comprises:
    样本获取模块,用于获取多个补光参数下采集的人脸图像,进行预处理并添加标签,得到训练样本集;其中,所述标签包括补光参数的RGB色彩参数α和亮度参数β;The sample acquisition module is used to obtain the face images collected under a plurality of supplementary light parameters, preprocess and add labels to obtain a training sample set; wherein, the labels include RGB color parameters α and brightness parameters β of supplementary light parameters;
    前向处理模块,用于将所述训练样本集输入卷积神经网络,提取用于判别真实人脸和假体人脸的特征向量样本,并根据训练样本集回归补光参数;The forward processing module is used for inputting the training sample set into the convolutional neural network, extracting feature vector samples for discriminating real faces and prosthetic faces, and regressing fill light parameters according to the training sample set;
    反向传播模块,用于通过ArcFace Loss计算特征向量样本的损失,通过欧式loss函数计算补光参数回归的损失,并通过反向传播更新卷积神经网络的参数;其中,在计算补光参数回归的损失时,假体训练样本的 回归量包括RGB色彩参数和亮度参数,其中,α=(0,0,0),β=0。The backpropagation module is used to calculate the loss of the eigenvector samples through ArcFace Loss, calculate the loss of the fill light parameter regression through the Euclidean loss function, and update the parameters of the convolutional neural network through backpropagation; wherein, when calculating the fill light parameter regression When the loss of , the regressor of the prosthetic training sample includes RGB color parameters and brightness parameters, where α=(0,0,0), β=0.
  13. 根据权利要求8-11任一项所述的人脸静默活体检测装置,其特征在于,所述图像获取模块包括:The human face silent living body detection device according to any one of claims 8-11, wherein the image acquisition module comprises:
    人脸检测和定位单元,用于对所述人脸图像进行人脸检测和人脸关键点定位,得到双眼坐标;A face detection and positioning unit, configured to perform face detection and key point positioning on the face image to obtain binocular coordinates;
    人脸归一化单元,用于通过双眼坐标对人脸图像进行对齐和缩放,得到预处理后的人脸图像。The human face normalization unit is used to align and scale the human face image through binocular coordinates to obtain the preprocessed human face image.
  14. 一种用于人脸静默活体检测的计算机可读存储介质,其特征在于,包括用于存储处理器可执行指令的存储器,所述指令被所述处理器执行时实现包括权利要求1-7任一所述人脸静默活体检测方法的步骤。A computer-readable storage medium for face silent liveness detection, characterized in that it includes a memory for storing processor-executable instructions, and when the instructions are executed by the processor, it implements any of claims 1-7. A step of the face silent liveness detection method.
  15. 一种用于人脸静默活体检测的设备,其特征在于,包括至少一个处理器以及存储计算机可执行指令的存储器,所述处理器执行所述指令时实现权利要求1-7中任意一项所述人脸静默活体检测方法的步骤。A device for face silent liveness detection, characterized in that it includes at least one processor and a memory storing computer-executable instructions, and when the processor executes the instructions, it implements any one of claims 1-7. Describe the steps of the human face silent liveness detection method.
PCT/CN2022/118048 2021-10-15 2022-09-09 Facial silent living body detection method and apparatus, and storage medium and device WO2023061123A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202111201839.8 2021-10-15
CN202111201839.8A CN115995102A (en) 2021-10-15 2021-10-15 Face silence living body detection method, device, storage medium and equipment

Publications (1)

Publication Number Publication Date
WO2023061123A1 true WO2023061123A1 (en) 2023-04-20

Family

ID=85987262

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/118048 WO2023061123A1 (en) 2021-10-15 2022-09-09 Facial silent living body detection method and apparatus, and storage medium and device

Country Status (2)

Country Link
CN (1) CN115995102A (en)
WO (1) WO2023061123A1 (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107992794A (en) * 2016-12-30 2018-05-04 腾讯科技(深圳)有限公司 A kind of biopsy method, device and storage medium
US20180173980A1 (en) * 2016-12-15 2018-06-21 Beijing Kuangshi Technology Co., Ltd. Method and device for face liveness detection
CN109558840A (en) * 2018-11-29 2019-04-02 中国科学院重庆绿色智能技术研究院 A kind of biopsy method of Fusion Features
CN110516644A (en) * 2019-08-30 2019-11-29 深圳前海微众银行股份有限公司 A kind of biopsy method and device
CN110969077A (en) * 2019-09-16 2020-04-07 成都恒道智融信息技术有限公司 Living body detection method based on color change
CN111241919A (en) * 2019-12-27 2020-06-05 深圳市宁远电子科技有限公司 Face living body detection method and system combining optical verification code
CN112906610A (en) * 2021-03-05 2021-06-04 上海肇观电子科技有限公司 Method for living body detection, electronic circuit, electronic apparatus, and medium
CN113361349A (en) * 2021-05-25 2021-09-07 北京百度网讯科技有限公司 Face living body detection method and device, electronic equipment and storage medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180173980A1 (en) * 2016-12-15 2018-06-21 Beijing Kuangshi Technology Co., Ltd. Method and device for face liveness detection
CN107992794A (en) * 2016-12-30 2018-05-04 腾讯科技(深圳)有限公司 A kind of biopsy method, device and storage medium
CN109558840A (en) * 2018-11-29 2019-04-02 中国科学院重庆绿色智能技术研究院 A kind of biopsy method of Fusion Features
CN110516644A (en) * 2019-08-30 2019-11-29 深圳前海微众银行股份有限公司 A kind of biopsy method and device
CN110969077A (en) * 2019-09-16 2020-04-07 成都恒道智融信息技术有限公司 Living body detection method based on color change
CN111241919A (en) * 2019-12-27 2020-06-05 深圳市宁远电子科技有限公司 Face living body detection method and system combining optical verification code
CN112906610A (en) * 2021-03-05 2021-06-04 上海肇观电子科技有限公司 Method for living body detection, electronic circuit, electronic apparatus, and medium
CN113361349A (en) * 2021-05-25 2021-09-07 北京百度网讯科技有限公司 Face living body detection method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN115995102A (en) 2023-04-21

Similar Documents

Publication Publication Date Title
Ge et al. Occluded face recognition in the wild by identity-diversity inpainting
Bak et al. Spatio-temporal saliency networks for dynamic saliency prediction
Wang et al. Robust face anti-spoofing with depth information
US9818024B2 (en) Identifying facial expressions in acquired digital images
Song et al. A robust eye detection method using combined binary edge and intensity information
CN108921041A (en) A kind of biopsy method and device based on RGB and IR binocular camera
Hatem et al. A survey of feature base methods for human face detection
WO2020140723A1 (en) Method, apparatus and device for detecting dynamic facial expression, and storage medium
Zhang et al. A survey on face anti-spoofing algorithms
CN111860394A (en) Gesture estimation and gesture detection-based action living body recognition method
WO2023279557A1 (en) Fake video inspection method and system based on blink synchronization and binocular movement detection
Huang et al. Attendance system based on dynamic face recognition
Selvaraj et al. Raspberry Pi based automatic door control system
CN112329727A (en) Living body detection method and device
Shrivastava et al. Conceptual model for proficient automated attendance system based on face recognition and gender classification using Haar-Cascade, LBPH algorithm along with LDA model
CN107862298B (en) Winking living body detection method based on infrared camera device
Li et al. Exploiting facial symmetry to expose deepfakes
Nikitin et al. Face anti-spoofing with joint spoofing medium detection and eye blinking analysis
WO2023061123A1 (en) Facial silent living body detection method and apparatus, and storage medium and device
WO2023061122A1 (en) Live face detection method and apparatus, computer-readable storage medium, and device
Sharma et al. Study and implementation of face detection algorithm using Matlab
Kainz et al. Students’ Attendance Monitoring through the Face Recognition
Gowda et al. Facial Expression Analysis and Estimation Based on Facial Salient Points and Action Unit (AUs)
WO2024045350A1 (en) Eye movement based liveness detection method and system based on deep learning
Shen et al. Unsupervised face anti-spoofing using dual cameras based feature matching