CN111222433B - Automatic face auditing method, system, equipment and readable storage medium - Google Patents

Automatic face auditing method, system, equipment and readable storage medium Download PDF

Info

Publication number
CN111222433B
CN111222433B CN201911387766.9A CN201911387766A CN111222433B CN 111222433 B CN111222433 B CN 111222433B CN 201911387766 A CN201911387766 A CN 201911387766A CN 111222433 B CN111222433 B CN 111222433B
Authority
CN
China
Prior art keywords
face
pictures
picture
quality
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911387766.9A
Other languages
Chinese (zh)
Other versions
CN111222433A (en
Inventor
刘小扬
王心莹
王欢
徐小丹
黄泽斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Newland Digital Technology Co ltd
Original Assignee
Newland Digital Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Newland Digital Technology Co ltd filed Critical Newland Digital Technology Co ltd
Priority to CN201911387766.9A priority Critical patent/CN111222433B/en
Publication of CN111222433A publication Critical patent/CN111222433A/en
Application granted granted Critical
Publication of CN111222433B publication Critical patent/CN111222433B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/26Government or public services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/40Spoof detection, e.g. liveness detection
    • G06V40/45Detection of the body part being alive

Abstract

The invention discloses an automatic face auditing method, which comprises the following steps: face detection, namely detecting coordinates of a rectangular frame of the face and coordinates of key points of the face through a cascade neural network algorithm; face quality evaluation and screening, namely performing face quality evaluation on quality attributes of a plurality of face pictures, and screening high-quality pictures; detecting living bodies, namely detecting whether the pictures are true persons or not by using a double-flow convolutional neural network, and filtering and judging the pictures to be non-true persons; and (3) face comparison authentication, namely extracting face feature vectors of the pictures, comparing the similarity degree between the face feature vectors of the pictures and the face feature vectors of the standard pictures, and outputting a comparison result. The invention realizes full automation of picture auditing, does not need manual operation, reduces labor cost, filters pictures with low similarity degree, and reduces manual errors. In addition, each invention uses a cascading picture filtering mode, so that the speed is higher.

Description

Automatic face auditing method, system, equipment and readable storage medium
Technical Field
The present invention relates to the field of image recognition processing technologies, and in particular, to a method, a system, an apparatus, and a readable storage medium for automatic face auditing.
Background
With the rapid development of mobile devices and mobile internet, many government platforms and enterprise platforms are increasingly beginning to merge offline and online services for the convenience and convenience of mass offices. For example, governments in various places have accelerated the construction of on-line government service platforms, mobile operators have opened on-line business halls, and so on. Many online services need to upload photos and keep a bottom according to requirements, for example, mobile phone operators need to upload personal photos for online handling of mobile phone numbers, front photos and back photos of identity cards, and independent application of shops for services such as kittens, treasures and the like need to upload personal handheld photos. In the prior art, manual verification and judgment are mainly adopted for personal photos uploaded by users. However, depending on manual operation to examine whether the personal photo uploaded by the user is compliant, whether the personal photo is compliant, etc., the working efficiency is low, the labor cost is high, human errors are inevitably generated in the process, and feedback to the user cannot be timely obtained.
Disclosure of Invention
The invention aims to provide a quick and efficient automatic face auditing method, system and device and a readable storage medium.
In order to solve the technical problems, the technical scheme of the invention is as follows:
in a first aspect, the present invention provides an automatic face auditing method, including the steps of:
face detection, namely detecting coordinates of a rectangular frame of the face and coordinates of key points of the face through a cascade neural network algorithm;
face quality evaluation and screening, namely performing face quality evaluation on quality attributes of a plurality of face pictures, and screening high-quality pictures;
detecting living bodies, namely detecting whether the pictures are true persons or not by using a double-flow convolutional neural network, and filtering and judging the pictures to be non-true persons;
and (3) face comparison authentication, namely extracting face feature vectors of the pictures, comparing the similarity degree between the face feature vectors of the pictures and the face feature vectors of the standard pictures, and filtering the pictures with low similarity degree.
Preferably, before face detection, a picture with a filtering resolution outside the threshold range is further included.
Preferably, after the face detection, the method further comprises:
filtering out the pictures with the size of the face rectangular frame accounting for the size of the picture and lower than a threshold value;
and filtering out pictures with the distance between the two eyes of the human face lower than a threshold value.
Preferably, after the face detection, the method further comprises:
and aligning the human face, calculating a transformation matrix between the key point coordinates of the human face of a picture and the key point coordinates of a pre-stored standard human face, and acting the transformation matrix on the picture to obtain an aligned human face image.
Preferably, the face comparison authentication process is as follows:
extracting the high-quality picture, outputting a floating point vector with 512 dimensions by using a 50-layer ResNet neural network, and recording the floating point vector as a face feature vector;
by comparing the similarity degree between the face feature vector of the current picture and the face feature vector of the standard picture, the formula is as follows:
Figure BDA0002344033060000021
wherein S is i Is the face feature vector of the current picture, S j The face feature vector is the face feature vector of the standard picture;
if the similarity is lower than a threshold, judging that the person evidence is not uniform; and if the similarity is higher than a threshold value, judging that the person and the evidence are uniform.
Preferably, the quality attributes used for face quality assessment include: the quality attributes used for face quality assessment comprise face gestures, eye states, mouth states, makeup states, overall brightness, left and right face brightness differences, ambiguity and shielding;
the face pose, the eye state, the mouth state, the dressing state, the ambiguity and the occlusion all adopt a MobileFaceNet structure as a main body to construct a multi-task convolutional neural network, and a plurality of tasks output corresponding to the quality attributes of the face respectively.
Eye state, mouth state, makeup state and face shielding are used as classification tasks, and a softmax loss function is used as an objective function;
the face gesture, the illuminance of the image and the image ambiguity are used as regression tasks, and a Euclidean loss function is used as an objective function;
the total objective function of the network training comprises a combination of a plurality of Softmax loss functions and a Euclidean loss function, and when a plurality of tasks are learned together, the total objective function is a linear combination of a plurality of loss functions.
Preferably, the in vivo detection process is:
acquiring a depth image, and carrying out normalization processing on a face region in the image to obtain a processed face depth image;
inputting RGB face images with a face ID of a preset frame number and the face depth map into a deep learning network for detection, and obtaining a living body judgment result of each frame of picture;
voting all living body judgment results of the face ID, when the number of frames of the living body is judged to be large, determining the object as a living body, and when the number of frames of the attack is judged to be large, determining the object as a non-living body.
Preferably, the in vivo detection process is:
the face is intercepted from the original image, RGB channels are converted into HSV and YCbCr spaces, and the converted HSV and YCbCr images are overlapped to obtain an overlapped image; extracting Sobel features from the face region through a Sobel operator, and obtaining a Sobel feature map;
respectively inputting the Sobel characteristic diagram and the superposition diagram of a preset frame number of a face ID from two input channels of a double-flow neural network to obtain a living body judgment result of each frame of picture;
voting all living body judgment results of the face ID, when the number of frames of the living body is judged to be large, the object is judged to be living body, and when the number of frames of the attack is judged to be large, the object is judged to be non-living body.
In a second aspect, the present invention further provides an automatic face auditing system, including:
the human face detection module is used for detecting a human face rectangular frame and human face key points through a cascade neural network algorithm;
the face quality evaluation module is used for carrying out face quality evaluation through quality attributes of a plurality of face pictures and screening high-quality pictures;
the living body detection module is used for detecting whether the pictures are true persons or not by using a double-flow convolutional neural network and filtering pictures without the true persons;
and a face comparison module for extracting face feature vectors of the pictures, comparing the similarity between the face feature vectors of the pictures and the face feature vectors of the pre-stored credentials, and filtering the pictures with low similarity.
In a third aspect, the present invention further provides an automatic face auditing apparatus, including a memory, a processor, and a computer program stored in the memory and capable of running on the processor, where the processor implements the steps of the automatic face auditing method described above when executing the program.
In a fourth aspect, the present invention further proposes a readable storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the above-described automatic face auditing method.
The technical scheme of the invention is an automatic face auditing method combining face detection, face quality analysis, face living body detection and face recognition technologies. The method examines whether the quality of the personal photo uploaded by the user is compliant, whether the personal photo is a true person or not, and whether the personal photo is authenticated. In the technical scheme, whether the quality of the personal photo is in compliance mainly considers whether the face in the picture is high in quality and is convenient to identify; the living body detection mainly considers whether the photo is a flip or fake; whether the personal certificate is unified mainly considers whether the certificate photograph and the personal photo are the same person. The invention realizes full automation of picture auditing, does not need manual operation, reduces labor cost, has stable algorithm and reduces manual error. In addition, each invention uses a cascading picture filtering mode, so that the speed is higher.
Drawings
Fig. 1 is a flowchart illustrating steps of an embodiment of an automatic face auditing method according to the present invention.
Detailed Description
The following describes the embodiments of the present invention further with reference to the drawings. The description of these embodiments is provided to assist understanding of the present invention, but is not intended to limit the present invention. In addition, the technical features of the embodiments of the present invention described below may be combined with each other as long as they do not collide with each other.
Referring to fig. 1, the invention provides an automatic face auditing method, which comprises the following steps:
filtering pictures with resolution outside the threshold range, and taking pictures with resolution lower than the threshold in the vertical direction or lower than the threshold in the horizontal direction as pictures with resolution not meeting the requirements. In the embodiment of the invention, the photos with the resolution lower than 640 in the vertical direction or the resolution lower than 480 in the horizontal direction are filtered and deleted.
S10: and detecting the coordinates of the rectangular frame of the human face and the coordinates of key points of the human face through a cascade neural network algorithm.
And predicting the coordinates of the face frame and the coordinates of the key points of the face in the image by using a cascade neural network algorithm. The face frame coordinates refer to a rectangular face frame containing a face area; the face key point coordinates refer to positions of 106 key points of the face area, and cover eyebrows, glasses, nose, mouth and face contours of the face area.
And calculating the size of the face frame according to the face frame coordinates. When the face duty ratio is greater than the threshold, the face duty ratio does not meet the requirement. In this embodiment, when the face ratio is greater than 0.4, the proportion of the face to the whole image is considered to be too large.
Face duty ratio = face frame size/image size
And calculating the distance between pupils of two eyes according to the key points of the human face, namely the number of pixels between the centers of the two eyes. When the interocular distance is smaller than the threshold value, the interocular distance does not satisfy the requirement. For example, when the distance between the left eye and the right eye is less than 40, the inter-eye distance is too small.
Face alignment: and aiming at each face, calculating a transformation matrix between the extracted face key point coordinates and the standard face key point coordinates, and applying the transformation matrix to the initial face image to obtain an aligned face image. The distribution of the aligned face key point coordinates is more consistent, and the face is corrected.
S20: face quality evaluation and screening, namely performing face quality evaluation on quality attributes of a plurality of face pictures, and screening high-quality pictures.
The face quality evaluation algorithm adopts a mode of combining deep learning and a traditional image analysis algorithm, and achieves the quality attributes of face brightness, left-right face brightness difference, face angle (yaw) around the y-axis direction, face angle (pitch) around the x-axis direction, face angle (roll) around the z-axis direction, expression classification, glasses classification, mask classification, eye state classification, mouth state classification, makeup state classification, face reality (the distinguishing of the items is stone statue image, CG face and real face), face ambiguity, face shielding degree and the like according to the face characteristics of the face image.
The face brightness and the left and right face brightness difference adopt a traditional algorithm, specifically, three RGB channels of a face image are converted into gray images according to a certain proportion, each region of the face is divided according to 106 key points of the face, the face brightness is calculated according to the gray average value of the face region, and the left and right face brightness difference is calculated according to the gray average value of the left and right faces.
The other attributes are realized by adopting a deep learning method, a lightweight MobileFaceNet structure is adopted as a main body to construct a multi-task convolutional neural network, and a plurality of tasks output each quality attribute corresponding to a face. The quality judgment of eye state, mouth state, dressing state, face shielding, mask classification and the like belongs to classification tasks, and a softmax loss function is adopted as an objective function; face pose angles, image ambiguities and the like belong to regression tasks, and a Euclidean loss function is adopted as an objective function. The total objective function of the network training is a combination of a plurality of Softmax loss functions and a Euclidean loss function, and when a plurality of tasks are learned together, the total objective function is a linear combination of a plurality of loss functions.
Calculate Softmax loss L: l= -log (p i ),
Wherein p is i Normalized probabilities calculated for each attribute class, i.e
Figure BDA0002344033060000041
x i Representing the ith neuron output, N representing the total number of categories;
calculate the Euclidean loss L:
Figure BDA0002344033060000042
wherein y is i As a true tag value of the tag,
Figure BDA0002344033060000043
is the predicted value of the regressor.
After face quality evaluation, face quality screening is also required, and reference factors for screening include the following:
face ratio: and calculating the size of the face frame according to the face frame coordinates, wherein when the face ratio is larger than a threshold value, the face ratio does not meet the requirement. For example: when the face ratio is greater than 0.4, the proportion of the face to the whole image is too large.
Face brightness: the brightness of the human face should be within a reasonable range. For example: the human face brightness value is between 0 and 1, and the reasonable human face brightness is more than 0.3 and less than 0.8.
Left-right face brightness difference: the left-right face luminance difference should be less than the threshold. For example: when the difference of the left and right face brightness is between 0 and 1, the reasonable difference of the left and right face brightness is less than 0.4.
Face pose: the face angle (yaw) around the y-axis, the face angle (pitch) around the x-axis, and the face angle (roll) around the z-axis should be within a reasonable range. For example, within + -10 deg..
Ambiguity: the ambiguity should be less than a threshold. For example: when the ambiguity is between 0 and 1, the face ambiguity should be less than 0.6.
Shielding: if the face map is judged to have shielding of the five sense organs and the outlines, including a sunglasses or a mask, filtering is carried out.
Expression: if the face image is judged to be an exaggerated expression, the eyes are closed, and the mouth is enlarged, filtering is performed.
Degree of realism: the degree of realism should be greater than a threshold, if the degree of realism is small, this may indicate that the face is a statue face/cartoon face, etc. For example: when the value of the reality is between 0 and 1, the reality of the human face is larger than 0.6.
And filtering out pictures which do not meet the quality requirements according to the requirements.
S30: and detecting living bodies, namely detecting whether the pictures are true persons or not by using a double-flow convolutional neural network, and filtering and judging the pictures to be non-true persons.
The following two methods can be used in performing the biopsy:
the first living body detection process is as follows:
acquiring a depth image, and carrying out normalization processing on a face region in the image to obtain a processed face depth image;
and inputting the RGB face image of the picture and the face depth image into a deep learning network for detection, and obtaining a living body judgment result of the picture.
Specifically, a deep learning network for performing living body judgment on a picture uses Resnet as a basic network, the deep learning network adopts a double-input channel of a face image and a face depth image, after feature extraction is performed on two input branches respectively, the features extracted by the two branches are selectively excited and fused through a se-module, and then feature extraction is performed on the fused features through a plurality of layers of convolution, so that a living body judgment result is obtained.
Specifically, the objective function of the deep learning network is a focal loss function.
Specifically, the actual depths of all the eyes and mouth angles in the key points of the human face are calculated, the average value of the actual depths of the points is calculated, the upper limit of normalization is taken as the average value plus a fixed value, the lower limit is taken as the average value minus the fixed value, and the depth of the human face area is normalized into a gray scale map with the pixel value in the interval of 0-255.
The gray value for the actual depth greater than the upper limit and less than the lower limit is set to 0.
Wherein, the normalization formula is:
Figure BDA0002344033060000051
wherein V is a gray value after depth normalization, the range is 0-255, dreal is the actual depth of the face region, dmax is the upper limit of the actual depth of the face, and Dmin is the lower limit of the actual depth of the face.
The second living body detection process is as follows:
the face is intercepted from the original image, RGB channels are converted into HSV and YCbCr spaces, and the converted HSV and YCbCr images are overlapped to obtain an overlapped image; extracting Sobel features from the face region through a Sobel operator to obtain a Sobel feature map;
and respectively inputting the Sobel characteristic diagram and the superposition diagram of the picture from two input channels of the double-flow neural network to obtain a living body judgment result of the picture.
Specifically, for each input image a, gx and Gy are convolved with image a to obtain AGx, AGy, respectively, and then image AG is output, the value of each pixel is:
Figure BDA0002344033060000052
where Gx represents the convolution kernel in the x-direction and Gy represents the convolution kernel in the y-direction.
S40: and (3) face comparison authentication, namely extracting face feature vectors of the pictures, comparing the similarity degree between the face feature vectors of the pictures and the face feature vectors of the standard pictures, and outputting a comparison result.
Extracting a high-quality picture, outputting a floating point vector with 512 dimensions by using a 50-layer ResNet neural network, and recording the floating point vector as a face feature vector;
by comparing the similarity degree between the face feature vector of the current picture and the face feature vector of the standard picture, the formula is as follows:
Figure BDA0002344033060000061
wherein S is i Is the face feature vector of the current picture, S j The face feature vector is the face feature vector of the standard picture;
if the similarity is lower than the threshold, judging that the credentials are not uniform; if the similarity is higher than the threshold, judging that the people are unified
The technical scheme of the invention is a face auditing method combining face detection, face quality analysis, face living body detection and face recognition technology, and is used for checking whether the quality of the personal photo uploaded by the user is compliant, whether the personal photo is a true person or not and whether the personal photo is authenticated. Whether the quality of the personal photo is compliant mainly considers whether the face in the picture is high in quality and is convenient to identify; the living body detection mainly considers whether the photo is a flip or fake; whether the personal certificate is unified mainly considers whether the certificate photograph and the personal photo are the same person. The invention realizes full automation of picture auditing, does not need manual operation, reduces labor cost, has stable algorithm and reduces manual error. In addition, each invention uses a cascading picture filtering mode, so that the speed is higher.
On the other hand, the invention also provides an automatic face auditing system, which comprises the following steps:
the human face detection module is used for detecting a human face rectangular frame and human face key points through a cascade neural network algorithm;
the face quality evaluation module is used for carrying out face quality evaluation through quality attributes of a plurality of face pictures and screening high-quality pictures;
the living body detection module is used for detecting whether the pictures are true persons or not by using a double-flow convolutional neural network and filtering pictures without the true persons;
and a face comparison module for extracting face feature vectors of the pictures, comparing the similarity between the face feature vectors of the pictures and the face feature vectors of the pre-stored credentials, and filtering the pictures with low similarity.
In still another aspect, the present invention further provides an automatic face auditing apparatus, including a memory, a processor, and a computer program stored in the memory and capable of running on the processor, where the processor implements the steps of the automatic face auditing method described above when executing the program.
In yet another aspect, the present invention further provides a readable storage medium having stored thereon a computer program which, when executed by a processor, performs the steps of the above-described automatic face auditing method.
Aiming at websites and applications requiring uploading images to meet certain standard requirements, the invention provides an automatic face auditing method and system based on a deep neural network. The method can be effectively used in the information verification category, and realizes rapid face filtering and person evidence comparison.
The embodiments of the present invention have been described in detail above with reference to the accompanying drawings, but the present invention is not limited to the described embodiments. It will be apparent to those skilled in the art that various changes, modifications, substitutions and alterations can be made to these embodiments without departing from the principles and spirit of the invention, and yet fall within the scope of the invention.

Claims (9)

1. An automatic face auditing method is characterized by comprising the following steps:
face detection, namely detecting coordinates of a rectangular frame of the face and coordinates of key points of the face through a cascade neural network algorithm;
face quality evaluation and screening, namely performing face quality evaluation on quality attributes of a plurality of face pictures, and screening high-quality pictures;
detecting living bodies, namely detecting whether the pictures are true persons or not by using a double-flow convolutional neural network, and filtering and judging the pictures to be non-true persons;
face comparison and authentication, namely extracting a face feature vector of a picture, comparing the similarity degree between the face feature vector of the picture and the face feature vector of a standard picture, and filtering a picture with low similarity degree; before face detection, filtering pictures with resolution outside a threshold range;
the quality attributes used for face quality assessment include: the quality attributes used for face quality assessment comprise face gestures, eye states, mouth states, makeup states, overall brightness, left and right face brightness differences, ambiguity and shielding;
the face pose, eye state, mouth state, dressing state, ambiguity and occlusion all adopt a MobileFaceNet structure as a main body to construct a multi-task convolutional neural network, and a plurality of task outputs respectively correspond to each quality attribute of the face;
eye state, mouth state, makeup state and face shielding are used as classification tasks, and a softmax loss function is used as an objective function;
the face gesture, the illuminance of the image and the image ambiguity are used as regression tasks, and a Euclidean loss function is used as an objective function;
the total objective function of the network training comprises a combination of a plurality of Softmax loss functions and a Euclidean loss function, and when a plurality of tasks are learned together, the total objective function is a linear combination of a plurality of loss functions.
2. The automatic face auditing method of claim 1, characterized by: after face detection, further comprising:
filtering out the pictures with the size of the face rectangular frame accounting for the size of the picture and lower than a threshold value;
and filtering out pictures with the distance between the two eyes of the human face lower than a threshold value.
3. The automatic face auditing method according to claim 1, further comprising, after face detection:
and aligning the human face, calculating a transformation matrix between the coordinates of the key points of the human face of a picture and the coordinates of the key points of the pre-stored standard human face, and acting the transformation matrix on the picture to obtain an aligned human face image.
4. A method of automatic face auditing according to any of claims 1 to 3, in which the process of face comparison authentication is:
extracting the high-quality picture, outputting a floating point vector with 512 dimensions by using a 50-layer ResNet neural network, and recording the floating point vector as a face feature vector;
by comparing the similarity degree between the face feature vector of the current picture and the face feature vector of the standard picture, the formula is as follows:
Figure QLYQS_1
wherein S is i Is the face feature vector of the current picture, S j The face feature vector is the face feature vector of the standard picture;
if the similarity is lower than a threshold, judging that the person evidence is not uniform; and if the similarity is higher than a threshold value, judging that the person and the evidence are unified.
5. A method of automatic face auditing according to any of claims 1 to 3, in which the process of in vivo detection is:
acquiring a depth image, and carrying out normalization processing on a face region in the image to obtain a processed face depth image;
and inputting the RGB face image of the picture and the face depth image into a deep learning network for detection, and obtaining a living body judgment result of the picture.
6. A method of automatic face auditing according to any of claims 1 to 3, in which the process of in vivo detection is:
the face is intercepted from the original image, RGB channels are converted into HSV and YCbCr spaces, and the converted HSV and YCbCr images are overlapped to obtain an overlapped image; extracting Sobel features from the face region through a Sobel operator to obtain a Sobel feature map;
and respectively inputting the Sobel characteristic diagram and the superposition diagram of the picture from two input channels of the double-flow neural network to obtain a living body judgment result of the picture.
7. An automated face auditing system, comprising:
the human face detection module is used for detecting a human face rectangular frame and human face key points through a cascade neural network algorithm;
the face quality evaluation module is used for carrying out face quality evaluation through quality attributes of a plurality of face pictures and screening high-quality pictures;
the living body detection module is used for detecting whether the pictures are true persons or not by using a double-flow convolutional neural network and filtering pictures without the true persons;
the face comparison module is used for extracting face feature vectors of the pictures, comparing the similarity degree between the face feature vectors of the pictures and the face feature vectors of the pre-stored credentials, and filtering the pictures with low similarity degree;
the quality attributes used for face quality assessment include: the quality attributes used for face quality assessment comprise face gestures, eye states, mouth states, makeup states, overall brightness, left and right face brightness differences, ambiguity and shielding;
the face pose, eye state, mouth state, dressing state, ambiguity and occlusion all adopt a MobileFaceNet structure as a main body to construct a multi-task convolutional neural network, and a plurality of task outputs respectively correspond to each quality attribute of the face;
eye state, mouth state, makeup state and face shielding are used as classification tasks, and a softmax loss function is used as an objective function;
the face gesture, the illuminance of the image and the image ambiguity are used as regression tasks, and a Euclidean loss function is used as an objective function;
the total objective function of the network training comprises a combination of a plurality of Softmax loss functions and a Euclidean loss function, and when a plurality of tasks are learned together, the total objective function is a linear combination of a plurality of loss functions.
8. An automated facial auditing apparatus comprising a memory, a processor, and a computer program stored on the memory and executable on the processor, characterized by: the processor, when executing the program, implements the steps of the automatic face auditing method of any of claims 1-6.
9. A readable storage medium having stored thereon a computer program for automated face auditing, characterized by: the computer program when executed by a processor implements the steps of the automatic face auditing method of any of claims 1-6.
CN201911387766.9A 2019-12-30 2019-12-30 Automatic face auditing method, system, equipment and readable storage medium Active CN111222433B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911387766.9A CN111222433B (en) 2019-12-30 2019-12-30 Automatic face auditing method, system, equipment and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911387766.9A CN111222433B (en) 2019-12-30 2019-12-30 Automatic face auditing method, system, equipment and readable storage medium

Publications (2)

Publication Number Publication Date
CN111222433A CN111222433A (en) 2020-06-02
CN111222433B true CN111222433B (en) 2023-06-20

Family

ID=70829143

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911387766.9A Active CN111222433B (en) 2019-12-30 2019-12-30 Automatic face auditing method, system, equipment and readable storage medium

Country Status (1)

Country Link
CN (1) CN111222433B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113807144A (en) * 2020-06-15 2021-12-17 福建新大陆支付技术有限公司 Testing method of living body detection equipment
CN112200108A (en) * 2020-10-16 2021-01-08 深圳市华付信息技术有限公司 Mask face recognition method
CN112329638A (en) * 2020-11-06 2021-02-05 上海优扬新媒信息技术有限公司 Image scoring method, device and system
CN114093004B (en) * 2021-11-25 2023-05-02 成都智元汇信息技术股份有限公司 Face fusion comparison method and device based on multiple cameras

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102622588A (en) * 2012-03-08 2012-08-01 无锡数字奥森科技有限公司 Dual-certification face anti-counterfeit method and device
CN104143086A (en) * 2014-07-18 2014-11-12 吴建忠 Application technology of portrait comparison to mobile terminal operating system
CN108280399A (en) * 2017-12-27 2018-07-13 武汉普利商用机器有限公司 A kind of scene adaptive face identification method
CN109815826A (en) * 2018-12-28 2019-05-28 新大陆数字技术股份有限公司 The generation method and device of face character model
WO2019100608A1 (en) * 2017-11-21 2019-05-31 平安科技(深圳)有限公司 Video capturing device, face recognition method, system, and computer-readable storage medium
CN110580445A (en) * 2019-07-12 2019-12-17 西北工业大学 Face key point detection method based on GIoU and weighted NMS improvement

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102622588A (en) * 2012-03-08 2012-08-01 无锡数字奥森科技有限公司 Dual-certification face anti-counterfeit method and device
CN104143086A (en) * 2014-07-18 2014-11-12 吴建忠 Application technology of portrait comparison to mobile terminal operating system
WO2019100608A1 (en) * 2017-11-21 2019-05-31 平安科技(深圳)有限公司 Video capturing device, face recognition method, system, and computer-readable storage medium
CN108280399A (en) * 2017-12-27 2018-07-13 武汉普利商用机器有限公司 A kind of scene adaptive face identification method
CN109815826A (en) * 2018-12-28 2019-05-28 新大陆数字技术股份有限公司 The generation method and device of face character model
CN110580445A (en) * 2019-07-12 2019-12-17 西北工业大学 Face key point detection method based on GIoU and weighted NMS improvement

Also Published As

Publication number Publication date
CN111222433A (en) 2020-06-02

Similar Documents

Publication Publication Date Title
CN111222433B (en) Automatic face auditing method, system, equipment and readable storage medium
US11830230B2 (en) Living body detection method based on facial recognition, and electronic device and storage medium
WO2022161286A1 (en) Image detection method, model training method, device, medium, and program product
KR102554724B1 (en) Method for identifying an object in an image and mobile device for practicing the method
US11403874B2 (en) Virtual avatar generation method and apparatus for generating virtual avatar including user selected face property, and storage medium
US8861800B2 (en) Rapid 3D face reconstruction from a 2D image and methods using such rapid 3D face reconstruction
CN108090511B (en) Image classification method and device, electronic equipment and readable storage medium
WO2021196389A1 (en) Facial action unit recognition method and apparatus, electronic device, and storage medium
US20110299741A1 (en) Distinguishing Live Faces from Flat Surfaces
CN110728209A (en) Gesture recognition method and device, electronic equipment and storage medium
CN112733802B (en) Image occlusion detection method and device, electronic equipment and storage medium
CN112396050B (en) Image processing method, device and storage medium
CN111382791B (en) Deep learning task processing method, image recognition task processing method and device
CN107369086A (en) A kind of identity card stamp system and method
CN115115552B (en) Image correction model training method, image correction device and computer equipment
CN113434912B (en) Material compliance verification method and device
CN112822393B (en) Image processing method and device and electronic equipment
CN115019364A (en) Identity authentication method and device based on face recognition, electronic equipment and medium
Shukla et al. Deep Learning Model to Identify Hide Images using CNN Algorithm
CN105760881A (en) Facial modeling detection method based on Haar classifier method
CN112132107A (en) Image processing method, image processing device, electronic equipment and computer readable storage medium
CN114140839A (en) Image sending method, device and equipment for face recognition and storage medium
Grinchuk et al. Training a multimodal neural network to determine the authenticity of images
CN116665283B (en) Construction method and system of class face data set
CN117496019B (en) Image animation processing method and system for driving static image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant