CN110796029B - Face correction and model training method and device, electronic equipment and storage medium - Google Patents

Face correction and model training method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN110796029B
CN110796029B CN201910963567.1A CN201910963567A CN110796029B CN 110796029 B CN110796029 B CN 110796029B CN 201910963567 A CN201910963567 A CN 201910963567A CN 110796029 B CN110796029 B CN 110796029B
Authority
CN
China
Prior art keywords
face
image
detected
detection model
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910963567.1A
Other languages
Chinese (zh)
Other versions
CN110796029A (en
Inventor
杨帆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Dajia Internet Information Technology Co Ltd
Original Assignee
Beijing Dajia Internet Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Dajia Internet Information Technology Co Ltd filed Critical Beijing Dajia Internet Information Technology Co Ltd
Priority to CN201910963567.1A priority Critical patent/CN110796029B/en
Publication of CN110796029A publication Critical patent/CN110796029A/en
Application granted granted Critical
Publication of CN110796029B publication Critical patent/CN110796029B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • G06V10/242Aligning, centring, orientation detection or correction of the image by image rotation, e.g. by 90 degrees
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Human Computer Interaction (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The disclosure relates to a face correction and model training method, a face correction and model training device, an electronic device and a storage medium, relates to the technical field of image processing, and aims to solve the problem of large calculation amount of face detection, and the method comprises the following steps: acquiring an image to be detected containing a human face, and inputting the image to be detected into a human face detection model; acquiring face class characteristics output by the face detection model, wherein the face class characteristics are used for representing the orientation angle range of the face in the image to be detected; determining a rotation direction and a rotation angle for performing direction correction on the face in the image to be detected according to the face type characteristics; the direction of the face in the image to be detected is corrected according to the rotating direction and the rotating angle, and because the output characteristics of the face detection module used in the method are face type characteristics, the orientation angle range of the face in the image to be detected can be directly determined, the face does not need to be rotated in the detection process, and the calculation is simpler.

Description

Face correction and model training method and device, electronic equipment and storage medium
Technical Field
The present disclosure relates to the field of image processing technologies, and in particular, to a method and an apparatus for face correction and model training, an electronic device, and a storage medium.
Background
The face detection is the first step of all face-related tasks, and face recognition, face attribute analysis, face key point extraction and the like can be performed only after the face is detected.
For example, in a large data short video platform, there is a large amount of face data, which in addition to the face in the forward direction, also includes the face in the oblique, lateral, or reverse direction, and in any case, the face in the 360 degree direction. Therefore, in an actual image to be detected, there are sometimes a plurality of faces, and directions of the faces are often different, for example, the image shown in fig. 1 includes faces in eight directions, in order to correctly recognize a face target in each direction, a commonly used face correction method at the present stage is to train a model to have a capability of detecting a forward face, then rotate the image in four directions, detect an image in each rotation direction by the model, and finally fuse detection results in the four directions to remove a repeatedly detected face.
However, in the detection process, the image needs to be rotated, and then four detection processes are performed, and generally, the calculation amount of the detection algorithm is large, and a large amount of calculation resources are consumed for performing four detections.
Disclosure of Invention
The present disclosure provides a method and an apparatus for face correction and model training, an electronic device, and a storage medium, so as to solve at least the problem that in the related art, face detection requires four detections for image rotation, and the amount of calculation is large. The technical scheme of the disclosure is as follows:
according to a first aspect of the embodiments of the present disclosure, there is provided a face correction method, including:
acquiring an image to be detected containing a human face, and inputting the image to be detected into a human face detection model;
acquiring an image to be detected containing a human face, and inputting the image to be detected into a human face detection model;
acquiring face class characteristics output by the face detection model, wherein the face class characteristics are used for representing the orientation angle range of the face in the image to be detected;
determining a rotation direction and a rotation angle for performing direction correction on the face in the image to be detected according to the face type characteristics;
and correcting the direction of the face in the image to be detected according to the rotating direction and the rotating angle.
In an optional implementation manner, the step of determining, according to the face class characteristics, a rotation direction and a rotation angle for performing direction correction on the face in the image to be detected includes:
determining the rotation angle according to the orientation angle range corresponding to the face class characteristic; and
and taking the opposite direction of the target direction as the rotation direction, wherein the target direction is the direction in which a training image used in training the face detection model rotates.
In an optional implementation manner, the step of performing direction correction on the face in the image to be detected according to the rotation direction and the rotation angle includes:
and rotating the image to be detected by the rotation angle according to the rotation direction, and performing direction correction on the face in the image to be detected.
According to a second aspect of the embodiments of the present disclosure, there is provided a face detection model training method, including:
rotating a plurality of identical training images containing human faces, wherein the rotation angles of the training images are different;
determining face class characteristics corresponding to a face orientation angle range to which the face orientation belongs in the rotated training image;
and taking the rotated training image as an input feature of a face detection model, taking the determined face class feature as an output feature of the face detection model, and training the face detection model.
In an alternative embodiment, there is at least a set number of rotated training images per range of orientation angles.
According to a third aspect of the embodiments of the present disclosure, there is provided a face correction apparatus, including:
a first acquisition unit configured to perform acquisition of an image to be detected including a human face and input the image to be detected to a human face detection model;
a second obtaining unit configured to perform obtaining a face class feature output by the face detection model, wherein the face class feature is used for representing an orientation angle range of a face in the image to be detected;
a first determination unit configured to determine a rotation direction and a rotation angle for performing direction correction on a face in the image to be detected according to the face class characteristics;
and the human face correction unit is configured to perform direction correction on the human face in the image to be detected according to the rotation direction and the rotation angle.
In an optional implementation manner, the first determining unit is specifically configured to perform:
determining the rotation angle according to the orientation angle range corresponding to the face class characteristics; and
and taking the opposite direction of the target direction as the rotation direction, wherein the target direction is the direction in which a training image used in training the face detection model rotates.
In an alternative embodiment, the face correction unit is specifically configured to perform:
and rotating the image to be detected by the rotation angle according to the rotation direction, and performing direction correction on the face in the image to be detected.
According to a fourth aspect of the embodiments of the present disclosure, there is provided a face detection model training apparatus, including:
an image rotation unit configured to perform rotation of a plurality of identical training images containing faces, each of which is rotated by a different angle;
a second determining unit configured to perform determining a face class feature corresponding to an orientation angle range to which the orientation of the face in the rotated training image belongs;
a model training unit configured to perform training of the face detection model using the rotated training image as an input feature of the face detection model and using the determined face class feature as an output feature of the face detection model.
In an alternative embodiment, there is at least a set number of rotated training images per range of orientation angles.
According to a fifth aspect of embodiments of the present disclosure, there is provided an electronic apparatus including:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the face correction method according to any one of the first aspect of the embodiments of the present disclosure.
According to a sixth aspect of an embodiment of the present disclosure, there is provided an electronic apparatus including:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the face detection model training method according to any one of the second aspect of the embodiments of the present disclosure.
According to a seventh aspect of the embodiments of the present disclosure, there is provided a non-volatile readable storage medium, where instructions when executed by a processor of an electronic device enable the electronic device to perform the face correction method according to any one of the first aspect of the embodiments of the present disclosure.
According to an eighth aspect of the embodiments of the present disclosure, there is provided a non-transitory readable storage medium, wherein instructions of the storage medium, when executed by a processor of an electronic device, enable the electronic device to perform the face detection model training method according to any one of the second aspects of the embodiments of the present disclosure.
According to a ninth aspect of embodiments of the present disclosure, there is provided a computer program product, which, when run on an electronic device, causes the electronic device to perform a method that implements the first aspect of embodiments of the present disclosure described above and any one of the first aspects related thereto.
According to a tenth aspect of the embodiments of the present disclosure, there is provided a computer program product, which, when run on an electronic device, causes the electronic device to perform a method that implements the second aspect of the embodiments of the present disclosure described above and any one of the possible references to the second aspect.
The technical scheme provided by the embodiment of the disclosure at least brings the following beneficial effects:
because the convolutional neural network is utilized to have no rotation invariance, in order to detect the faces in all directions in the image, the face detection model is divided into the orientation angle ranges of a plurality of faces, each orientation angle range corresponds to one category respectively, and the face detection model used in the embodiment of the invention can directly output the face category characteristics, so that the face detection model can be used for detecting the image to be detected to directly determine the orientation of the face, and further the direction of the face in the image to be detected can be corrected according to the determined rotation direction and the determined rotation angle.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and, together with the description, serve to explain the principles of the disclosure and are not to be construed as limiting the disclosure.
FIG. 1 is a schematic diagram illustrating an image to be detected in accordance with an exemplary embodiment;
FIG. 2 is a flow diagram illustrating a method for training a face detection model in accordance with an exemplary embodiment;
FIG. 3 is a flow diagram illustrating a method of face correction according to an exemplary embodiment;
FIG. 4A is a block diagram illustrating a face detection model in accordance with an exemplary embodiment;
FIG. 4B is a diagram illustrating a categorization by angular range according to an exemplary embodiment;
FIG. 5A is a schematic diagram illustrating a first type of image rotation, according to an exemplary embodiment;
FIG. 5B is a schematic diagram illustrating a second type of image rotation, according to an exemplary embodiment;
FIG. 5C is a schematic diagram illustrating a third image rotation according to an exemplary embodiment;
FIG. 5D is a schematic diagram illustrating a first type of image correction, according to an exemplary embodiment;
FIG. 5E is a schematic diagram illustrating a second type of image correction, according to an exemplary embodiment;
FIG. 5F is a schematic illustration of a third type of image correction shown in accordance with an exemplary embodiment;
FIG. 6 is a flow diagram illustrating a complete method of face detection in accordance with an exemplary embodiment;
FIG. 7 is a flow diagram illustrating a complete method of face correction according to an exemplary embodiment;
FIG. 8 is a block diagram illustrating a face correction device according to an exemplary embodiment;
FIG. 9 is a block diagram of an electronic device shown in accordance with an exemplary embodiment;
FIG. 10 is a block diagram illustrating a face detection model training apparatus in accordance with an exemplary embodiment;
FIG. 11 is a block diagram illustrating an electronic device in accordance with an example embodiment.
Detailed Description
In order to make the technical solutions of the present disclosure better understood by those of ordinary skill in the art, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the accompanying drawings.
It should be noted that the terms "first," "second," and the like in the description and claims of the present disclosure and in the foregoing drawings are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the disclosure described herein are capable of operation in sequences other than those illustrated or otherwise described herein. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.
Some of the words that appear in the text are explained below:
1. the term "and/or" in the embodiments of the present disclosure describes an association relationship of associated objects, and means that there may be three relationships, for example, a and/or B, which may mean: a exists alone, A and B exist simultaneously, and B exists alone. The character "/" generally indicates that the former and latter associated objects are in an "or" relationship.
2. The term "electronic equipment" in the embodiments of the present disclosure refers to equipment that is composed of electronic components such as integrated circuits, transistors, and electronic tubes, and functions by applying electronic technology (including) software, and includes electronic computers, robots controlled by the electronic computers, numerical control or program control systems, and the like.
3. The term "positive sample" in the embodiments of the present disclosure refers to a target object to be detected by a task, and the target object in face detection is a face in an image, such as faces of different ethnic ages, faces of different expressions, faces wearing different decorations, and the like, and may be a face of any scale.
4. In the embodiment of the present disclosure, the term "negative sample" refers to a background where an object to be detected by a task is located, for example, in human face detection, more than 99% of the background is a non-human face area, for example, a human face may appear in different environments, such as a street, a room, and the like.
5. The term "RetinaFace" in the disclosed embodiments is a powerful single-stage face detector that performs pixel-wise face localization on various face scales using joint supervised and self-supervised multitask learning.
6. The term "tenserflow" in the embodiment of the present disclosure is a system that transmits a complex data structure to an artificial intelligent neural network for analysis and processing, can be used in the field of deep learning of multiple machines such as voice recognition or image recognition, and can be operated on various devices as small as one smart phone and as large as thousands of data center servers.
The application scenario described in the embodiment of the present disclosure is for more clearly illustrating the technical solution of the embodiment of the present disclosure, and does not form a limitation on the technical solution provided in the embodiment of the present disclosure, and as a person having ordinary skill in the art knows, with the occurrence of a new application scenario, the technical solution provided in the embodiment of the present disclosure is also applicable to similar technical problems. Wherein, in the description of the present disclosure, unless otherwise indicated, "plurality" means.
Face detection is a key link in face recognition, which refers to a computer technology for identity identification by analyzing and comparing face visual characteristic information, is the key point of artificial intelligence and mode recognition at present, and is widely applied to the fields of national security, military security, identity recognition, bank and customs monitoring, access control systems, video conferences and the like.
The face detection means that any given image is searched by adopting a certain strategy to determine whether the image contains a face, and if so, the position, size and posture of the face are returned.
In an actual image to be detected, a plurality of faces sometimes exist, the directions of the faces are often different and have fronts, and the faces also deflect a certain angle towards the left shoulder or the right shoulder, so that the face targets in all directions are correctly identified, and the method is a very popular and very challenging algorithm problem in the field of computer vision. Face detection is also one of the most important business scenarios for artificial intelligence algorithms. In order to improve the Artificial Intelligence algorithm and externally display the technical strength of the Artificial Intelligence, many AI (Artificial Intelligence) companies select an open data set to verify the algorithm capability thereof. Among many data sets, the wildface is the largest-scale and most difficult-to-detect face detection data set disclosed in the industry, and the data set contains 2 ten thousand face images, including about 18 thousand face data, and is basically a forward face.
At present, the most effective face detection algorithm is basically based on deep learning, and a deep learning model does not have rotation invariance to images, so that faces in different rotation directions need to learn a specific pattern (pattern) for the model.
Fig. 2 is a flowchart illustrating a face detection model training method according to an exemplary embodiment, as shown in fig. 2, including the following steps.
In step S21, rotating a plurality of identical training images containing human faces, wherein each training image is rotated by a different angle;
in step S22, a face type feature corresponding to a face direction angle range to which the face direction belongs in the rotated training image is determined;
in step S23, the rotated training image is used as an input feature of a face detection model, and the determined face type feature is used as an output feature of the face detection model, so as to train the face detection model.
Fig. 3 is a flow chart illustrating a method of face correction according to an exemplary embodiment, as shown in fig. 3, including the following steps.
In step S31, an image to be detected including a human face is acquired, and the image to be detected is input to a human face detection model;
the face detection model is obtained by training through the training method shown in fig. 2.
In step S32, a face class feature output by the face detection model is obtained, where the face class feature is used to represent an orientation angle range of a face in the image to be detected;
in step S33, determining a rotation direction and a rotation angle for performing direction correction on the face in the image to be detected according to the face type feature;
in step S34, the direction of the face in the image to be detected is corrected according to the rotation direction and the rotation angle.
By the scheme, the convolutional neural network is utilized to have no rotation invariance, in order to detect faces in all directions in the image, the orientation angle ranges of the faces are divided into several face orientation angle ranges, each orientation angle range corresponds to one category, a plurality of identical training images containing the faces are rotated at different angles before the face detection model is trained, the rotated training images are used as input features when the model is trained, and the face category features are used as output features.
In an embodiment of the present disclosure, the face detection model is a neural network model.
In the disclosed embodiment, there are at least a set number of rotated training images in each orientation angle range.
For example, there are 4 orientation angle ranges, which are-45 to 45 degrees, 45 to 135 degrees, 135 to 225 degrees, 225 to 315 degrees, and if the set number is 100, the 4 orientation angle ranges are guaranteed to have 100 rotated training images, respectively, and the orientation of the face in the rotated training images is in the corresponding orientation angle range, for example, in the face detection, in the training images in the range of-45 to 45 degrees, there may be 10 faces with 10, 20 with 20 degrees, 30 with 30 degrees, 30 with 0 degrees, 10 with-10 degrees, and the like.
The output characteristic of the face detection model is a face type characteristic, and the face type characteristic represents the orientation angle range of the face in the image to be detected, so that the orientation of the face can be determined according to the face type characteristic, namely the rotation direction and the rotation angle for performing direction correction on the face in the image to be detected are determined, and further the direction correction can be performed on the face in the image to be detected.
In the embodiment of the present disclosure, in order to achieve a good detection effect, a retinaFace is used as a basic model, and a basic framework of the retinaFace is shown in fig. 4A.
The basic Network structure selects a resnet50 (Residual Neural Network), and selects block1 (p 2), block2 (p 3), block3 (p 4), block4 (p 5), and a convolution layer (c 6/p 6) with stride (step length) of 2 is made on the basis of block4 to output as a characteristic layer.
In the embodiment of the disclosure, blocks 1 to 4 are 4 blocks included in the face detection model, and correspond to 4 feature layers, which are p2 to p5, where c2 to c5 represent convolutional layers, and c6/p6 is both a convolutional layer and a feature layer.
Optionally, assuming that the number of anchors corresponding to each feature layer is 3, the size of the anchors is selected according to the scale of each feature layer, as shown in the following table.
Figure BDA0002229754140000101
Table 1 Anchor settings table
Wherein, the setting of the anchors in the table 1 is selected according to the dimension of each feature layer, and the larger the dimension of the feature layer is, the larger the anchors are.
The anchor represents a range of a face scale detected by the feature layer, and the larger the anchor is, the larger the range of the detected face scale is, for example, anchor =16 indicates that a face with a size of 16 pixels can be detected, that is, the range of the detectable face scale is about 16x 16, and can be detected by an anchor box (anchor frame) of 16x 16; anchor =25 indicates that a face of 25 pixels in size can be detected, i.e., the range of detectable face dimensions is about 25x 25.
In the embodiment of the present disclosure, the number of negative samples corresponding to different feature layers in the face detection model is determined according to the scale of the feature layer, the number of negative samples corresponding to each feature layer is about the scale of the feature layer, and taking the P2 feature layer as an example, the scale of the P2 feature layer is 160x 160=25600, and thus the number of negative samples corresponding to the P2 feature layer is 25600.
In the embodiment of the present disclosure, the number of negative samples corresponding to each feature layer is approximately the scale of the feature layer, and the larger the scale of the feature layer is, the larger the area is, the larger the number of negative samples corresponding to the feature layer is, as shown in the following table:
feature layer Dimension Number of negative examples
P2 160x 160 25600
P3 80x 80 6400
P4 40x 40 1600
P5 20x 20 400
P6 10x 10 100
TABLE 2 characteristic layer Scale Table
Table 2 is an example of determining the number of negative samples corresponding to each feature layer according to the scale of each feature layer, which is enumerated in the embodiment of the present disclosure.
Wherein each anchor corresponds to one classifier, taking a P2 feature layer as an example, three anchors (16, 20.16, 25.40) correspond to 3 classifiers on the P2 feature layer, the number of negative samples corresponding to the 3 classifiers is 25600, the number of negative samples corresponding to the 3 anchors on the P3 feature layer is 6400, and so on, the number of negative samples corresponding to the 3 anchors on the P4 feature layer is 1600, the number of negative samples corresponding to the classifiers on the P5 feature layer is 400, and the number of negative samples corresponding to the classifiers on the P6 feature layer is 100. Wherein, the ratio of the number of negative samples between the respective feature layers refers to the ratio of the number of negative samples corresponding to each feature layer, so that the ratio of the number of negative samples between the respective feature layers is 25600.
In the embodiment of the present disclosure, a negative sample refers to a non-human face region in an image, taking as an example that three anchors (16, 20.16, 25.40) on a P2 feature layer correspond to 3 classifiers, the number of the negative samples corresponding to the 3 classifiers is 25600, and for a classifier corresponding to anchor =16, 25600 negative samples corresponding to the classifier, that is, 25600 non-human face regions with a scale of about 16x 16; for the classifier corresponding to anchor =20.16, 25600 negative samples, namely 25600 non-face regions with a scale of about 20.16x 20.16, are corresponding to the classifier, wherein multiple negative samples can exist in the same target training image, and the number of negative samples refers to the number of non-face regions, but not the number of non-training images.
In the embodiment of the present disclosure, in order to ensure that the classifier corresponding to each anchor can be well trained, it is necessary to ensure that the positive and negative samples of each classifier are relatively balanced, so the ratio of the number of positive samples on each feature layer should also be 256.
In an alternative embodiment, for 640 × 640 images, the range of positive sample pixels detected by the P2 feature layer is 16 to 28, the range of positive sample pixels detected by the P3 feature layer is 28 to 57, the range of positive sample pixels detected by the P4 feature layer is 57 to 114, the range of positive sample pixels detected by the P5 feature layer is 114 to 227, and the range of positive sample pixels detected by the P6 feature layer is 227 to 640.
It should be noted that the number of anchors and negative samples and the like can ensure that the face detection model realizes multi-scale face detection, and can detect faces of different scales, where the sizes of the faces detected by different feature layers are different, the size of the face detected by the P6 feature layer is the largest, and the size of the face detected by the P2 feature layer is the smallest.
Aiming at the network structure and the Anchor setting, data enhancement is firstly carried out before the face detection model is carried out: collecting an open-source face data set, then obtaining a plurality of identical training images containing human faces by copying images in the face data set, rotating the identical training images containing human faces, and enabling the rotation angles of all the training images to be different.
In an alternative embodiment, in order to detect a face in 360 directions, 360 degrees are divided into four directions, as shown in fig. 4B, the four directions correspond to four orientation angle ranges, each face in the orientation angle range represents a category, faces in different directions are taken as different "objects", a forward face corresponds to data in a first category, a 90-degree face corresponds to data in a second category, a 180-degree face corresponds to data in a third category, and a 270-degree face corresponds to data in a fourth category.
Before the face detection model is trained, a plurality of identical face images obtained after copying are rotated by different angles, for example, the face images are rotated by 90 degrees, 180 degrees and 270 degrees clockwise respectively, and then the change of the corresponding face position is calculated, so that face data of 360 degrees can be obtained.
For example, the face images shown in fig. 5A to 5C, wherein the face on the right side in fig. 5A is obtained by clockwise rotating the face on the left side in fig. 5A by 90 degrees; similarly, the face on the right side in fig. 5B is obtained by clockwise rotating the face on the left side in fig. 5B by 180 degrees; the face on the right side in fig. 5C is obtained by clockwise rotating the face on the left side in fig. 5C by 270 degrees, and the rotation direction of the training image used in the face detection model training in this manner is clockwise.
Optionally, the face on the right side in fig. 5C may also be obtained by rotating the face on the left side in fig. 5C by 90 degrees in the counterclockwise direction, the face on the right side in fig. 5B may also be obtained by rotating the face on the left side in fig. 5B by 180 degrees in the counterclockwise direction, and the face on the right side in fig. 5A may also be obtained by rotating the face on the left side in fig. 5A by 270 degrees in the counterclockwise direction.
In an alternative embodiment, a rotated training image is used as an input feature of a face detection model, and a determined face class feature is used as an output feature of the face detection model, when the face detection model is trained, a tenserflow frame is used for training the face detection model, wherein a cos (cosine) function is used as a learning rate variation mode, a classified cross entropy loss function and a position regression L2 loss function (square loss function) are used as loss functions, a weight ratio is 1, and when the loss functions are reduced to be not changed any more, the training can be stopped.
In the embodiment of the present disclosure, the trained face detection model may detect the orientation of the face in the image or directly correct the face in the image.
In an optional implementation manner, the output feature of the face detection model is a face class feature, when the face detection model is used, an image to be detected may be input to the trained face detection model, and a face is further corrected according to the face class feature output by the face detection model.
An optional implementation manner is that, when further correcting a face according to the face class features output by the face detection model, specifically:
determining a rotation angle for performing direction correction on the face in the image to be detected according to the orientation angle range corresponding to the face class characteristics; and taking the opposite direction of the target direction as the rotation direction, wherein the target direction is the direction in which the training image used in the training of the face detection model is rotated.
For example, in the process of training the face detection model, when a plurality of identical training images containing faces are rotated clockwise, the target direction is clockwise, and thus the rotation direction in the correction is the opposite direction of the target direction, that is, the counterclockwise direction; on the contrary, if the training images containing faces are rotated counterclockwise in the training process of the face detection model, the target direction is counterclockwise, and thus the rotation direction in the correction is the opposite direction of the target direction, that is, clockwise.
When the rotation angle during face correction is determined, assuming that the corresponding orientation angle range of the category is-45 (315) -45 degrees, the rotation angle determined according to the orientation angle range is 0 degree, and correction is not needed; the facing angle range corresponding to the category is 45-135 degrees, the rotation angle determined according to the facing angle range is 90 degrees, and the rotation angle is rotated by 90 degrees anticlockwise during correction; the third category corresponds to the orientation angle range of 135-225 degrees, the rotation angle determined according to the orientation angle range is 180 degrees, and the rotation angle is rotated by 180 degrees anticlockwise during correction; the category four corresponds to a heading angle range of 225 to 315 degrees, and the rotation angle determined from the heading angle range is 270 degrees, and the rotation angle is rotated counterclockwise by 270 degrees during correction.
Taking the target direction as a clockwise direction as an example, when the human face detection is carried out, if the human face class characteristics indicate a class one, the human face in the image to be detected can be judged to face within a range of-45 to 45 degrees in the clockwise direction, the human face is indicated in a forward direction, and the determined rotation angle is 0 degree, so that the correction is not needed;
if the detection result human face class characteristics represent class two, the human face in the image to be detected can be judged to face in the range of 45-135 degrees in the clockwise direction, and the human face image of the class can be corrected by rotating 90 degrees in the anticlockwise direction;
if the detection result shows that the face type feature represents type three, the face orientation in the image to be detected can be judged to be a face within the range of 135-225 degrees in the clockwise direction, and the correction can be completed by rotating the face image of the type by 180 degrees in the anticlockwise direction;
if the detection result indicates that the face type feature represents the type four, the face orientation in the image to be detected can be determined to be a face in the range of 225-315 degrees in the clockwise direction, and the face image of the type can be corrected by rotating 270 degrees in the counterclockwise direction.
As shown in fig. 5D, the face orientation is 90 degrees clockwise, and belongs to the orientation angle range corresponding to the category two, so that the face is rotated 90 degrees counterclockwise when the face correction is performed; similarly, as shown in fig. 5E, the face orientation is 180 degrees clockwise, and belongs to the orientation angle range corresponding to category three, so that the face is rotated 180 degrees counterclockwise when performing face correction; as shown in fig. 5F, the face orientation is clockwise 270 degrees, and belongs to the orientation angle range corresponding to the category four, so that the face is rotated counterclockwise by 270 degrees when the face correction is performed again.
The embodiment completes omnidirectional face detection through one model, can directly judge the face orientation, and provides a good preprocessing mode for subsequent face-related tasks.
It should be noted that the calibration method recited in the above embodiment is only an example, and the method of dividing the orientation angle range is also an example, optionally, when performing face calibration, the face may also be rotated counterclockwise by a corresponding angle, for example, if the orientation of the face is 135 degrees clockwise, the face may be rotated counterclockwise by 135 degrees, and the like when performing face calibration.
In this case, the face detection model finally outputs a corrected face image, that is, a face position feature, where the face position feature is determined after performing direction correction on the face in the image to be detected according to the face category feature corresponding to the image to be detected. In the method, the face class characteristic is only used as an intermediate result when the face detection model carries out face correction.
Fig. 6 is a flowchart illustrating a complete method for training a face detection model according to an exemplary embodiment, which specifically includes the following steps:
s600, collecting an open source face detection data set widget face;
s601, copying the face images in the face detection data set for multiple times;
s602, rotating the copied multiple identical training images containing the human faces clockwise by 90 degrees, 180 degrees and 270 degrees respectively to perform data enhancement;
s603, calculating the position change of the face corresponding to the position before and after rotation, and classifying the rotated training images according to the calculation result, wherein the forward face corresponds to the first class of data, the 90-degree face corresponds to the second class of data, the 180-degree face corresponds to the third class of data, and the 260-degree face corresponds to the fourth class of data;
s604, training a face detection model by using a tensoflow frame and 4 types of data prepared in the step S603;
and S605, stopping training when the loss function is reduced to be not changed any more.
The face detection model in step S604 may have a structure as shown in fig. 4A.
It should be noted that, by using the model trained in step S603, a face in any direction can be detected, if the detection result is the first type, it can be determined that the face is a forward face, and if the detection result is the second type, it can be determined that the face is a face within a range of 90 degrees clockwise, and the face is rotated 90 degrees counterclockwise, so that the face can be corrected.
Fig. 7 is a flowchart illustrating a complete method for face correction according to an exemplary embodiment, which specifically includes the following steps:
s700, acquiring an image to be detected containing a human face;
s701, inputting an image to be detected into a trained face detection model;
s702, determining the face type characteristics of the image to be detected through a face detection model;
s703, determining a rotation angle for performing direction correction on the face in the image to be detected according to the orientation angle range corresponding to the face class characteristics, and taking the opposite direction of a target direction as a rotation direction, wherein the target direction is the direction in which a training image used when a face detection model is trained rotates;
and S704, rotating the image to be detected by a rotation angle according to the target direction, and performing direction correction on the face in the image to be detected.
Fig. 8 is a block diagram illustrating a face correction apparatus according to an exemplary embodiment. Referring to fig. 8, the apparatus includes a first acquisition unit 800, a second acquisition unit 801, a first determination unit 802, and a face correction unit 803.
A first acquisition unit 800 configured to perform acquisition of an image to be detected including a human face and input the image to be detected to a human face detection model;
a second obtaining unit 801 configured to determine a rotation direction and a rotation angle for performing direction correction on the face in the image to be detected according to the face class characteristics;
a first determining unit 802 configured to determine a rotation direction and a rotation angle for performing direction correction on a face in the image to be detected according to the face class characteristics;
a face correction unit 803 configured to perform direction correction on the face in the image to be detected according to the rotation direction and the rotation angle.
In an optional implementation manner, the first determining unit 802 is specifically configured to perform:
determining the rotation angle according to the orientation angle range corresponding to the face class characteristics; and
and taking the opposite direction of the target direction as the rotation direction, wherein the target direction is the direction in which a training image used in training the face detection model rotates.
In an alternative embodiment, the face correction unit 803 is specifically configured to perform:
and rotating the image to be detected by the rotation angle according to the rotation direction, and performing direction correction on the face in the image to be detected.
With regard to the apparatus in the above-described embodiment, the specific manner in which each unit executes the request has been described in detail in the embodiment related to the method, and will not be elaborated here.
Fig. 9 is a block diagram illustrating an electronic device 900 according to an example embodiment, the apparatus comprising:
a processor 910;
a memory 920 for storing instructions executable by the processor 910;
wherein the processor 910 is configured to execute the instructions to implement the face correction method in any of the embodiments of the present disclosure.
In an exemplary embodiment, a storage medium comprising instructions, such as the memory 920 comprising instructions, executable by the processor 910 of the electronic device 900 to perform the above-described method is also provided. Alternatively, the storage medium may be a non-transitory computer readable storage medium, which may be, for example, a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
The embodiments of the present disclosure further provide a computer program product, which, when run on an electronic device, causes the electronic device to execute any one of the above-mentioned face correction methods or any one of the methods that may be involved in implementing any one of the face correction methods in the embodiments of the present disclosure.
FIG. 10 is a block diagram illustrating a face detection model training apparatus according to an exemplary embodiment. Referring to fig. 10, the apparatus includes an image rotation unit 1000, a second determination unit 1001, and a model training unit 1002:
the image rotating unit 1000 is configured to perform rotation on a plurality of identical training images containing human faces, wherein the rotation angle of each training image is different;
the second determining unit 1001 is configured to determine a face class feature corresponding to an orientation angle range to which the orientation of the face in the rotated training image belongs;
the model training unit 1002 is configured to perform training on the face detection model by using the rotated training image as an input feature of the face detection model and using the determined face class feature as an output feature of the face detection model.
In an alternative embodiment, there is at least a set number of rotated training images per range of orientation angles.
Fig. 11 is a block diagram illustrating an electronic device 1100 according to an example embodiment, the apparatus comprising:
a processor 1110;
a memory 1120 for storing instructions executable by the processor 1110;
wherein the processor 1110 is configured to execute the instructions to implement the face detection model training method according to any one of the embodiments of the present disclosure.
In an exemplary embodiment, a storage medium comprising instructions, such as the memory 1120 comprising instructions, executable by the processor 1110 of the electronic device 1100 to perform the method described above is also provided. Alternatively, the storage medium may be a non-transitory computer readable storage medium, which may be, for example, a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
The embodiment of the present disclosure further provides a computer program product, which, when running on an electronic device, enables the electronic device to execute any one of the above methods for training a face detection model or any one of methods for training a face detection model, which may be involved in implementing the embodiments of the present disclosure.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This disclosure is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice in the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements that have been described above and shown in the drawings, and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.

Claims (14)

1. A face correction method, comprising:
acquiring an image to be detected containing a human face, inputting the image to be detected into a human face detection model, and detecting the human face in the image to be detected based on a characteristic layer in the human face detection model; the face detection model comprises a feature layer for detecting faces with different scales and is used for realizing the detection of the faces with different scales;
acquiring face class characteristics output by the face detection model and aiming at the detected face in the image to be detected, wherein the face class characteristics are used for representing the orientation angle range of the face in the image to be detected;
determining a rotation direction and a rotation angle for performing direction correction on the face in the image to be detected according to the face class characteristics;
and correcting the direction of the detected face in the image to be detected according to the rotating direction and the rotating angle.
2. The method according to claim 1, wherein the step of determining a rotation direction and a rotation angle for performing direction correction on the detected face in the image to be detected according to the face classification features comprises:
determining the rotation angle according to the orientation angle range corresponding to the face class characteristics; and
and taking the opposite direction of the target direction as the rotation direction, wherein the target direction is the direction in which a training image used in training the face detection model rotates.
3. The method for correcting human face according to claim 2, wherein the step of correcting direction of the detected human face in the image to be detected according to the rotation direction and the rotation angle comprises:
and rotating the image to be detected by the rotation angle according to the rotation direction, and performing direction correction on the detected face in the image to be detected.
4. A face detection model training method is characterized by comprising the following steps:
rotating a plurality of identical training images containing human faces, wherein the rotation angles of the training images are different;
determining face class characteristics corresponding to a face orientation angle range to which the face orientation belongs in the rotated training image;
and taking the rotated training image as an input feature of a face detection model, taking the determined face class feature as an output feature of the face detection model, training the face detection model, and obtaining the trained face detection model, wherein the face detection model comprises a feature layer for detecting faces with different scales, and is used for carrying out direction correction on the detected face in the image to be detected based on the feature layer in the face detection model after detecting the face in the image to be detected.
5. The training method of the face detection model according to claim 4, wherein there are at least a set number of rotated training images in each orientation angle range.
6. A face correction apparatus, comprising:
the image processing device comprises a first acquisition unit, a second acquisition unit and a third acquisition unit, wherein the first acquisition unit is configured to acquire an image to be detected containing a human face, input the image to be detected into a human face detection model and detect the human face in the image to be detected based on a feature layer in the human face detection model; the face detection model comprises a feature layer for detecting faces with different scales and is used for realizing the detection of the faces with different scales;
a second obtaining unit configured to perform obtaining a face class feature output by the face detection model and aiming at the detected face in the image to be detected, wherein the face class feature is used for representing an orientation angle range of the face in the image to be detected;
a first determination unit configured to determine a rotation direction and a rotation angle for performing direction correction on a face in the image to be detected according to the face class characteristics;
and the human face correction unit is configured to perform direction correction on the detected human face in the image to be detected according to the rotation direction and the rotation angle.
7. The face correction device according to claim 6, characterized in that the first determination unit is specifically configured to perform:
determining the rotation angle according to the orientation angle range corresponding to the face class characteristics; and
and taking the opposite direction of the target direction as the rotation direction, wherein the target direction is the direction in which a training image used in training the face detection model rotates.
8. The face correction apparatus according to claim 6, characterized in that the face correction unit is specifically configured to perform:
and rotating the image to be detected by the rotation angle according to the rotation direction, and performing direction correction on the detected face in the image to be detected.
9. A face detection model training device is characterized by comprising:
an image rotation unit configured to perform rotation of a plurality of identical training images containing faces, each of which is rotated by a different angle;
a second determining unit configured to perform determining a face class feature corresponding to an orientation angle range to which the orientation of the face in the rotated training image belongs;
and the model training unit is configured to execute the operation of taking the rotated training image as the input characteristic of the face detection model and taking the determined face class characteristic as the output characteristic of the face detection model, and then the face detection model is trained to obtain the trained face detection model, wherein the face detection model comprises a characteristic layer for detecting faces with different scales, and the direction of the face in the image to be detected is corrected after the face in the image to be detected is detected based on the characteristic layer in the face detection model.
10. The training apparatus for face detection model according to claim 9, wherein there are at least a predetermined number of rotated training images in each orientation angle range.
11. An electronic device, comprising:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the face correction method of any one of claims 1 to 3.
12. An electronic device, comprising:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the face detection model training method as claimed in claim 4 or claim 5.
13. A storage medium, wherein instructions in the storage medium, when executed by a processor of an electronic device, enable the electronic device to perform the face correction method of any one of claims 1 to 3.
14. A storage medium, wherein instructions in the storage medium, when executed by a processor of an electronic device, enable the electronic device to perform a face detection model training method as claimed in claim 4 or claim 5.
CN201910963567.1A 2019-10-11 2019-10-11 Face correction and model training method and device, electronic equipment and storage medium Active CN110796029B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910963567.1A CN110796029B (en) 2019-10-11 2019-10-11 Face correction and model training method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910963567.1A CN110796029B (en) 2019-10-11 2019-10-11 Face correction and model training method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN110796029A CN110796029A (en) 2020-02-14
CN110796029B true CN110796029B (en) 2022-11-11

Family

ID=69440228

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910963567.1A Active CN110796029B (en) 2019-10-11 2019-10-11 Face correction and model training method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN110796029B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111950567B (en) * 2020-08-18 2024-04-09 创新奇智(成都)科技有限公司 Extractor training method and device, electronic equipment and storage medium
CN112633221B (en) * 2020-12-30 2024-08-09 深圳市捷顺科技实业股份有限公司 Face direction detection method and related device
CN112733849B (en) * 2021-01-11 2024-07-16 浙江智慧视频安防创新中心有限公司 Model training method, image rotation angle correction method and device
CN113542692A (en) * 2021-07-19 2021-10-22 临沂边锋自动化设备有限公司 Face recognition system and method based on monitoring video
CN115941859A (en) * 2021-08-16 2023-04-07 广州视源电子科技股份有限公司 Image direction adjusting method and device, storage medium and electronic equipment
CN114882489B (en) * 2022-07-07 2022-12-16 浙江智慧视频安防创新中心有限公司 Method, device, equipment and medium for horizontally correcting rotating license plate

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106778684A (en) * 2017-01-12 2017-05-31 易视腾科技股份有限公司 deep neural network training method and face identification method
CN108038474A (en) * 2017-12-28 2018-05-15 深圳云天励飞技术有限公司 Method for detecting human face, the training method of convolutional neural networks parameter, device and medium
CN108090470A (en) * 2018-01-10 2018-05-29 浙江大华技术股份有限公司 A kind of face alignment method and device
JP2018101212A (en) * 2016-12-19 2018-06-28 矢崎総業株式会社 On-vehicle device and method for calculating degree of face directed to front side
CN109002769A (en) * 2018-06-22 2018-12-14 深源恒际科技有限公司 A kind of ox face alignment schemes and system based on deep neural network
CN109993137A (en) * 2019-04-09 2019-07-09 安徽大学 A kind of fast face antidote based on convolutional neural networks
CN110163114A (en) * 2019-04-25 2019-08-23 厦门瑞为信息技术有限公司 A kind of facial angle and face method for analyzing ambiguity, system and computer equipment
CN110188667A (en) * 2019-05-28 2019-08-30 复旦大学 It is a kind of based on tripartite fight generate network face ajust method
CN110276274A (en) * 2019-05-31 2019-09-24 东南大学 A kind of depth characteristic spatial attitude face identification method of multitask

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101271515B (en) * 2007-03-21 2014-03-19 株式会社理光 Image detection device capable of recognizing multi-angle objective
US8396263B2 (en) * 2008-12-30 2013-03-12 Nokia Corporation Method, apparatus and computer program product for providing face pose estimation
CN102147851B (en) * 2010-02-08 2014-06-04 株式会社理光 Device and method for judging specific object in multi-angles
CN105760836A (en) * 2016-02-17 2016-07-13 厦门美图之家科技有限公司 Multi-angle face alignment method based on deep learning and system thereof and photographing terminal
US10347218B2 (en) * 2016-07-12 2019-07-09 Qualcomm Incorporated Multiple orientation detection
CN106874861A (en) * 2017-01-22 2017-06-20 北京飞搜科技有限公司 A kind of face antidote and system
CN107368797A (en) * 2017-07-06 2017-11-21 湖南中云飞华信息技术有限公司 The parallel method for detecting human face of multi-angle, device and terminal device
CN107506702B (en) * 2017-08-08 2020-09-11 江西高创保安服务技术有限公司 Multi-angle-based face recognition model training and testing system and method
CN108256490A (en) * 2018-01-25 2018-07-06 上海交通大学 Method for detecting human face and listen to the teacher rate appraisal procedure, system based on Face datection
CN108563997B (en) * 2018-03-16 2021-10-12 新智认知数据服务有限公司 Method and device for establishing face detection model and face recognition
CN108596089A (en) * 2018-04-24 2018-09-28 北京达佳互联信息技术有限公司 Human face posture detection method, device, computer equipment and storage medium
CN109145765B (en) * 2018-07-27 2021-01-15 华南理工大学 Face detection method and device, computer equipment and storage medium
CN109800648B (en) * 2018-12-18 2021-09-28 北京英索科技发展有限公司 Face detection and recognition method and device based on face key point correction
CN109948559A (en) * 2019-03-25 2019-06-28 厦门美图之家科技有限公司 Method for detecting human face and device

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018101212A (en) * 2016-12-19 2018-06-28 矢崎総業株式会社 On-vehicle device and method for calculating degree of face directed to front side
CN106778684A (en) * 2017-01-12 2017-05-31 易视腾科技股份有限公司 deep neural network training method and face identification method
CN108038474A (en) * 2017-12-28 2018-05-15 深圳云天励飞技术有限公司 Method for detecting human face, the training method of convolutional neural networks parameter, device and medium
CN108090470A (en) * 2018-01-10 2018-05-29 浙江大华技术股份有限公司 A kind of face alignment method and device
CN109002769A (en) * 2018-06-22 2018-12-14 深源恒际科技有限公司 A kind of ox face alignment schemes and system based on deep neural network
CN109993137A (en) * 2019-04-09 2019-07-09 安徽大学 A kind of fast face antidote based on convolutional neural networks
CN110163114A (en) * 2019-04-25 2019-08-23 厦门瑞为信息技术有限公司 A kind of facial angle and face method for analyzing ambiguity, system and computer equipment
CN110188667A (en) * 2019-05-28 2019-08-30 复旦大学 It is a kind of based on tripartite fight generate network face ajust method
CN110276274A (en) * 2019-05-31 2019-09-24 东南大学 A kind of depth characteristic spatial attitude face identification method of multitask

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
Detecting Orientation of In-Plain Rotated Face Images Based on Category Classification by Deep Learning;Yoshihiro Shima 等;《Proc. of the 2017 IEEE Region 10 Conference (TENCON)》;20171108;第127-132页 *
Score Calibration in Face Recognition;Miranti Indar Mandasari 等;《WILEY》;20141201;第1-23页 *
基于二级神经网络的多角度人脸检测;王丽萍 等;《计算机工程与应用》;20040921;第46-47、92页 *
基于级联卷积网络的人脸特征点检测;张衡 等;《南京大学学报(自然科学版)》;20190627;第39卷(第3期);第104-110页 *
基于肤色模型、神经网络和人脸结构模型的平面旋转人脸检测;张洪明 等;《计算机学报》;20021112;第25卷(第11期);第1250-1256页 *

Also Published As

Publication number Publication date
CN110796029A (en) 2020-02-14

Similar Documents

Publication Publication Date Title
CN110796029B (en) Face correction and model training method and device, electronic equipment and storage medium
CN110728209B (en) Gesture recognition method and device, electronic equipment and storage medium
US10991074B2 (en) Transforming source domain images into target domain images
CN111738231B (en) Target object detection method and device, computer equipment and storage medium
Kluger et al. Deep learning for vanishing point detection using an inverse gnomonic projection
CN111241989A (en) Image recognition method and device and electronic equipment
CN109584300A (en) A kind of method and device of determining headstock towards angle
CN108960412B (en) Image recognition method, device and computer readable storage medium
Hu et al. Towards requirements specification for machine-learned perception based on human performance
Buoncompagni et al. Saliency-based keypoint selection for fast object detection and matching
Li et al. Visual slam in dynamic scenes based on object tracking and static points detection
CN104239843A (en) Positioning method and device for face feature points
Chidananda et al. Entropy-cum-Hough-transform-based ear detection using ellipsoid particle swarm optimization
CN115937596A (en) Target detection method, training method and device of model thereof, and storage medium
Madessa et al. A deep learning approach for specular highlight removal from transmissive materials
Warif et al. A comprehensive evaluation procedure for copy-move forgery detection methods: results from a systematic review
CN112560880A (en) Object classification method, object classification apparatus, and computer-readable storage medium
Kumar et al. Pose image generation for video content creation using controlled human pose image generation GAN
Berger et al. Visual tracking with vg-ram weightless neural networks
Vacavant A novel definition of robustness for image processing algorithms
CN115147469A (en) Registration method, device, equipment and storage medium
Pasqualino et al. Unsupervised multi-camera domain adaptation for object detection in cultural sites
CN115375618A (en) Defect detection and training method and device, storage medium and equipment
Tang et al. Learning Hough regression models via bridge partial least squares for object detection
CN110705479A (en) Model training method, target recognition method, device, equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant