CN110956082B - Face key point detection method and detection system based on deep learning - Google Patents

Face key point detection method and detection system based on deep learning Download PDF

Info

Publication number
CN110956082B
CN110956082B CN201910986156.4A CN201910986156A CN110956082B CN 110956082 B CN110956082 B CN 110956082B CN 201910986156 A CN201910986156 A CN 201910986156A CN 110956082 B CN110956082 B CN 110956082B
Authority
CN
China
Prior art keywords
key point
face
training
image
point detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910986156.4A
Other languages
Chinese (zh)
Other versions
CN110956082A (en
Inventor
马国军
马道懿
朱琎
唐跃
曾庆军
夏健
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu University of Science and Technology
Original Assignee
Jiangsu University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu University of Science and Technology filed Critical Jiangsu University of Science and Technology
Priority to CN201910986156.4A priority Critical patent/CN110956082B/en
Publication of CN110956082A publication Critical patent/CN110956082A/en
Application granted granted Critical
Publication of CN110956082B publication Critical patent/CN110956082B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The invention discloses a face key point detection method and a face key point detection system based on deep learning, wherein the detection method comprises the following steps: 1. constructing a face key point detection network based on a MobileNet V1 architecture; the input of the face key point detection network is a face image, and the output of the face key point detection network is a face key point coordinate; 2. taking a face data set of Facial Key-points Detection on Kaggle as a sample image, and training the face Key point Detection network constructed in the step 1; 3. collecting an image of a face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier; 4. inputting the face region acquired in the step 3 into a trained face key point detection network to obtain coordinates of a left eye and a nose tip; 5. and calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points. The method can obtain higher detection precision on a deep learning network with smaller scale.

Description

Face key point detection method and detection system based on deep learning
Technical Field
The invention belongs to the technical field of face key point detection, and particularly relates to a method and a system for detecting face key points by applying deep learning.
Background
Face detection technology was first generated with face recognition technology, in the sixth and seventies of the 20 th century, and has been developed for more than half a century since then. Many scientific researchers do a lot of research in this period to make different types of face key point detection systems with different modes. However, until the end of the last century, many systems obtain the position of the face through an external auxiliary sensor or the movement of the face is not obviously and easily captured, so the technology is immature, and the development of the face detection technology is slow in this period.
With the development of society, especially with the rapid development of computer technology and the extensive application of biometric identification in the last two decades, face detection technology has been used not only for face identification. The application range of the face detection technology is developed from the original pure scientific research to the aspects of business, military, security and the like at present, and the attention degree of the technology also reaches a new level. In which the convolutional neural network CNN plays an irreplaceable role.
Convolutional neural is a feedforward network, and its artificial neuron can cover units in the range, including convolutional layer and pooling layer. At present, the convolutional neural network CNN has been widely applied to the computer vision domain, and has achieved good effect. In recent years, the performance of CNN in ImageNet competition, it can be seen that in pursuit of classification accuracy, the depth of the model is deeper and deeper, and the complexity of the model is higher and higher, such as a depth residual error network, the number of layers is up to 152.
However, in some real application scenarios, such as mobile or embedded devices, such large and complex models are difficult to apply. Firstly, the models are too large and face the problem of insufficient memory, and secondly, the scenes require low delay or high response speed, for example, the pedestrian detection system of an automatic driving automobile has high requirement on the response speed. Therefore, it is very important to construct a small-scale and efficient CNN model.
Disclosure of Invention
The purpose of the invention is as follows: the invention aims to provide a face key point detection method, which can obtain higher detection precision on a deep learning network with smaller scale.
The technical scheme is as follows: the invention discloses a face key point detection method based on deep learning, which comprises a training stage and a detection stage, wherein the training stage comprises the following steps:
(1) Constructing a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the key points of the human face are coordinates of the left eye and the nose tip;
the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a common convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-connection layer, and the sixteenth layer is a softmax layer;
(2) Training the face Key point Detection network constructed in the step (1) by taking a face data set of Facial Key-points Detection on Kaggle as a sample image to obtain a trained face Key point Detection network;
the detection stage comprises the following steps:
(3) Collecting an image of a face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open visual constant value) cascade classifier;
(4) Inputting the face region acquired in the step (3) into a trained face key point detection network to obtain coordinates of a left eye and a nose tip;
(5) Calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points;
Figure GDA0003947861470000021
wherein (X) left_eye ,Y left_eye ),(X noise ,Y noise ),(X right_eye ,Y right_eye ),(X left_mouth ,Y left_mouth ),(X right_mouth ,Y right_mouth ) The coordinates of the left eye, the nose tip, the right eye, the left mouth corner and the right mouth corner are respectively.
Further, the step (2) further comprises preprocessing the training sample, including:
(2.1) rejecting training samples with missing data;
(2.2) standardizing the training samples: normalized gray value of pixel i in sample image A
Figure GDA0003947861470000023
Comprises the following steps:
Figure GDA0003947861470000022
wherein g is i Is the gray value, g, of pixel i in sample image A before normalization min Is the minimum value of the pixel gray levels in all sample images, g max Is the maximum value of the pixel gray levels in all sample images.
In order to reduce the dependence of the network on initialization, in the first layer to the thirteenth layer of the face key point detection network, the input of each layer is subjected to batch standardization through Batchnormalization and then convolution operation is performed.
In order to prevent overfitting, an early-stop method is adopted for training in the step (2), the sample images are divided into a training set and a verification set, a face key point detection network is trained on the training set, and errors of the face key point detection network are verified by the verification set;
the training steps are as follows:
(3.1) initializing the error of the optimal verification set into the error of the verification set after the first training is finished, and initializing the error training times of the optimal verification set to 0;
(3.2) after each training is finished, verifying the error of the face key point detection network by using a verification set, and comparing the error of the current verification set with the error of the optimal verification set; if the error of the current verification set is smaller than the error of the optimal verification set, updating the error of the optimal verification set into the error of the current verification set, and setting the error training times of the optimal verification set to be 0; if the error of the current verification set is larger than the error of the optimal verification set, adding one to the error training times of the optimal verification set;
and (3.3) stopping training if the error training times of the optimal verification set reach a preset error training time threshold of the optimal verification set or the total training times reach a preset total training time threshold, and taking the face key point detection network parameters corresponding to the errors of the optimal verification set as training results.
On the other hand, the invention discloses a detection system for realizing the human face key point detection method, which comprises an image acquisition module, a human face detection module and a human face key point detection and labeling module;
the image acquisition module is used for acquiring an image of a face to be detected;
the face detection module is used for acquiring a face area in an image of a face to be detected;
the face key point detection and labeling module is used for acquiring and labeling coordinates of a left eye, a nose tip, a right eye, a left mouth corner and a right mouth corner in a face region.
The face key point detection and labeling module is a face key point detection network based on a MobileNet V1 architecture; the input of the face key point detection network is a face image, and the output of the face key point detection network is a face key point coordinate; the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
The face key point detection and labeling module is a computer provided with a NVIDIAGTX 1080 GPU.
Has the advantages that: compared with the prior art, the face key point detection method based on deep learning disclosed by the invention has the following advantages: 1. compared with the common convolutional neural network, the constructed human face key point detection network based on the MobileNet V1 framework has fewer model parameters under the condition of the same detection effect; meanwhile, only 2 key points of the face are detected, and the training and detection speeds are further increased; 2. on the basis of the detected 2 key points, other feature point coordinates can be obtained.
Drawings
FIG. 1 is a flow chart of a face key point detection method disclosed by the present invention;
FIG. 2 is a schematic structural diagram of a face key point detection network constructed by the present invention;
FIG. 3 is a schematic diagram of the operation of depth convolution and point-by-point convolution;
fig. 4 is a schematic structural diagram of a face key point detection system disclosed in the present invention.
Detailed Description
The invention is further elucidated with reference to the drawings and the detailed description.
As shown in fig. 1, the invention discloses a method for detecting key points of a human face based on deep learning, which comprises a training stage and a detection stage, wherein the training stage comprises the following steps:
step 1, constructing a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the key points of the human face are coordinates of the left eye and the nose tip;
as shown in fig. 2, the face key point detection network includes 16 layers of networks connected in sequence, where the first layer is a common convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
The 1-13 layers are feature extraction sub-networks and are used for extracting local features of the image; and the 14-16 layers are feature classification sub-networks, and face key points are obtained according to local features.
The depth convolution is calculated as:
the method comprises the following steps: splitting an image of size M x N into N images of size M x 1;
step two: defining 1 convolution kernel of F x F1 (F is usually an odd number) for each single-channel image, a total of N need to be defined;
step three: in order to make the convolved output image and the input image equal in size, padding is required, and the size of the padding for the upper, lower, left, and right sides is P = (F-1)/2.
Step four: convolving each pixel point in each single-channel image with the convolution kernel of the corresponding single channel, and assuming that F =2a +1, the calculation formula is as follows because the size of the convolution kernel is usually odd:
Figure GDA0003947861470000051
where w (i, j) is the coefficient in the convolution kernel of F x F1, F (x, y) is any point pixel in the image, and g (x, y) is the convolution output of F (x, y), where x, y are variable so that each pixel in w can access each pixel in F.
The calculation of the point-by-point convolution is:
and adding the results of the N convolution operations by using 1 × N convolution to obtain an output of M × 1.
The process of depth separable convolution of a three channel image using the convolution kernel of 3*3 is illustrated in fig. 3. The three-channel image is changed into three single-channel images, and then convolution operation is carried out on each single channel. At any point f (x, y) in a single channel image, the output g (x, y) is the sum of the products of the convolution kernel coefficients and the image pixels surrounded by the convolution kernel:
g(x,y)=w(-1,-1)f(x-1,y-1)+w(-1,0)f(x-1,y)+...........+w(1,1)f(x+1,y+1)
the three single-channel image outputs g (x, y) are then summed by a 1 x 3 convolution kernel to give the final output z (x, y).
In order to reduce the dependence of the network on initialization, the input of each layer in the first layer to the tenth layer in the face key point detection network is subjected to batch standardization through BatchNormalization and then is subjected to convolution operation. Meanwhile, a ReLU linear activation function is used on each convolution layer, and convolution layer results are output.
Step 2, taking a Facial data set of Facial Key-points Detection on Kaggle as a sample image, and training the Facial Key point Detection network constructed in the step 1 to obtain a trained Facial Key point Detection network;
firstly, training samples are preprocessed, and the preprocessing comprises the following steps:
(2.1) eliminating training samples with data missing;
(2.2) standardizing the training samples: normalized gray value of pixel i in sample image A
Figure GDA0003947861470000052
Comprises the following steps:
Figure GDA0003947861470000061
wherein g is i For the gray value of pixel i in sample image A before normalization, g min Is the minimum value of the pixel gray levels in all sample images, g max Is the maximum value of the pixel gray levels in all sample images.
The method adopts a random gradient descent algorithm (Adam) minimum loss function of adaptive matrix estimation to train the face key point detection network, and the loss function adopts a mean square error loss function.
In order to prevent overfitting, an early-stop method is adopted for training in the step (2), the sample images are divided into a training set and a verification set, a face key point detection network is trained on the training set, and errors of the face key point detection network are verified by the verification set;
the training steps are as follows:
(3.1) initializing the error of the optimal verification set to be the error of the verification set after the first training is finished, and initializing the error training times of the optimal verification set to be 0;
(3.2) after each training is finished, verifying errors of the face key point detection network by using a verification set, and comparing the current verification set errors with the optimal verification set errors; if the error of the current verification set is smaller than the error of the optimal verification set, updating the error of the optimal verification set into the error of the current verification set, and setting the error training times of the optimal verification set to be 0; if the error of the current verification set is larger than the error of the optimal verification set, adding one to the error training times of the optimal verification set;
and (3.3) stopping training if the error training times of the optimal verification set reach a preset error training time threshold of the optimal verification set or the total training times reach a preset total training time threshold, and taking the face key point detection network parameters corresponding to the errors of the optimal verification set as training results.
The detection stage comprises the following steps:
step 3, collecting the image of the face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open visual constant value) cascade classifier;
in the present invention, the face area resize to be acquired is 96 × 96 × 1.
Step 4, inputting the face region obtained in the step 3 into a trained face key point detection network to obtain coordinates of a left eye and a nose tip;
step 5, calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points;
the longitudinal distance from the left eye to the nose tip is found to be equal to the longitudinal distance from the left mouth corner to the nose tip through statistics, and the abscissa of the left mouth corner is equal to the abscissa of the left eye; the longitudinal distance from the right eye to the tip of the nose and the longitudinal distance from the right mouth corner to the tip of the nose are equal and the abscissa of the right mouth corner and the right eye is equal. Obtaining the coordinates of other key points on the face according to the corresponding relation, wherein the calculation formula is as follows:
Figure GDA0003947861470000071
wherein (X) left_eye ,Y left_eye ),(X noise ,Y noise ),(X right_eye ,Y right_eye ),(X left_mouth ,Y left_mouth ),(X right_mouth ,Y right_mouth ) The coordinates of the left eye, the nose tip, the right eye, the left mouth corner and the right mouth corner are respectively.
As shown in fig. 4, a detection system for implementing the face key point detection method includes an image acquisition module, a face detection module, and a face key point detection and labeling module;
the image acquisition module is used for acquiring an image of a human face to be detected; in the invention, a monocular camera is adopted to acquire an image of a face to be detected, and the acquired image is preprocessed and converted into a gray image;
the face detection module is used for acquiring a face area in an image of a face to be detected and framing the face area by using a rectangular frame;
the face key point detection and labeling module is used for acquiring and labeling coordinates of a left eye, a nose tip, a right eye, a left mouth angle and a right mouth angle in a face region.
The face key point detection and labeling module is a face key point detection network based on a MobileNet V1 framework; the input of the human face key point detection network is a human face image, and the output is coordinates of a left eye and a nose tip; the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
The coordinates of the right eye, the left mouth angle and the right mouth angle are calculated according to the formula (1).
The face key point detection and labeling module is a computer provided with a NVIDIAGTX 1080 GPU.

Claims (1)

1. The method for detecting the key points of the human face based on deep learning is characterized by comprising a training stage and a detection stage, wherein the training stage comprises the following steps:
(1) Constructing a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the key points of the human face are coordinates of the left eye and the nose tip;
the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a common convolution layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-connection layer, and the sixteenth layer is a softmax layer; the 1-13 layers are feature extraction sub-networks and are used for extracting local features of the image; the 14-16 layers are feature classification sub-networks, and face key points are obtained according to local features;
the depth convolution is calculated as:
the method comprises the following steps: splitting an image of size M x N into N images of size M x 1;
step two: defining 1 convolution kernel of F1 for each single-channel image, wherein F is an odd number; a total of N is required to be defined;
step three: in order to make the size of the convolved output image equal to that of the input image, white filling is needed, and the size of the white filling is P = (F-1)/2;
step four: convolving each pixel point in each single-channel image with the convolution kernel of the corresponding single channel, and assuming that F =2a +1, the calculation formula is as follows because the size of the convolution kernel is odd:
Figure FDA0003947861460000011
wherein w (i, j) is the coefficient in the convolution kernel of F1, F (x, y) is any point pixel in the image, g (x, y) is the convolution output of F (x, y), where x, y are variable, so that each pixel in w can access each pixel in F;
the calculation of the point-by-point convolution is:
adding the results of the N convolution operations by using 1 × N convolution to obtain an output of M × 1;
firstly, converting a three-channel image into three single-channel images, and then performing convolution operation on each single channel; at any point f (x, y) in a single channel image, the output g (x, y) is the sum of the products of the convolution kernel coefficients and the image pixels surrounded by the convolution kernel:
g(x,y)=w(-1,-1)f(x-1,y-1)+w(-1,0)f(x-1,y)+...........+w(1,1)f(x+1,y+1)
then adding up the three single-channel image outputs g (x, y) by a 1 x 3 convolution kernel to obtain a final output z (x, y);
(2) Training the face Key point Detection network constructed in the step (1) by taking a face data set of Facial Key-points Detection on Kaggle as a sample image to obtain a trained face Key point Detection network;
the detection stage comprises the following steps:
(3) Collecting an image of a face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier;
(4) Inputting the face region acquired in the step (3) into a trained face key point detection network to obtain coordinates of a left eye and a nose tip;
(5) Calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points;
Figure FDA0003947861460000021
wherein (X) left_eye ,Y left_eye ),(X noise ,Y noise ),(X right_eye ,Y right_eye ),(X left_mouth ,Y left_mouth ),(X right_mouth ,Y right_mouth ) Coordinates of a left eye, a nose tip, a right eye, a left mouth corner and a right mouth corner are respectively;
the step (2) further comprises preprocessing a training sample, and the preprocessing comprises the following steps:
(2.1) rejecting training samples with missing data;
(2.2) standardizing the training samples: normalized gray value of pixel i in sample image A
Figure FDA0003947861460000023
Comprises the following steps:
Figure FDA0003947861460000022
wherein g is i Is the gray value, g, of pixel i in sample image A before normalization min Is the minimum value of the pixel gray levels in all sample images, g max The maximum value of the pixel gray levels in all sample images is obtained;
in the first layer to the thirteenth layer in the face key point detection network, the input of each layer is subjected to batch standardization through BatchNormal and then convolution operation is carried out;
in the step (2), the sample image is divided into a training set and a verification set, a face key point detection network is trained on the training set, and the error of the face key point detection network is verified by using the verification set; the training steps are as follows:
(3.1) initializing the error of the optimal verification set to be the error of the verification set after the first training is finished, and initializing the error training times of the optimal verification set to be 0;
(3.2) after each training is finished, verifying errors of the face key point detection network by using a verification set, and comparing the current verification set errors with the optimal verification set errors; if the error of the current verification set is smaller than the error of the optimal verification set, updating the error of the optimal verification set into the error of the current verification set, and setting the error training times of the optimal verification set to be 0; if the error of the current verification set is larger than the error of the optimal verification set, adding one to the error training times of the optimal verification set;
(3.3) if the error training times of the optimal verification set reach a preset error training time threshold of the optimal verification set or the total training times reach a preset total training time threshold, stopping training and taking the face key point detection network parameters corresponding to the errors of the optimal verification set as training results;
and when the error of the face key point detection network on the verification set is larger than that of the last training result, stopping training, and adopting the parameters of the last training result as the final parameters of the face key point detection network.
CN201910986156.4A 2019-10-17 2019-10-17 Face key point detection method and detection system based on deep learning Active CN110956082B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910986156.4A CN110956082B (en) 2019-10-17 2019-10-17 Face key point detection method and detection system based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910986156.4A CN110956082B (en) 2019-10-17 2019-10-17 Face key point detection method and detection system based on deep learning

Publications (2)

Publication Number Publication Date
CN110956082A CN110956082A (en) 2020-04-03
CN110956082B true CN110956082B (en) 2023-03-24

Family

ID=69975622

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910986156.4A Active CN110956082B (en) 2019-10-17 2019-10-17 Face key point detection method and detection system based on deep learning

Country Status (1)

Country Link
CN (1) CN110956082B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112200008A (en) * 2020-09-15 2021-01-08 青岛邃智信息科技有限公司 Face attribute recognition method in community monitoring scene
CN112597888B (en) * 2020-12-22 2024-03-08 西北工业大学 Online education scene student attention recognition method aiming at CPU operation optimization
CN112883854A (en) * 2021-02-04 2021-06-01 新绎健康科技有限公司 Facial region positioning method and system based on deep learning
CN113011356A (en) * 2021-03-26 2021-06-22 杭州朗和科技有限公司 Face feature detection method, device, medium and electronic equipment
CN113822256B (en) * 2021-11-24 2022-03-25 北京的卢深视科技有限公司 Face recognition method, electronic device and storage medium
CN114881893B (en) * 2022-07-05 2022-10-21 腾讯科技(深圳)有限公司 Image processing method, device, equipment and computer readable storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105678248A (en) * 2015-12-31 2016-06-15 上海科技大学 Face key point alignment algorithm based on deep learning
WO2017031089A1 (en) * 2015-08-15 2017-02-23 Eyefluence, Inc. Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects
CN108229442A (en) * 2018-02-07 2018-06-29 西南科技大学 Face fast and stable detection method in image sequence based on MS-KCF
CN108898112A (en) * 2018-07-03 2018-11-27 东北大学 A kind of near-infrared human face in-vivo detection method and system
CN109344693A (en) * 2018-08-13 2019-02-15 华南理工大学 A kind of face multizone fusion expression recognition method based on deep learning
CN110119702A (en) * 2019-04-30 2019-08-13 西安理工大学 Facial expression recognizing method based on deep learning priori
CN110263691A (en) * 2019-06-12 2019-09-20 合肥中科奔巴科技有限公司 Head movement detection method based on android system

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017031089A1 (en) * 2015-08-15 2017-02-23 Eyefluence, Inc. Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects
EP3335096A1 (en) * 2015-08-15 2018-06-20 Google LLC Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects
CN105678248A (en) * 2015-12-31 2016-06-15 上海科技大学 Face key point alignment algorithm based on deep learning
CN108229442A (en) * 2018-02-07 2018-06-29 西南科技大学 Face fast and stable detection method in image sequence based on MS-KCF
CN108898112A (en) * 2018-07-03 2018-11-27 东北大学 A kind of near-infrared human face in-vivo detection method and system
CN109344693A (en) * 2018-08-13 2019-02-15 华南理工大学 A kind of face multizone fusion expression recognition method based on deep learning
CN110119702A (en) * 2019-04-30 2019-08-13 西安理工大学 Facial expression recognizing method based on deep learning priori
CN110263691A (en) * 2019-06-12 2019-09-20 合肥中科奔巴科技有限公司 Head movement detection method based on android system

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"Deep Convolutional Network Cascade for Facial Point Detection";Yi Sun et al.;《2013 IEEE Conference on Computer Vision and Pattern Recognition》;20131003;第3476-3483页 *
"基于深度学习的人脸美丽预测模型及其应用";曾家建;《中国优秀硕士学位论文全文数据库 (信息科技辑)》;20181215(第12期);第I138-1347页 *
基于深度学习的人脸美丽预测模型及其应用;姜开永等;《五邑大学学报(自然科学版)》;20180515(第02期);第45-52页 *

Also Published As

Publication number Publication date
CN110956082A (en) 2020-04-03

Similar Documents

Publication Publication Date Title
CN110956082B (en) Face key point detection method and detection system based on deep learning
CN109522818B (en) Expression recognition method and device, terminal equipment and storage medium
CN110348376B (en) Pedestrian real-time detection method based on neural network
CN106599883B (en) CNN-based multilayer image semantic face recognition method
CN106529447B (en) Method for identifying face of thumbnail
CN108416266B (en) Method for rapidly identifying video behaviors by extracting moving object through optical flow
CN112766160A (en) Face replacement method based on multi-stage attribute encoder and attention mechanism
CN107909005A (en) Personage's gesture recognition method under monitoring scene based on deep learning
CN110399821B (en) Customer satisfaction acquisition method based on facial expression recognition
CN108009493B (en) Human face anti-cheating recognition method based on motion enhancement
CN108182397B (en) Multi-pose multi-scale human face verification method
CN104361316B (en) Dimension emotion recognition method based on multi-scale time sequence modeling
CN106909909B (en) Face detection and alignment method based on shared convolution characteristics
CN112434655A (en) Gait recognition method based on adaptive confidence map convolution network
CN110135277B (en) Human behavior recognition method based on convolutional neural network
CN112464844A (en) Human behavior and action recognition method based on deep learning and moving target detection
CN111382601A (en) Illumination face image recognition preprocessing system and method for generating confrontation network model
EP2790130A1 (en) Method for object recognition
Huang et al. Human emotion recognition based on face and facial expression detection using deep belief network under complicated backgrounds
CN113763417B (en) Target tracking method based on twin network and residual error structure
CN113963333A (en) Traffic sign board detection method based on improved YOLOF model
CN109522865A (en) A kind of characteristic weighing fusion face identification method based on deep neural network
Cho et al. Modified perceptual cycle generative adversarial network-based image enhancement for improving accuracy of low light image segmentation
CN116246338B (en) Behavior recognition method based on graph convolution and transducer composite neural network
KR20180092453A (en) Face recognition method Using convolutional neural network and stereo image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant