CN110956082A - Face key point detection method and detection system based on deep learning - Google Patents

Face key point detection method and detection system based on deep learning Download PDF

Info

Publication number
CN110956082A
CN110956082A CN201910986156.4A CN201910986156A CN110956082A CN 110956082 A CN110956082 A CN 110956082A CN 201910986156 A CN201910986156 A CN 201910986156A CN 110956082 A CN110956082 A CN 110956082A
Authority
CN
China
Prior art keywords
face
key point
training
point detection
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910986156.4A
Other languages
Chinese (zh)
Other versions
CN110956082B (en
Inventor
马国军
马道懿
朱琎
唐跃
曾庆军
夏健
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu University of Science and Technology
Original Assignee
Jiangsu University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu University of Science and Technology filed Critical Jiangsu University of Science and Technology
Priority to CN201910986156.4A priority Critical patent/CN110956082B/en
Publication of CN110956082A publication Critical patent/CN110956082A/en
Application granted granted Critical
Publication of CN110956082B publication Critical patent/CN110956082B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The invention discloses a face key point detection method and a face key point detection system based on deep learning, wherein the detection method comprises the following steps: 1. constructing a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; 2. taking a face data set of Facial Key-points Detection on Kaggle as a sample image, and training the face Key point Detection network constructed in the step 1; 3. collecting an image of a face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier; 4. inputting the face region acquired in the step 3 into a trained face key point detection network to obtain coordinates of a left eye and a nose tip; 5. and calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points. The method can obtain higher detection precision on a deep learning network with smaller scale.

Description

Face key point detection method and detection system based on deep learning
Technical Field
The invention belongs to the technical field of face key point detection, and particularly relates to a method and a system for detecting face key points by applying deep learning.
Background
Face detection technology was first generated with face recognition technology, in the sixth and seventies of the 20 th century, and has been developed for more than half a century since then. Many scientific researchers do a lot of research in this period to make different types of face key point detection systems with different modes. However, until the end of the last century, many systems obtain the position of the face through an external auxiliary sensor or the movement of the face is not obviously and easily captured, so the technology is immature, and the development of the face detection technology is slow in this period.
With the development of society, especially with the rapid development of computer technology and the massive application of biometric identification in the last two decades, face detection technology has been used for not only face identification. The application range of the face detection technology is developed from the original pure scientific research to the aspects of business, military, security and the like at present, and the attention degree of the technology also reaches a new level. In which the convolutional neural network CNN plays an irreplaceable role.
Convolutional neural is a feedforward network, and its artificial neuron can cover units in the range, including convolutional layer and pooling layer. At present, the convolutional neural network CNN has been widely applied to the computer vision domain, and has achieved good effect. In recent years, the performance of CNN in ImageNet competition, it can be seen that in pursuit of classification accuracy, the depth of the model is deeper and deeper, and the complexity of the model is higher and higher, such as a depth residual error network, the number of layers is up to 152.
However, in some real application scenarios, such as mobile or embedded devices, such large and complex models are difficult to apply. Firstly, the models are too large and face the problem of insufficient memory, and secondly, the scenes require low delay or high response speed, for example, the pedestrian detection system of an automatic driving automobile has high requirement on the response speed. Therefore, it is very important to construct a small-scale and efficient CNN model.
Disclosure of Invention
The purpose of the invention is as follows: the invention aims to provide a face key point detection method, which can obtain higher detection precision on a deep learning network with smaller scale.
The technical scheme is as follows: the invention discloses a face key point detection method based on deep learning, which comprises a training stage and a detection stage, wherein the training stage comprises the following steps:
(1) constructing a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the key points of the human face are coordinates of the left eye and the nose tip;
the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a common convolution layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-connection layer, and the sixteenth layer is a softmax layer;
(2) training the face Key point Detection network constructed in the step (1) by taking a face data set of Facial Key-points Detection on Kaggle as a sample image to obtain a trained face Key point Detection network;
the detection stage comprises the following steps:
(3) collecting an image of a face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier;
(4) inputting the face region acquired in the step (3) into a trained face key point detection network to obtain coordinates of a left eye and a nose tip;
(5) calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points;
Figure BDA0002236756520000021
wherein (X)left_eye,Yleft_eye),(Xnoise,Ynoise),(Xright_eye,Yright_eye),(Xleft_mouth,Yleft_mouth),(Xright_mouth,Yright_mouth) The coordinates of the left eye, the nose tip, the right eye, the left mouth corner and the right mouth corner are respectively.
Further, the step (2) further comprises preprocessing the training sample, including:
(2.1) eliminating training samples with missing data;
(2.2) standardizing the training samples: normalized gray value of pixel i in sample image A
Figure BDA0002236756520000022
Comprises the following steps:
Figure BDA0002236756520000023
wherein g isiFor the gray value of pixel i in sample image A before normalization, gminIs the minimum value of the pixel gray levels in all sample images, gmaxIs the maximum value of the pixel gray levels in all sample images.
In order to reduce the dependence of the network on initialization, in the first layer to the thirteenth layer of the face key point detection network, the input of each layer is subjected to batch standardization through Batchnormalization and then convolution operation is performed.
In order to prevent overfitting, an early-stop method is adopted for training in the step (2), the sample images are divided into a training set and a verification set, a face key point detection network is trained on the training set, and errors of the face key point detection network are verified by the verification set;
the training steps are as follows:
(3.1) initializing the error of the optimal verification set to be the error of the verification set after the first training is finished, and initializing the error training times of the optimal verification set to be 0;
(3.2) after each training is finished, verifying errors of the face key point detection network by using a verification set, and comparing the current verification set errors with the optimal verification set errors; if the error of the current verification set is smaller than the error of the optimal verification set, updating the error of the optimal verification set into the error of the current verification set, and setting the error training times of the optimal verification set to be 0; if the error of the current verification set is larger than the error of the optimal verification set, adding one to the error training times of the optimal verification set;
and (3.3) stopping training if the error training times of the optimal verification set reach a preset error training time threshold of the optimal verification set or the total training times reach a preset total training time threshold, and taking the face key point detection network parameters corresponding to the errors of the optimal verification set as training results.
On the other hand, the invention discloses a detection system for realizing the human face key point detection method, which comprises an image acquisition module, a human face detection module and a human face key point detection and labeling module;
the image acquisition module is used for acquiring an image of a face to be detected;
the face detection module is used for acquiring a face area in an image of a face to be detected;
the face key point detection and labeling module is used for acquiring and labeling coordinates of a left eye, a nose tip, a right eye, a left mouth corner and a right mouth corner in a face region.
The face key point detection and labeling module is a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
The face key point detection and labeling module is a computer provided with a NVIDIAGTX 1080 GPU.
Has the advantages that: compared with the prior art, the face key point detection method based on deep learning disclosed by the invention has the following advantages: 1. compared with the common convolutional neural network, the constructed human face key point detection network based on the MobileNet V1 framework has fewer model parameters under the condition of the same detection effect; meanwhile, only 2 key points of the face are detected, and the training and detecting speeds are further increased; 2. on the basis of the detected 2 key points, other feature point coordinates can be obtained.
Drawings
FIG. 1 is a flow chart of a face key point detection method disclosed by the present invention;
FIG. 2 is a schematic structural diagram of a face key point detection network constructed by the present invention;
FIG. 3 is a schematic diagram of the operation of depth convolution and point-by-point convolution;
fig. 4 is a schematic structural diagram of a face key point detection system disclosed in the present invention.
Detailed Description
The invention is further elucidated with reference to the drawings and the detailed description.
As shown in fig. 1, the invention discloses a method for detecting key points of a human face based on deep learning, which comprises a training stage and a detection stage, wherein the training stage comprises the following steps:
step 1, constructing a face key point detection network based on a MobileNet V1 framework; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the key points of the human face are coordinates of the left eye and the nose tip;
as shown in fig. 2, the face key point detection network includes 16 layers of networks connected in sequence, where the first layer is a common convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
The 1-13 layers are feature extraction sub-networks and are used for extracting local features of the image; and the 14-16 layers are feature classification sub-networks, and face key points are obtained according to local features.
The depth convolution is calculated as:
the method comprises the following steps: splitting an image of size M x N into N images of size M x 1;
step two: defining 1 convolution kernel of F x F1 (F is usually an odd number) for each single-channel image, a total of N need to be defined;
step three: in order to make the convolved output image equal in size to the input image, padding is performed, and the size of the padding for the upper, lower, left, and right sides is (F-1)/2.
Step four: convolving each pixel point in each single-channel image with the convolution kernel of the corresponding single channel, and assuming that F is 2a +1, since the size of the convolution kernel is usually odd, the calculation formula is as follows:
Figure BDA0002236756520000051
where w (i, j) is the coefficient in the convolution kernel of F x F1, F (x, y) is any point pixel in the image, and g (x, y) is the convolution output of F (x, y), where x, y are variable so that each pixel in w can access each pixel in F.
The calculation of the point-by-point convolution is:
and adding the results of the N convolution operations by using 1 × N convolution to obtain an output of M × 1.
The process of depth separable convolution of a three channel image using 3 x 3 convolution kernel is illustrated in fig. 3. The three-channel image is changed into three single-channel images, and then convolution operation is carried out on each single channel. At any point f (x, y) in a single channel image, the output g (x, y) is the sum of the products of the convolution kernel coefficients and the image pixels surrounded by the convolution kernel:
g(x,y)=w(-1,-1)f(x-1,y-1)+w(-1,0)f(x-1,y)+...........+w(1,1)f(x+1,y+1)
the three single-channel image outputs g (x, y) are then summed by a 1 x 3 convolution kernel to give the final output z (x, y).
In order to reduce the dependence of the network on initialization, in the first layer to the thirteenth layer of the face key point detection network, the input of each layer is subjected to batch standardization through Batchnormalization and then convolution operation is performed. And simultaneously, using a ReLU linear activation function on each convolutional layer to output convolutional layer results.
Step 2, taking a Facial data set of Facial Key-points Detection on Kaggle as a sample image, and training the Facial Key point Detection network constructed in the step 1 to obtain a trained Facial Key point Detection network;
firstly, training samples are preprocessed, and the preprocessing comprises the following steps:
(2.1) eliminating training samples with missing data;
(2.2) standardizing the training samples: normalized gray value of pixel i in sample image A
Figure BDA0002236756520000052
Comprises the following steps:
Figure BDA0002236756520000061
wherein g isiFor the gray value of pixel i in sample image A before normalization, gminIs the minimum value of the pixel gray levels in all sample images, gmaxIs the maximum value of the pixel gray levels in all sample images.
The method adopts a random gradient descent algorithm (Adam) minimum loss function of adaptive matrix estimation to train the face key point detection network, and the loss function adopts a mean square error loss function.
In order to prevent overfitting, an early-stop method is adopted for training in the step (2), the sample images are divided into a training set and a verification set, a face key point detection network is trained on the training set, and errors of the face key point detection network are verified by the verification set;
the training steps are as follows:
(3.1) initializing the error of the optimal verification set to be the error of the verification set after the first training is finished, and initializing the error training times of the optimal verification set to be 0;
(3.2) after each training is finished, verifying errors of the face key point detection network by using a verification set, and comparing the current verification set errors with the optimal verification set errors; if the error of the current verification set is smaller than the error of the optimal verification set, updating the error of the optimal verification set into the error of the current verification set, and setting the error training times of the optimal verification set to be 0; if the error of the current verification set is larger than the error of the optimal verification set, adding one to the error training times of the optimal verification set;
and (3.3) stopping training if the error training times of the optimal verification set reach a preset error training time threshold of the optimal verification set or the total training times reach a preset total training time threshold, and taking the face key point detection network parameters corresponding to the errors of the optimal verification set as training results.
The detection stage comprises the following steps:
step 3, collecting the image of the face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier;
in the present invention, the face area resize to be acquired is 96 × 96 × 1.
Step 4, inputting the face region obtained in the step 3 into a trained face key point detection network to obtain coordinates of a left eye and a nose tip;
step 5, calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points;
the longitudinal distance from the left eye to the nose tip is found to be equal to the longitudinal distance from the left mouth corner to the nose tip through statistics, and the abscissa of the left mouth corner is equal to the abscissa of the left eye; the longitudinal distance from the right eye to the tip of the nose and the longitudinal distance from the right mouth corner to the tip of the nose are equal and the abscissa of the right mouth corner and the right eye is equal. Obtaining the coordinates of other key points on the face according to the corresponding relation, wherein the calculation formula is as follows:
Figure BDA0002236756520000071
wherein (X)left_eye,Yleft_eye),(Xnoise,Ynoise),(Xright_eye,Yright_eye),(Xleft_mouth,Yleft_mouth),(Xright_mouth,Yright_mouth) The coordinates of the left eye, the nose tip, the right eye, the left mouth corner and the right mouth corner are respectively.
As shown in fig. 4, a detection system for implementing the face key point detection method includes an image acquisition module, a face detection module, and a face key point detection and labeling module;
the image acquisition module is used for acquiring an image of a face to be detected; in the invention, a monocular camera is adopted to acquire an image of a face to be detected, and the acquired image is preprocessed and converted into a gray image;
the face detection module is used for acquiring a face area in an image of a face to be detected and framing the face area by using a rectangular frame;
the face key point detection and labeling module is used for acquiring and labeling coordinates of a left eye, a nose tip, a right eye, a left mouth corner and a right mouth corner in a face region.
The face key point detection and labeling module is a face key point detection network based on a MobileNet V1 framework; the input of the human face key point detection network is a human face image, and the output is coordinates of a left eye and a nose tip; the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
The coordinates of the right eye, the left mouth angle and the right mouth angle are calculated according to the formula (1).
The face key point detection and labeling module is a computer provided with a NVIDIAGTX 1080 GPU.

Claims (7)

1. The method for detecting the key points of the human face based on deep learning is characterized by comprising a training stage and a detection stage, wherein the training stage comprises the following steps:
(1) constructing a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the key points of the human face are coordinates of the left eye and the nose tip;
the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a common convolution layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-connection layer, and the sixteenth layer is a softmax layer;
(2) training the face Key point Detection network constructed in the step (1) by taking a face data set of Facial Key-points Detection on Kaggle as a sample image to obtain a trained face Key point Detection network;
the detection stage comprises the following steps:
(3) collecting an image of a face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier;
(4) inputting the face region acquired in the step (3) into a trained face key point detection network to obtain coordinates of a left eye and a nose tip;
(5) calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points;
Figure FDA0002236756510000011
wherein (X)left_eye,Yleft_eye),(Xnoise,Ynoise),(Xright_eye,Yright_eye),(Xleft_mouth,Yleft_mouth),(Xright_mouth,Yright_mouth) The coordinates of the left eye, the nose tip, the right eye, the left mouth corner and the right mouth corner are respectively.
2. The method for detecting key points of a human face based on deep learning of claim 1, wherein the step (2) further comprises preprocessing training samples, comprising:
(2.1) eliminating training samples with missing data;
(2.2) standardizing the training samples: pixel i criteria in sample image ANormalized gray scale value
Figure FDA0002236756510000012
Comprises the following steps:
Figure FDA0002236756510000021
wherein g isiFor the gray value of pixel i in sample image A before normalization, gminIs the minimum value of the pixel gray levels in all sample images, gmaxIs the maximum value of the pixel gray levels in all sample images.
3. The method according to claim 1, wherein the input of each layer is standardized in batch by BatchNormal and then convolution is performed in the first to thirteenth layers of the face key point detection network.
4. The method for detecting key points of a human face based on deep learning of claim 1, wherein in the step (2), the sample images are divided into a training set and a verification set, a human face key point detection network is trained on the training set, and the verification set is used for verifying the error of the human face key point detection network; the training steps are as follows:
(3.1) initializing the error of the optimal verification set to be the error of the verification set after the first training is finished, and initializing the error training times of the optimal verification set to be 0;
(3.2) after each training is finished, verifying errors of the face key point detection network by using a verification set, and comparing the current verification set errors with the optimal verification set errors; if the error of the current verification set is smaller than the error of the optimal verification set, updating the error of the optimal verification set into the error of the current verification set, and setting the error training times of the optimal verification set to be 0; if the error of the current verification set is larger than the error of the optimal verification set, adding one to the error training times of the optimal verification set;
and (3.3) stopping training if the error training times of the optimal verification set reach a preset error training time threshold of the optimal verification set or the total training times reach a preset total training time threshold, and taking the face key point detection network parameters corresponding to the errors of the optimal verification set as training results.
And when the error of the face key point detection network on the verification set is larger than that of the last training result, stopping training, and adopting the parameters of the last training result as the final parameters of the face key point detection network.
5. The human face key point detection system based on deep learning is characterized by comprising an image acquisition module, a human face detection module and a human face key point detection and labeling module;
the image acquisition module is used for acquiring an image of a face to be detected;
the face detection module is used for acquiring a face area in an image of a face to be detected;
the face key point detection and labeling module is used for acquiring and labeling coordinates of a left eye, a nose tip, a right eye, a left mouth corner and a right mouth corner in a face region.
6. The deep learning based face keypoint detection system of claim 5, wherein the face keypoint detection and labeling module is a face keypoint detection network based on the MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
7. The deep learning based face keypoint detection system of claim 6, wherein the face keypoint detection and labeling module is a computer equipped with a NVIDIAGTX 1080 GPU.
CN201910986156.4A 2019-10-17 2019-10-17 Face key point detection method and detection system based on deep learning Active CN110956082B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910986156.4A CN110956082B (en) 2019-10-17 2019-10-17 Face key point detection method and detection system based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910986156.4A CN110956082B (en) 2019-10-17 2019-10-17 Face key point detection method and detection system based on deep learning

Publications (2)

Publication Number Publication Date
CN110956082A true CN110956082A (en) 2020-04-03
CN110956082B CN110956082B (en) 2023-03-24

Family

ID=69975622

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910986156.4A Active CN110956082B (en) 2019-10-17 2019-10-17 Face key point detection method and detection system based on deep learning

Country Status (1)

Country Link
CN (1) CN110956082B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112200008A (en) * 2020-09-15 2021-01-08 青岛邃智信息科技有限公司 Face attribute recognition method in community monitoring scene
CN112597888A (en) * 2020-12-22 2021-04-02 西北工业大学 On-line education scene student attention recognition method aiming at CPU operation optimization
CN112883854A (en) * 2021-02-04 2021-06-01 新绎健康科技有限公司 Facial region positioning method and system based on deep learning
CN113011356A (en) * 2021-03-26 2021-06-22 杭州朗和科技有限公司 Face feature detection method, device, medium and electronic equipment
CN113822256A (en) * 2021-11-24 2021-12-21 北京的卢深视科技有限公司 Face recognition method, electronic device and storage medium
CN114881893A (en) * 2022-07-05 2022-08-09 腾讯科技(深圳)有限公司 Image processing method, device, equipment and computer readable storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105678248A (en) * 2015-12-31 2016-06-15 上海科技大学 Face key point alignment algorithm based on deep learning
WO2017031089A1 (en) * 2015-08-15 2017-02-23 Eyefluence, Inc. Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects
CN108229442A (en) * 2018-02-07 2018-06-29 西南科技大学 Face fast and stable detection method in image sequence based on MS-KCF
CN108898112A (en) * 2018-07-03 2018-11-27 东北大学 A kind of near-infrared human face in-vivo detection method and system
CN109344693A (en) * 2018-08-13 2019-02-15 华南理工大学 A kind of face multizone fusion expression recognition method based on deep learning
CN110119702A (en) * 2019-04-30 2019-08-13 西安理工大学 Facial expression recognizing method based on deep learning priori
CN110263691A (en) * 2019-06-12 2019-09-20 合肥中科奔巴科技有限公司 Head movement detection method based on android system

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017031089A1 (en) * 2015-08-15 2017-02-23 Eyefluence, Inc. Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects
EP3335096A1 (en) * 2015-08-15 2018-06-20 Google LLC Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects
CN105678248A (en) * 2015-12-31 2016-06-15 上海科技大学 Face key point alignment algorithm based on deep learning
CN108229442A (en) * 2018-02-07 2018-06-29 西南科技大学 Face fast and stable detection method in image sequence based on MS-KCF
CN108898112A (en) * 2018-07-03 2018-11-27 东北大学 A kind of near-infrared human face in-vivo detection method and system
CN109344693A (en) * 2018-08-13 2019-02-15 华南理工大学 A kind of face multizone fusion expression recognition method based on deep learning
CN110119702A (en) * 2019-04-30 2019-08-13 西安理工大学 Facial expression recognizing method based on deep learning priori
CN110263691A (en) * 2019-06-12 2019-09-20 合肥中科奔巴科技有限公司 Head movement detection method based on android system

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
YI SUN ET AL.: ""Deep Convolutional Network Cascade for Facial Point Detection"", 《2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
姜开永等: "基于深度学习的人脸美丽预测模型及其应用", 《五邑大学学报(自然科学版)》 *
曾家建: ""基于深度学习的人脸美丽预测模型及其应用"", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112200008A (en) * 2020-09-15 2021-01-08 青岛邃智信息科技有限公司 Face attribute recognition method in community monitoring scene
CN112597888A (en) * 2020-12-22 2021-04-02 西北工业大学 On-line education scene student attention recognition method aiming at CPU operation optimization
CN112597888B (en) * 2020-12-22 2024-03-08 西北工业大学 Online education scene student attention recognition method aiming at CPU operation optimization
CN112883854A (en) * 2021-02-04 2021-06-01 新绎健康科技有限公司 Facial region positioning method and system based on deep learning
CN113011356A (en) * 2021-03-26 2021-06-22 杭州朗和科技有限公司 Face feature detection method, device, medium and electronic equipment
CN113822256A (en) * 2021-11-24 2021-12-21 北京的卢深视科技有限公司 Face recognition method, electronic device and storage medium
CN113822256B (en) * 2021-11-24 2022-03-25 北京的卢深视科技有限公司 Face recognition method, electronic device and storage medium
CN114881893A (en) * 2022-07-05 2022-08-09 腾讯科技(深圳)有限公司 Image processing method, device, equipment and computer readable storage medium

Also Published As

Publication number Publication date
CN110956082B (en) 2023-03-24

Similar Documents

Publication Publication Date Title
CN110956082B (en) Face key point detection method and detection system based on deep learning
CN109522818B (en) Expression recognition method and device, terminal equipment and storage medium
Rao et al. Selfie video based continuous Indian sign language recognition system
CN106599883B (en) CNN-based multilayer image semantic face recognition method
CN110348376B (en) Pedestrian real-time detection method based on neural network
CN111160269A (en) Face key point detection method and device
CN108009493B (en) Human face anti-cheating recognition method based on motion enhancement
CN108182397B (en) Multi-pose multi-scale human face verification method
CN110570443B (en) Image linear target extraction method based on structural constraint condition generation model
CN111639580B (en) Gait recognition method combining feature separation model and visual angle conversion model
CN113011253B (en) Facial expression recognition method, device, equipment and storage medium based on ResNeXt network
CN106611156B (en) Pedestrian identification method and system based on self-adaptive depth space characteristics
Huang et al. Human emotion recognition based on face and facial expression detection using deep belief network under complicated backgrounds
CN111126250A (en) Pedestrian re-identification method and device based on PTGAN
CN113763417B (en) Target tracking method based on twin network and residual error structure
CN110826534A (en) Face key point detection method and system based on local principal component analysis
CN103235943A (en) Principal component analysis-based (PCA-based) three-dimensional (3D) face recognition system
CN116246338B (en) Behavior recognition method based on graph convolution and transducer composite neural network
KR20180092453A (en) Face recognition method Using convolutional neural network and stereo image
CN112613430B (en) Gait recognition method based on deep migration learning
CN108830166B (en) Real-time bus passenger flow volume statistical method
CN114743257A (en) Method for detecting and identifying image target behaviors
CN114038011A (en) Method for detecting abnormal behaviors of human body in indoor scene
Monica et al. Recognition of medicine using cnn for visually impaired
WO2021056531A1 (en) Face gender recognition method, face gender classifier training method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant