CN110956082A - Face key point detection method and detection system based on deep learning - Google Patents
Face key point detection method and detection system based on deep learning Download PDFInfo
- Publication number
- CN110956082A CN110956082A CN201910986156.4A CN201910986156A CN110956082A CN 110956082 A CN110956082 A CN 110956082A CN 201910986156 A CN201910986156 A CN 201910986156A CN 110956082 A CN110956082 A CN 110956082A
- Authority
- CN
- China
- Prior art keywords
- face
- key point
- training
- point detection
- layer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Abstract
The invention discloses a face key point detection method and a face key point detection system based on deep learning, wherein the detection method comprises the following steps: 1. constructing a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; 2. taking a face data set of Facial Key-points Detection on Kaggle as a sample image, and training the face Key point Detection network constructed in the step 1; 3. collecting an image of a face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier; 4. inputting the face region acquired in the step 3 into a trained face key point detection network to obtain coordinates of a left eye and a nose tip; 5. and calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points. The method can obtain higher detection precision on a deep learning network with smaller scale.
Description
Technical Field
The invention belongs to the technical field of face key point detection, and particularly relates to a method and a system for detecting face key points by applying deep learning.
Background
Face detection technology was first generated with face recognition technology, in the sixth and seventies of the 20 th century, and has been developed for more than half a century since then. Many scientific researchers do a lot of research in this period to make different types of face key point detection systems with different modes. However, until the end of the last century, many systems obtain the position of the face through an external auxiliary sensor or the movement of the face is not obviously and easily captured, so the technology is immature, and the development of the face detection technology is slow in this period.
With the development of society, especially with the rapid development of computer technology and the massive application of biometric identification in the last two decades, face detection technology has been used for not only face identification. The application range of the face detection technology is developed from the original pure scientific research to the aspects of business, military, security and the like at present, and the attention degree of the technology also reaches a new level. In which the convolutional neural network CNN plays an irreplaceable role.
Convolutional neural is a feedforward network, and its artificial neuron can cover units in the range, including convolutional layer and pooling layer. At present, the convolutional neural network CNN has been widely applied to the computer vision domain, and has achieved good effect. In recent years, the performance of CNN in ImageNet competition, it can be seen that in pursuit of classification accuracy, the depth of the model is deeper and deeper, and the complexity of the model is higher and higher, such as a depth residual error network, the number of layers is up to 152.
However, in some real application scenarios, such as mobile or embedded devices, such large and complex models are difficult to apply. Firstly, the models are too large and face the problem of insufficient memory, and secondly, the scenes require low delay or high response speed, for example, the pedestrian detection system of an automatic driving automobile has high requirement on the response speed. Therefore, it is very important to construct a small-scale and efficient CNN model.
Disclosure of Invention
The purpose of the invention is as follows: the invention aims to provide a face key point detection method, which can obtain higher detection precision on a deep learning network with smaller scale.
The technical scheme is as follows: the invention discloses a face key point detection method based on deep learning, which comprises a training stage and a detection stage, wherein the training stage comprises the following steps:
(1) constructing a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the key points of the human face are coordinates of the left eye and the nose tip;
the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a common convolution layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-connection layer, and the sixteenth layer is a softmax layer;
(2) training the face Key point Detection network constructed in the step (1) by taking a face data set of Facial Key-points Detection on Kaggle as a sample image to obtain a trained face Key point Detection network;
the detection stage comprises the following steps:
(3) collecting an image of a face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier;
(4) inputting the face region acquired in the step (3) into a trained face key point detection network to obtain coordinates of a left eye and a nose tip;
(5) calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points;
wherein (X)left_eye,Yleft_eye),(Xnoise,Ynoise),(Xright_eye,Yright_eye),(Xleft_mouth,Yleft_mouth),(Xright_mouth,Yright_mouth) The coordinates of the left eye, the nose tip, the right eye, the left mouth corner and the right mouth corner are respectively.
Further, the step (2) further comprises preprocessing the training sample, including:
(2.1) eliminating training samples with missing data;
(2.2) standardizing the training samples: normalized gray value of pixel i in sample image AComprises the following steps:
wherein g isiFor the gray value of pixel i in sample image A before normalization, gminIs the minimum value of the pixel gray levels in all sample images, gmaxIs the maximum value of the pixel gray levels in all sample images.
In order to reduce the dependence of the network on initialization, in the first layer to the thirteenth layer of the face key point detection network, the input of each layer is subjected to batch standardization through Batchnormalization and then convolution operation is performed.
In order to prevent overfitting, an early-stop method is adopted for training in the step (2), the sample images are divided into a training set and a verification set, a face key point detection network is trained on the training set, and errors of the face key point detection network are verified by the verification set;
the training steps are as follows:
(3.1) initializing the error of the optimal verification set to be the error of the verification set after the first training is finished, and initializing the error training times of the optimal verification set to be 0;
(3.2) after each training is finished, verifying errors of the face key point detection network by using a verification set, and comparing the current verification set errors with the optimal verification set errors; if the error of the current verification set is smaller than the error of the optimal verification set, updating the error of the optimal verification set into the error of the current verification set, and setting the error training times of the optimal verification set to be 0; if the error of the current verification set is larger than the error of the optimal verification set, adding one to the error training times of the optimal verification set;
and (3.3) stopping training if the error training times of the optimal verification set reach a preset error training time threshold of the optimal verification set or the total training times reach a preset total training time threshold, and taking the face key point detection network parameters corresponding to the errors of the optimal verification set as training results.
On the other hand, the invention discloses a detection system for realizing the human face key point detection method, which comprises an image acquisition module, a human face detection module and a human face key point detection and labeling module;
the image acquisition module is used for acquiring an image of a face to be detected;
the face detection module is used for acquiring a face area in an image of a face to be detected;
the face key point detection and labeling module is used for acquiring and labeling coordinates of a left eye, a nose tip, a right eye, a left mouth corner and a right mouth corner in a face region.
The face key point detection and labeling module is a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
The face key point detection and labeling module is a computer provided with a NVIDIAGTX 1080 GPU.
Has the advantages that: compared with the prior art, the face key point detection method based on deep learning disclosed by the invention has the following advantages: 1. compared with the common convolutional neural network, the constructed human face key point detection network based on the MobileNet V1 framework has fewer model parameters under the condition of the same detection effect; meanwhile, only 2 key points of the face are detected, and the training and detecting speeds are further increased; 2. on the basis of the detected 2 key points, other feature point coordinates can be obtained.
Drawings
FIG. 1 is a flow chart of a face key point detection method disclosed by the present invention;
FIG. 2 is a schematic structural diagram of a face key point detection network constructed by the present invention;
FIG. 3 is a schematic diagram of the operation of depth convolution and point-by-point convolution;
fig. 4 is a schematic structural diagram of a face key point detection system disclosed in the present invention.
Detailed Description
The invention is further elucidated with reference to the drawings and the detailed description.
As shown in fig. 1, the invention discloses a method for detecting key points of a human face based on deep learning, which comprises a training stage and a detection stage, wherein the training stage comprises the following steps:
as shown in fig. 2, the face key point detection network includes 16 layers of networks connected in sequence, where the first layer is a common convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
The 1-13 layers are feature extraction sub-networks and are used for extracting local features of the image; and the 14-16 layers are feature classification sub-networks, and face key points are obtained according to local features.
The depth convolution is calculated as:
the method comprises the following steps: splitting an image of size M x N into N images of size M x 1;
step two: defining 1 convolution kernel of F x F1 (F is usually an odd number) for each single-channel image, a total of N need to be defined;
step three: in order to make the convolved output image equal in size to the input image, padding is performed, and the size of the padding for the upper, lower, left, and right sides is (F-1)/2.
Step four: convolving each pixel point in each single-channel image with the convolution kernel of the corresponding single channel, and assuming that F is 2a +1, since the size of the convolution kernel is usually odd, the calculation formula is as follows:
where w (i, j) is the coefficient in the convolution kernel of F x F1, F (x, y) is any point pixel in the image, and g (x, y) is the convolution output of F (x, y), where x, y are variable so that each pixel in w can access each pixel in F.
The calculation of the point-by-point convolution is:
and adding the results of the N convolution operations by using 1 × N convolution to obtain an output of M × 1.
The process of depth separable convolution of a three channel image using 3 x 3 convolution kernel is illustrated in fig. 3. The three-channel image is changed into three single-channel images, and then convolution operation is carried out on each single channel. At any point f (x, y) in a single channel image, the output g (x, y) is the sum of the products of the convolution kernel coefficients and the image pixels surrounded by the convolution kernel:
g(x,y)=w(-1,-1)f(x-1,y-1)+w(-1,0)f(x-1,y)+...........+w(1,1)f(x+1,y+1)
the three single-channel image outputs g (x, y) are then summed by a 1 x 3 convolution kernel to give the final output z (x, y).
In order to reduce the dependence of the network on initialization, in the first layer to the thirteenth layer of the face key point detection network, the input of each layer is subjected to batch standardization through Batchnormalization and then convolution operation is performed. And simultaneously, using a ReLU linear activation function on each convolutional layer to output convolutional layer results.
Step 2, taking a Facial data set of Facial Key-points Detection on Kaggle as a sample image, and training the Facial Key point Detection network constructed in the step 1 to obtain a trained Facial Key point Detection network;
firstly, training samples are preprocessed, and the preprocessing comprises the following steps:
(2.1) eliminating training samples with missing data;
(2.2) standardizing the training samples: normalized gray value of pixel i in sample image AComprises the following steps:
wherein g isiFor the gray value of pixel i in sample image A before normalization, gminIs the minimum value of the pixel gray levels in all sample images, gmaxIs the maximum value of the pixel gray levels in all sample images.
The method adopts a random gradient descent algorithm (Adam) minimum loss function of adaptive matrix estimation to train the face key point detection network, and the loss function adopts a mean square error loss function.
In order to prevent overfitting, an early-stop method is adopted for training in the step (2), the sample images are divided into a training set and a verification set, a face key point detection network is trained on the training set, and errors of the face key point detection network are verified by the verification set;
the training steps are as follows:
(3.1) initializing the error of the optimal verification set to be the error of the verification set after the first training is finished, and initializing the error training times of the optimal verification set to be 0;
(3.2) after each training is finished, verifying errors of the face key point detection network by using a verification set, and comparing the current verification set errors with the optimal verification set errors; if the error of the current verification set is smaller than the error of the optimal verification set, updating the error of the optimal verification set into the error of the current verification set, and setting the error training times of the optimal verification set to be 0; if the error of the current verification set is larger than the error of the optimal verification set, adding one to the error training times of the optimal verification set;
and (3.3) stopping training if the error training times of the optimal verification set reach a preset error training time threshold of the optimal verification set or the total training times reach a preset total training time threshold, and taking the face key point detection network parameters corresponding to the errors of the optimal verification set as training results.
The detection stage comprises the following steps:
step 3, collecting the image of the face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier;
in the present invention, the face area resize to be acquired is 96 × 96 × 1.
Step 4, inputting the face region obtained in the step 3 into a trained face key point detection network to obtain coordinates of a left eye and a nose tip;
step 5, calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points;
the longitudinal distance from the left eye to the nose tip is found to be equal to the longitudinal distance from the left mouth corner to the nose tip through statistics, and the abscissa of the left mouth corner is equal to the abscissa of the left eye; the longitudinal distance from the right eye to the tip of the nose and the longitudinal distance from the right mouth corner to the tip of the nose are equal and the abscissa of the right mouth corner and the right eye is equal. Obtaining the coordinates of other key points on the face according to the corresponding relation, wherein the calculation formula is as follows:
wherein (X)left_eye,Yleft_eye),(Xnoise,Ynoise),(Xright_eye,Yright_eye),(Xleft_mouth,Yleft_mouth),(Xright_mouth,Yright_mouth) The coordinates of the left eye, the nose tip, the right eye, the left mouth corner and the right mouth corner are respectively.
As shown in fig. 4, a detection system for implementing the face key point detection method includes an image acquisition module, a face detection module, and a face key point detection and labeling module;
the image acquisition module is used for acquiring an image of a face to be detected; in the invention, a monocular camera is adopted to acquire an image of a face to be detected, and the acquired image is preprocessed and converted into a gray image;
the face detection module is used for acquiring a face area in an image of a face to be detected and framing the face area by using a rectangular frame;
the face key point detection and labeling module is used for acquiring and labeling coordinates of a left eye, a nose tip, a right eye, a left mouth corner and a right mouth corner in a face region.
The face key point detection and labeling module is a face key point detection network based on a MobileNet V1 framework; the input of the human face key point detection network is a human face image, and the output is coordinates of a left eye and a nose tip; the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
The coordinates of the right eye, the left mouth angle and the right mouth angle are calculated according to the formula (1).
The face key point detection and labeling module is a computer provided with a NVIDIAGTX 1080 GPU.
Claims (7)
1. The method for detecting the key points of the human face based on deep learning is characterized by comprising a training stage and a detection stage, wherein the training stage comprises the following steps:
(1) constructing a face key point detection network based on a MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the key points of the human face are coordinates of the left eye and the nose tip;
the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a common convolution layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-connection layer, and the sixteenth layer is a softmax layer;
(2) training the face Key point Detection network constructed in the step (1) by taking a face data set of Facial Key-points Detection on Kaggle as a sample image to obtain a trained face Key point Detection network;
the detection stage comprises the following steps:
(3) collecting an image of a face to be detected; acquiring a face region of an acquired image to be detected by adopting an OpenCV (open circuit vehicle vision) cascade classifier;
(4) inputting the face region acquired in the step (3) into a trained face key point detection network to obtain coordinates of a left eye and a nose tip;
(5) calculating coordinates of the right eye and the left and right mouth angles of the human face in the image to be detected, and labeling key points;
wherein (X)left_eye,Yleft_eye),(Xnoise,Ynoise),(Xright_eye,Yright_eye),(Xleft_mouth,Yleft_mouth),(Xright_mouth,Yright_mouth) The coordinates of the left eye, the nose tip, the right eye, the left mouth corner and the right mouth corner are respectively.
2. The method for detecting key points of a human face based on deep learning of claim 1, wherein the step (2) further comprises preprocessing training samples, comprising:
(2.1) eliminating training samples with missing data;
(2.2) standardizing the training samples: pixel i criteria in sample image ANormalized gray scale valueComprises the following steps:
wherein g isiFor the gray value of pixel i in sample image A before normalization, gminIs the minimum value of the pixel gray levels in all sample images, gmaxIs the maximum value of the pixel gray levels in all sample images.
3. The method according to claim 1, wherein the input of each layer is standardized in batch by BatchNormal and then convolution is performed in the first to thirteenth layers of the face key point detection network.
4. The method for detecting key points of a human face based on deep learning of claim 1, wherein in the step (2), the sample images are divided into a training set and a verification set, a human face key point detection network is trained on the training set, and the verification set is used for verifying the error of the human face key point detection network; the training steps are as follows:
(3.1) initializing the error of the optimal verification set to be the error of the verification set after the first training is finished, and initializing the error training times of the optimal verification set to be 0;
(3.2) after each training is finished, verifying errors of the face key point detection network by using a verification set, and comparing the current verification set errors with the optimal verification set errors; if the error of the current verification set is smaller than the error of the optimal verification set, updating the error of the optimal verification set into the error of the current verification set, and setting the error training times of the optimal verification set to be 0; if the error of the current verification set is larger than the error of the optimal verification set, adding one to the error training times of the optimal verification set;
and (3.3) stopping training if the error training times of the optimal verification set reach a preset error training time threshold of the optimal verification set or the total training times reach a preset total training time threshold, and taking the face key point detection network parameters corresponding to the errors of the optimal verification set as training results.
And when the error of the face key point detection network on the verification set is larger than that of the last training result, stopping training, and adopting the parameters of the last training result as the final parameters of the face key point detection network.
5. The human face key point detection system based on deep learning is characterized by comprising an image acquisition module, a human face detection module and a human face key point detection and labeling module;
the image acquisition module is used for acquiring an image of a face to be detected;
the face detection module is used for acquiring a face area in an image of a face to be detected;
the face key point detection and labeling module is used for acquiring and labeling coordinates of a left eye, a nose tip, a right eye, a left mouth corner and a right mouth corner in a face region.
6. The deep learning based face keypoint detection system of claim 5, wherein the face keypoint detection and labeling module is a face keypoint detection network based on the MobileNet V1 architecture; the input of the human face key point detection network is a human face image, and the output is a human face key point coordinate; the face key point detection network comprises 16 layers of networks which are connected in sequence, wherein the first layer is a convolutional layer; the second layer to the tenth layer are 6 depth separable convolution units in which depth convolution layers and point-by-point convolution layers are alternately connected; the fourteenth layer is a pooling layer, the fifteenth layer is a full-link layer, and the sixteenth layer is a softmax layer.
7. The deep learning based face keypoint detection system of claim 6, wherein the face keypoint detection and labeling module is a computer equipped with a NVIDIAGTX 1080 GPU.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910986156.4A CN110956082B (en) | 2019-10-17 | 2019-10-17 | Face key point detection method and detection system based on deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910986156.4A CN110956082B (en) | 2019-10-17 | 2019-10-17 | Face key point detection method and detection system based on deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110956082A true CN110956082A (en) | 2020-04-03 |
CN110956082B CN110956082B (en) | 2023-03-24 |
Family
ID=69975622
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910986156.4A Active CN110956082B (en) | 2019-10-17 | 2019-10-17 | Face key point detection method and detection system based on deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110956082B (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112200008A (en) * | 2020-09-15 | 2021-01-08 | 青岛邃智信息科技有限公司 | Face attribute recognition method in community monitoring scene |
CN112597888A (en) * | 2020-12-22 | 2021-04-02 | 西北工业大学 | On-line education scene student attention recognition method aiming at CPU operation optimization |
CN112883854A (en) * | 2021-02-04 | 2021-06-01 | 新绎健康科技有限公司 | Facial region positioning method and system based on deep learning |
CN113011356A (en) * | 2021-03-26 | 2021-06-22 | 杭州朗和科技有限公司 | Face feature detection method, device, medium and electronic equipment |
CN113822256A (en) * | 2021-11-24 | 2021-12-21 | 北京的卢深视科技有限公司 | Face recognition method, electronic device and storage medium |
CN114881893A (en) * | 2022-07-05 | 2022-08-09 | 腾讯科技(深圳)有限公司 | Image processing method, device, equipment and computer readable storage medium |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105678248A (en) * | 2015-12-31 | 2016-06-15 | 上海科技大学 | Face key point alignment algorithm based on deep learning |
WO2017031089A1 (en) * | 2015-08-15 | 2017-02-23 | Eyefluence, Inc. | Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects |
CN108229442A (en) * | 2018-02-07 | 2018-06-29 | 西南科技大学 | Face fast and stable detection method in image sequence based on MS-KCF |
CN108898112A (en) * | 2018-07-03 | 2018-11-27 | 东北大学 | A kind of near-infrared human face in-vivo detection method and system |
CN109344693A (en) * | 2018-08-13 | 2019-02-15 | 华南理工大学 | A kind of face multizone fusion expression recognition method based on deep learning |
CN110119702A (en) * | 2019-04-30 | 2019-08-13 | 西安理工大学 | Facial expression recognizing method based on deep learning priori |
CN110263691A (en) * | 2019-06-12 | 2019-09-20 | 合肥中科奔巴科技有限公司 | Head movement detection method based on android system |
-
2019
- 2019-10-17 CN CN201910986156.4A patent/CN110956082B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017031089A1 (en) * | 2015-08-15 | 2017-02-23 | Eyefluence, Inc. | Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects |
EP3335096A1 (en) * | 2015-08-15 | 2018-06-20 | Google LLC | Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects |
CN105678248A (en) * | 2015-12-31 | 2016-06-15 | 上海科技大学 | Face key point alignment algorithm based on deep learning |
CN108229442A (en) * | 2018-02-07 | 2018-06-29 | 西南科技大学 | Face fast and stable detection method in image sequence based on MS-KCF |
CN108898112A (en) * | 2018-07-03 | 2018-11-27 | 东北大学 | A kind of near-infrared human face in-vivo detection method and system |
CN109344693A (en) * | 2018-08-13 | 2019-02-15 | 华南理工大学 | A kind of face multizone fusion expression recognition method based on deep learning |
CN110119702A (en) * | 2019-04-30 | 2019-08-13 | 西安理工大学 | Facial expression recognizing method based on deep learning priori |
CN110263691A (en) * | 2019-06-12 | 2019-09-20 | 合肥中科奔巴科技有限公司 | Head movement detection method based on android system |
Non-Patent Citations (3)
Title |
---|
YI SUN ET AL.: ""Deep Convolutional Network Cascade for Facial Point Detection"", 《2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 * |
姜开永等: "基于深度学习的人脸美丽预测模型及其应用", 《五邑大学学报(自然科学版)》 * |
曾家建: ""基于深度学习的人脸美丽预测模型及其应用"", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》 * |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112200008A (en) * | 2020-09-15 | 2021-01-08 | 青岛邃智信息科技有限公司 | Face attribute recognition method in community monitoring scene |
CN112597888A (en) * | 2020-12-22 | 2021-04-02 | 西北工业大学 | On-line education scene student attention recognition method aiming at CPU operation optimization |
CN112597888B (en) * | 2020-12-22 | 2024-03-08 | 西北工业大学 | Online education scene student attention recognition method aiming at CPU operation optimization |
CN112883854A (en) * | 2021-02-04 | 2021-06-01 | 新绎健康科技有限公司 | Facial region positioning method and system based on deep learning |
CN113011356A (en) * | 2021-03-26 | 2021-06-22 | 杭州朗和科技有限公司 | Face feature detection method, device, medium and electronic equipment |
CN113822256A (en) * | 2021-11-24 | 2021-12-21 | 北京的卢深视科技有限公司 | Face recognition method, electronic device and storage medium |
CN113822256B (en) * | 2021-11-24 | 2022-03-25 | 北京的卢深视科技有限公司 | Face recognition method, electronic device and storage medium |
CN114881893A (en) * | 2022-07-05 | 2022-08-09 | 腾讯科技(深圳)有限公司 | Image processing method, device, equipment and computer readable storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN110956082B (en) | 2023-03-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110956082B (en) | Face key point detection method and detection system based on deep learning | |
CN109522818B (en) | Expression recognition method and device, terminal equipment and storage medium | |
Rao et al. | Selfie video based continuous Indian sign language recognition system | |
CN106599883B (en) | CNN-based multilayer image semantic face recognition method | |
CN110348376B (en) | Pedestrian real-time detection method based on neural network | |
CN111160269A (en) | Face key point detection method and device | |
CN108009493B (en) | Human face anti-cheating recognition method based on motion enhancement | |
CN108182397B (en) | Multi-pose multi-scale human face verification method | |
CN110570443B (en) | Image linear target extraction method based on structural constraint condition generation model | |
CN111639580B (en) | Gait recognition method combining feature separation model and visual angle conversion model | |
CN113011253B (en) | Facial expression recognition method, device, equipment and storage medium based on ResNeXt network | |
CN106611156B (en) | Pedestrian identification method and system based on self-adaptive depth space characteristics | |
Huang et al. | Human emotion recognition based on face and facial expression detection using deep belief network under complicated backgrounds | |
CN111126250A (en) | Pedestrian re-identification method and device based on PTGAN | |
CN113763417B (en) | Target tracking method based on twin network and residual error structure | |
CN110826534A (en) | Face key point detection method and system based on local principal component analysis | |
CN103235943A (en) | Principal component analysis-based (PCA-based) three-dimensional (3D) face recognition system | |
CN116246338B (en) | Behavior recognition method based on graph convolution and transducer composite neural network | |
KR20180092453A (en) | Face recognition method Using convolutional neural network and stereo image | |
CN112613430B (en) | Gait recognition method based on deep migration learning | |
CN108830166B (en) | Real-time bus passenger flow volume statistical method | |
CN114743257A (en) | Method for detecting and identifying image target behaviors | |
CN114038011A (en) | Method for detecting abnormal behaviors of human body in indoor scene | |
Monica et al. | Recognition of medicine using cnn for visually impaired | |
WO2021056531A1 (en) | Face gender recognition method, face gender classifier training method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |