WO2020187160A1 - Procédé et système de reconnaissance faciale basés sur un réseau neuronal à convolution profonde en cascade - Google Patents

Procédé et système de reconnaissance faciale basés sur un réseau neuronal à convolution profonde en cascade Download PDF

Info

Publication number
WO2020187160A1
WO2020187160A1 PCT/CN2020/079281 CN2020079281W WO2020187160A1 WO 2020187160 A1 WO2020187160 A1 WO 2020187160A1 CN 2020079281 W CN2020079281 W CN 2020079281W WO 2020187160 A1 WO2020187160 A1 WO 2020187160A1
Authority
WO
WIPO (PCT)
Prior art keywords
network
face
face recognition
output
convolutional neural
Prior art date
Application number
PCT/CN2020/079281
Other languages
English (en)
Chinese (zh)
Inventor
翟新刚
张楠赓
Original Assignee
北京嘉楠捷思信息技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京嘉楠捷思信息技术有限公司 filed Critical 北京嘉楠捷思信息技术有限公司
Publication of WO2020187160A1 publication Critical patent/WO2020187160A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation

Definitions

  • the field of artificial intelligence technology of the present invention particularly relates to a face recognition method and system based on a cascaded deep convolutional neural network.
  • Face recognition technology is a kind of biometric recognition technology based on the facial feature information of people.
  • the face recognition process is mainly to collect video streams with a camera, automatically detect and track faces in the images, and then perform face recognition on the detected faces.
  • face recognition systems have been widely used in various fields, such as community access control, company attendance, judicial and criminal investigations, etc.
  • face recognition is a kind of multi-task combination work in deep learning, but existing technical solutions often focus on the realization of a certain task, and ignore the relationship between multiple tasks.
  • the main purpose of the present invention is to provide a face recognition method and system based on a cascaded deep convolutional neural network to solve at least one of the above-mentioned problems.
  • a face recognition method based on a cascaded deep convolutional neural network including:
  • the extracting facial features using a cascaded deep convolutional neural network includes:
  • the output of the second network is sent to the third network to extract facial features.
  • the sending the output of the first network to the second network to predict the location of key points on the face includes:
  • the sending the output of the second network to the third network to extract facial features includes:
  • the first network is a face detection network (Face Detection Network, referred to as FDNet)
  • the second network is a key-point detection network (Key-point Detection Network, referred to as KDNet)
  • the third network is a feature Extraction network (Feature Extraction Network, FENet for short).
  • the method before extracting facial features by using the cascaded deep convolutional neural network, the method further includes: collecting facial image data.
  • a face recognition system based on a cascaded deep convolutional neural network including:
  • a feature extraction module for extracting facial features using cascaded deep convolutional neural networks
  • the face recognition module is connected to the feature extraction module and is used to perform face recognition based on the extracted facial features.
  • the feature extraction module includes:
  • the first network is used to receive face image data and predict the regression of the face frame
  • the frame interception unit is used to receive the output of the first network and perform frame interception and size conversion operations
  • the second network is used to receive the output of the frame interception unit and predict the position of key points on the face;
  • the similarity transformation unit is used to receive the output of the second network and perform similarity transformation, mapping and size transformation operations;
  • the third network is used to receive the output of the similarity transformation unit and extract facial features.
  • the first network is a face detection network (Face Detection Network, referred to as FDNet)
  • the second network is a key-point detection network (Key-point Detection Network, referred to as KDNet)
  • the third network is a feature Extraction network (Feature Extraction Network, FENet for short).
  • it further includes a collection module for collecting face image data.
  • the present invention uses a cascaded deep convolutional neural network for feature extraction, and performs face recognition based on the extracted features.
  • Each level of the cascaded deep convolutional neural network only needs to be executed for everyone Once, the control is simple, the amount of calculation is small, and it is easy to accelerate; and extracting facial features through deep learning for face recognition can easily cope with face recognition tasks of various security levels.
  • the present invention uses similar transformations when performing face recognition based on cascaded deep convolutional neural networks, which further reduces the background effect caused by different frame sizes and reduces the demand for the network.
  • the present invention maps face recognition to multiple different deep learning models independently according to different tasks, which has strong replaceability, avoids waste of computing power, and facilitates intuitive determination of the need to upgrade the network part.
  • Fig. 1 is a schematic flow chart of the face recognition method of the present invention.
  • Fig. 2 is a schematic diagram of a frame cut in the face recognition method shown in Fig. 1.
  • FIG. 3 is a flowchart of the face recognition method of the present invention.
  • Fig. 4 is another flowchart of the face recognition method of the present invention.
  • Figure 5 is a flow chart of the inventors extracting facial features.
  • Fig. 6 is a flow chart of predicting the position of key points on a face according to the present invention.
  • Fig. 7 is another flow chart of extracting facial features according to the present invention.
  • Figure 8 is a schematic diagram of the structure of the face recognition system of the present invention.
  • Fig. 9 is a schematic diagram of another structure of the face recognition system of the present invention.
  • Figure 10 is a schematic diagram of the structure of the feature extraction module of the present invention.
  • FIG. 11 is another flowchart of the face recognition method according to the embodiment of the present invention.
  • Face recognition usually includes face detection, face feature extraction, and classification of the extracted face features to complete face recognition.
  • face detection is to find out whether there are one or more faces in a given picture, and return the position and range of each face in the picture. Face detection algorithms are divided into four categories: knowledge-based, feature-based, template matching-based, and appearance-based methods. With the use of DPM (Direct Part Model) algorithm (variable component model) and deep learning Convolutional Neural Networks (CNN), all face detection algorithms can be divided into two categories: (1) Based on Template matching (Based on rigid templates): Among them, there are algorithms (Boosting) + features (Features) and CNN; (2) Based on parts model (Based on parts model).
  • DPM Direct Part Model
  • CNN deep learning Convolutional Neural Networks
  • Facial feature extraction is a process of obtaining facial feature information in the area where the face is located on the basis of face detection.
  • Face feature extraction methods include: Eigenface (Eigenface), Principal Component Analysis (Principal Component Analysis, referred to as PAC).
  • PAC Principal Component Analysis
  • Classification refers to classifying according to type, level or nature, and classifying the extracted features to complete face recognition.
  • Classification methods mainly include: decision tree method, Bayesian method, and artificial neural network.
  • the process of the face recognition method of the present invention is: Pyramid scale transformation is performed on a new picture, and the transformed picture is input into a network to generate a large number of face classification scores.
  • Regression vector with the face rectangle also called box, border, bounding box, window, window, etc.
  • weed out the face rectangle with lower score for example, lower than a threshold M1
  • replace the remaining face rectangle The frame performs non-maximum suppression to obtain the final prediction result; then the predicted result is input into another network, and the face rectangle with a lower score (for example, lower than the threshold M2) is also eliminated, and then non-maximum suppression is used
  • the algorithm filters out the large overlapping face rectangles, displays the key points of the face, and performs feature extraction and face recognition.
  • the face recognition method is introduced by taking the face network (FaceNet) as an example. As shown in Figure 1-2, the face recognition method includes the following steps:
  • FaceNet extracts facial features in two steps:
  • MTCNN Multi-task Cascaded Convolutional Networks
  • the MTCNN predicts the Bounding Box of the face, as shown in Figure 1, including the following sub-steps:
  • the input original image is scaled to various sizes, that is, the original image is subjected to different Scale Resize operations to build an image pyramid, and each layer of the pyramid is sent to the shallow CNN candidate box network (Proposal Network, referred to as PNet) And perform Bounding Box Regression and Non-maximum Suppression (NMS) to quickly generate candidate forms;
  • PNet Proposal Network
  • NMS Non-maximum Suppression
  • a more powerful CNN output network (Output network, referred to as ONet) is used to realize the selection of candidate forms and display the location of five facial key points at the same time.
  • the above method uses MTCNN to predict Bounding Box that requires repeated PNet and RNet multiple times, and the control is relatively complicated and the amount of calculation is large.
  • the Bounding Box predicted by MTCNN is added with a fixed-length Margin and sent to the feature extraction network. Since the Bounding Box of the face in the figure will have various sizes, if a fixed Margin is added to the face of different size, the size will be different. The background information of the face will be very different, so it will weaken the generalization ability of the feature extraction network.
  • the present invention also provides a method for realizing sensorless face recognition based on a cascaded deep convolutional neural network.
  • the face recognition method based on the cascaded deep convolutional neural network of the present invention includes the following steps:
  • S2 Perform face recognition according to the extracted facial features.
  • the present invention uses cascaded deep convolutional neural networks for feature extraction, and performs face recognition based on the extracted features.
  • Each level of network in the cascaded deep convolutional neural network only needs to be executed once for each person, and control Simple, small amount of calculation, easy to accelerate.
  • the face recognition method may further include: S0, collecting facial image data.
  • the extraction of facial features using the cascaded deep convolutional neural network includes:
  • S11 Send the face image data to the first network to predict the regression of the face frame
  • S12 Send the output of the first network to the second network to predict the location of key points on the face;
  • S13 Send the output of the second network to the third network to extract facial features.
  • the cascaded deep convolutional neural network may include three networks, and the three networks form a three-level cascaded deep convolutional neural network; wherein, the first network is a face detection network (Face Detection Network). Network, referred to as FDNet), the second network is a Key-point Detection Network (Key-point Detection Network, referred to as KDNet), and the third network is a Feature Extraction Network (Feature Extraction Network, referred to as FENet).
  • FDNet face detection network
  • KDNet Key-point Detection Network
  • FENet Feature Extraction Network
  • the present invention disassembles the face recognition task and fully considers the relationship between multiple tasks. Specifically, the present invention disassembles the face recognition task
  • the convolutional neural networks provided by deep learning corresponding to the three-level tasks are FDNet, KDNet and FENet.
  • face recognition is independently mapped to multiple different deep learning models according to different tasks, which is highly replaceable, avoids waste of computing power, and facilitates intuitive determination of the need to upgrade the network part.
  • the sending the output of the first network to the second network to predict the location of key points on the face includes:
  • S121 Perform frame interception and size conversion operations on the output of the first network before sending it to the second network;
  • S122 Use the second network to predict the location of key points on the face.
  • the sending the output of the second network to the third network to extract facial features includes:
  • the present invention uses similar transformations to further reduce the background effect caused by different frame sizes, reduce the demand for FDNet, and improve the accuracy of feature extraction .
  • the present invention also provides a face recognition system based on a cascaded deep convolutional neural network.
  • the face recognition system based on a cascaded deep convolutional neural network includes:
  • the feature extraction module 11 is used for extracting facial features using cascaded deep convolutional neural networks.
  • the face recognition module 12 is connected to the feature extraction module 11 and is used to perform face recognition according to the extracted facial features.
  • the face recognition system may further include a collection module 10 for collecting face image data.
  • the feature extraction module 11 is connected to the collection module 10, and is configured to receive face image data sent by the collection module 10, and extract facial features by using a cascaded deep convolutional neural network.
  • the feature extraction module includes:
  • the first network 110 is configured to receive the face image data and predict the regression of the face frame
  • the frame interception unit 111 is configured to receive the output of the first network 110 and perform frame interception and size conversion operations;
  • the second network 112 is configured to receive the output of the frame interception unit 111 and predict the position of key points on the face;
  • the similarity transformation unit 113 is configured to receive the output of the second network 112 and perform similarity transformation, mapping and size transformation operations;
  • the third network 114 is configured to receive the output of the similarity transformation unit 113 and extract facial features.
  • the first network is a face detection network (Face Detection Network, referred to as FDNet)
  • the second network is a key-point detection network (Key-point Detection Network, referred to as KDNet)
  • the third network is a feature extraction network (Feature Extraction Network, referred to as FENet).
  • the face recognition method based on the cascaded deep convolutional neural network specifically includes:
  • Step 1 FDNet is based on YOLO's design idea, using MobileNet as the backbone, directly Bounding Box Regression on the face, and predicting the confidence at the same time; in other words, the task of FDNet is to find The position of all faces in the image, and intercept all face images as the input of KDNet in turn.
  • FDNet includes but is not limited to MobileNet-YOLO.
  • the face is marked in the data set, and the golden truth of the face frame is set as follows:
  • Location is set to among them, Represents the coordinates of the upper left corner of the i-th face frame, Respectively represent the width and height of the face frame;
  • the corresponding YOLO prediction result is x i y i w i h i C i p i ;
  • the MobileNet-YOLO network in the stable state can be used for the prediction of the face frame.
  • Step 2 Based on the output of FDNet, cut out the bounding box, transform the size (Resize) to a fixed size, and send it to KDNet (Keypoints Detection Net) to directly predict the positions of five facial key points; that is, KDNet's task
  • KDNet Keypoints Detection Net
  • this embodiment takes five points as an example (left eye corner, right eye corner, nose, left mouth corner, right mouth corner), traverse all the inputs of FDNet, you can get everyone The key points of the face.
  • KPNet can choose a variety of networks, and only needs to predict the location of key points on the face.
  • the following takes the left eye, right eye, nose, left mouth corner, and right mouth corner of the face as examples to introduce the process of marking the key points of the face in the data set:
  • the face data is preprocessed by the FDNet in the aforementioned step 1, and the face part is cut out by the face frame, and the position of the left eye, right eye, nose, left mouth corner, and right mouth corner of the face is calculated in the face part.
  • Position ratio get five pairs of position coordinates (x, y), x ⁇ [0,1], y ⁇ [0,1]
  • the Sigmoid function is as follows:
  • Step 3 Based on the output of five facial key points in KDNet, perform five-point similarity transformation on the entire frame of image, map it to five points at a fixed golden position, and transform the mapped face image to a fixed size (Resize)
  • the size is sent to FENet (Feature Extraction Net, referred to as FENet) to extract facial features; that is to say, a similar transformation is required between KDNet and FENet as a bridge, and the key point position information obtained by KDNet is compared with the position of golden (Golden).
  • the key point position information obtains the similarity transformation matrix, and then the entire frame image is similarly transformed corresponding to the key points to obtain the similarly transformed face image.
  • the task of FENet is to abstract the face image information into a feature vector representation. This feature Vector representation has the following characteristics: all faces of the same subject can be mapped to similar feature vectors, while the feature vectors obtained by all faces of different subjects are quite different.
  • This embodiment uses a five-point similarity transformation, of course, it is not limited to the five-point similarity transformation.
  • first set the golden positions of the left eye, right eye, nose, left mouth corner, and right mouth corner of the face to resemble the transformed image Take 112*112 as an example, the Golden position can be selected as follows
  • Nose_g [56.0252, 71.7366]
  • Golden key point Landmark_Golden [le_g, re_g, nose_g, l_mouth_g, r_mouth_g]
  • Landmark_get [le, re, nose, l_mouth, r_mouth]
  • Landmark_Get Landmark_Golden can perform similar transformations, as follows:
  • the FENet includes but is not limited to Mobilefacenets, and AM-softmax Loss can be selected as the loss function.
  • the face recognition method based on the cascaded deep convolutional neural network described in this embodiment uses a cascaded deep convolutional neural network (Cascaded-Deep CNN, referred to as CDCNN) to extract facial features and perform human Face recognition.
  • CDCNN cascaded deep convolutional neural network
  • Each level of CDCNN only needs to be executed once for everyone, which is simple to control, small in calculation, and easy to accelerate; and compared to the aforementioned Bounding Box method of adding Margin, it uses a five-point similar transformation to further reduce
  • the background effect caused by the different sizes of the Bounding Box reduces the demand for FDNet (as long as the five key points of the face are accurate, the face detection box does not have to be generated by the MTCNN network).
  • the face recognition method and system based on the cascaded deep convolutional neural network of the present invention may also include other parts, which are not related to the innovation of the present invention, so they will not be repeated here.
  • modules or units or components in the embodiments can be combined into one module or unit or component, and in addition, they can be divided into multiple sub-modules or sub-units or sub-components. Except that at least some of such features and/or processes or units are mutually exclusive, any combination can be used to compare all the features of the invention in this specification (including the accompanying claims, abstract and drawings) and any method or method of such invention. All the processes or units of the equipment are combined. Unless expressly stated otherwise, each feature of the invention in this specification (including the accompanying claims, abstract and drawings) may be replaced by an alternative feature providing the same, equivalent or similar purpose.
  • the various component embodiments of the present invention may be implemented by hardware, or by software modules running on one or more processors, or by their combination.
  • a microprocessor or a digital signal processor (DSP) may be used in practice to implement some or all of the functions of some or all of the components in the relevant device according to the embodiments of the present invention.
  • DSP digital signal processor
  • the present invention can also be implemented as a device or device program (for example, a computer program and a computer program product) for executing part or all of the methods described herein.
  • Such a program for realizing the present invention may be stored on a computer-readable medium, or may have the form of one or more signals. Such signals can be downloaded from Internet websites, or provided on carrier signals, or provided in any other form.
  • ordinal numbers used in the specification and claims such as “first”, “second”, “third”, etc., are used to modify the corresponding elements, and they do not imply or represent that the elements have any
  • the ordinal number does not represent the order of a certain element and another element, or the order in the manufacturing method. The use of these ordinal numbers is only used to make it clear that one element with a certain name can be made clear from another element with the same name distinguish.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Human Computer Interaction (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

L'invention concerne un procédé et un système de reconnaissance faciale basés sur un réseau neuronal à convolution profonde en cascade. Le procédé de reconnaissance faciale basé sur un réseau neuronal à convolution profonde en cascade consiste à : extraire une caractéristique faciale au moyen d'un réseau neuronal à convolution profonde en cascade ; et effectuer une reconnaissance faciale en fonction de la caractéristique faciale extraite. Le procédé et le système de reconnaissance faciale basés sur un réseau neuronal à convolution profonde en cascade de l'invention sont simples à commander, ce qui requiert une petite quantité de calcul et facilite l'accélération.
PCT/CN2020/079281 2019-03-15 2020-03-13 Procédé et système de reconnaissance faciale basés sur un réseau neuronal à convolution profonde en cascade WO2020187160A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910201162.4 2019-03-15
CN201910201162.4A CN111695392B (zh) 2019-03-15 2019-03-15 基于级联的深层卷积神经网络的人脸识别方法及系统

Publications (1)

Publication Number Publication Date
WO2020187160A1 true WO2020187160A1 (fr) 2020-09-24

Family

ID=72475529

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/079281 WO2020187160A1 (fr) 2019-03-15 2020-03-13 Procédé et système de reconnaissance faciale basés sur un réseau neuronal à convolution profonde en cascade

Country Status (2)

Country Link
CN (1) CN111695392B (fr)
WO (1) WO2020187160A1 (fr)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112749687A (zh) * 2021-01-31 2021-05-04 云知声智能科技股份有限公司 一种图片质量和静默活体检测多任务训练方法和设备
CN112818772A (zh) * 2021-01-19 2021-05-18 网易(杭州)网络有限公司 一种面部参数的识别方法、装置、电子设备及存储介质
CN113362110A (zh) * 2021-06-03 2021-09-07 中国电信股份有限公司 营销信息的推送方法、装置、电子设备和可读介质
CN113516146A (zh) * 2020-12-21 2021-10-19 腾讯科技(深圳)有限公司 一种数据分类方法、计算机及可读存储介质
CN114723756A (zh) * 2022-06-09 2022-07-08 北京理工大学 基于双监督网络的低分时序遥感目标检测方法及装置

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112395393B (zh) * 2020-11-27 2022-09-30 华东师范大学 一种基于多任务多示例的远程监督关系抽取方法
CN113160171B (zh) * 2021-04-20 2023-09-05 中日友好医院(中日友好临床医学研究所) 一种弹性超声成像图像处理的方法和装置

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105868689A (zh) * 2016-02-16 2016-08-17 杭州景联文科技有限公司 一种基于级联卷积神经网络的人脸遮挡检测方法
CN106339680A (zh) * 2016-08-25 2017-01-18 北京小米移动软件有限公司 人脸关键点定位方法及装置
CN106485215A (zh) * 2016-09-29 2017-03-08 西交利物浦大学 基于深度卷积神经网络的人脸遮挡检测方法
CN106951867A (zh) * 2017-03-22 2017-07-14 成都擎天树科技有限公司 基于卷积神经网络的人脸识别方法、装置、系统及设备
CN107967456A (zh) * 2017-11-27 2018-04-27 电子科技大学 一种基于人脸关键点的多神经网络级联识别人脸方法
CN108875833A (zh) * 2018-06-22 2018-11-23 北京智能管家科技有限公司 神经网络的训练方法、人脸识别方法及装置

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2723118B2 (ja) * 1992-08-31 1998-03-09 インターナショナル・ビジネス・マシーンズ・コーポレイション 2次元オブジェクトの認識に用いるためのニューラル・ネットワーク及び光学式文字認識装置
US6751354B2 (en) * 1999-03-11 2004-06-15 Fuji Xerox Co., Ltd Methods and apparatuses for video segmentation, classification, and retrieval using image class statistical models
CN103824054B (zh) * 2014-02-17 2018-08-07 北京旷视科技有限公司 一种基于级联深度神经网络的人脸属性识别方法
CN104463172B (zh) * 2014-12-09 2017-12-22 重庆中科云丛科技有限公司 基于人脸特征点形状驱动深度模型的人脸特征提取方法
CN107832700A (zh) * 2017-11-03 2018-03-23 全悉科技(北京)有限公司 一种人脸识别方法与系统
CN108304788B (zh) * 2018-01-18 2022-06-14 陕西炬云信息科技有限公司 基于深度神经网络的人脸识别方法
CN108564049A (zh) * 2018-04-22 2018-09-21 北京工业大学 一种基于深度学习的快速人脸检测识别方法
CN109448707A (zh) * 2018-12-18 2019-03-08 北京嘉楠捷思信息技术有限公司 一种语音识别方法及装置、设备、介质
CN109447053A (zh) * 2019-01-09 2019-03-08 江苏星云网格信息技术有限公司 一种基于双重限制注意力神经网络模型的人脸识别方法

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105868689A (zh) * 2016-02-16 2016-08-17 杭州景联文科技有限公司 一种基于级联卷积神经网络的人脸遮挡检测方法
CN106339680A (zh) * 2016-08-25 2017-01-18 北京小米移动软件有限公司 人脸关键点定位方法及装置
CN106485215A (zh) * 2016-09-29 2017-03-08 西交利物浦大学 基于深度卷积神经网络的人脸遮挡检测方法
CN106951867A (zh) * 2017-03-22 2017-07-14 成都擎天树科技有限公司 基于卷积神经网络的人脸识别方法、装置、系统及设备
CN107967456A (zh) * 2017-11-27 2018-04-27 电子科技大学 一种基于人脸关键点的多神经网络级联识别人脸方法
CN108875833A (zh) * 2018-06-22 2018-11-23 北京智能管家科技有限公司 神经网络的训练方法、人脸识别方法及装置

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113516146A (zh) * 2020-12-21 2021-10-19 腾讯科技(深圳)有限公司 一种数据分类方法、计算机及可读存储介质
CN112818772A (zh) * 2021-01-19 2021-05-18 网易(杭州)网络有限公司 一种面部参数的识别方法、装置、电子设备及存储介质
CN112749687A (zh) * 2021-01-31 2021-05-04 云知声智能科技股份有限公司 一种图片质量和静默活体检测多任务训练方法和设备
CN113362110A (zh) * 2021-06-03 2021-09-07 中国电信股份有限公司 营销信息的推送方法、装置、电子设备和可读介质
CN114723756A (zh) * 2022-06-09 2022-07-08 北京理工大学 基于双监督网络的低分时序遥感目标检测方法及装置
CN114723756B (zh) * 2022-06-09 2022-08-12 北京理工大学 基于双监督网络的低分时序遥感目标检测方法及装置

Also Published As

Publication number Publication date
CN111695392B (zh) 2023-09-15
CN111695392A (zh) 2020-09-22

Similar Documents

Publication Publication Date Title
WO2020187160A1 (fr) Procédé et système de reconnaissance faciale basés sur un réseau neuronal à convolution profonde en cascade
Gou et al. Vehicle license plate recognition based on extremal regions and restricted Boltzmann machines
CN111401257B (zh) 一种基于余弦损失在非约束条件下的人脸识别方法
WO2019232866A1 (fr) Procédé d'apprentissage de modèle d'œil humain, procédé de reconnaissance de modèle d'œil humain, appareil, dispositif et support
WO2020182121A1 (fr) Procédé de reconnaissance d'expression et dispositif associé
Ban et al. Face detection based on skin color likelihood
WO2019232862A1 (fr) Procédé et appareil d'apprentissage de modèle de bouche, procédé et appareil de reconnaissance de bouche, dispositif et support
WO2019114036A1 (fr) Procédé et dispositif de détection de visage, dispositif informatique et support d'informations lisible par ordinateur
Li et al. Robust visual tracking based on convolutional features with illumination and occlusion handing
WO2021139324A1 (fr) Procédé et appareil de reconnaissance d'image, support de stockage lisible par ordinateur et dispositif électronique
CN105512638B (zh) 一种基于融合特征的人脸检测与对齐方法
WO2020015752A1 (fr) Procédé, appareil et système d'identification d'attribut d'objet et dispositif informatique
US20180349716A1 (en) Apparatus and method for recognizing traffic signs
WO2023082784A1 (fr) Procédé et appareil de ré-identification de personne sur la base d'une attention d'attribut local
CN112232184B (zh) 一种基于深度学习和空间转换网络的多角度人脸识别方法
WO2021238586A1 (fr) Procédé et appareil d'entraînement, dispositif, et support de stockage lisible par ordinateur
CN112101208A (zh) 高龄老人特征串联融合手势识别方法及装置
Du High-precision portrait classification based on mtcnn and its application on similarity judgement
Yang et al. A Face Detection Method Based on Skin Color Model and Improved AdaBoost Algorithm.
Wang et al. Accurate playground localisation based on multi-feature extraction and cascade classifier in optical remote sensing images
Bhatia et al. Face detection using fuzzy logic and skin color segmentation in images
Obaida et al. Comparative of Viola-Jones and YOLO v3 for Face Detection in Real time
Chaki et al. Fragmented handwritten digit recognition using grading scheme and fuzzy rules
Sarmah et al. Facial identification expression-based attendance monitoring and emotion detection—A deep CNN approach
Huang et al. Eye landmarks detection via two-level cascaded CNNs with multi-task learning

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20773064

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS (EPO FORM 1205A DATED 02.02.2022)

122 Ep: pct application non-entry in european phase

Ref document number: 20773064

Country of ref document: EP

Kind code of ref document: A1