WO2016110005A1 - Gray level and depth information based multi-layer fusion multi-modal face recognition device and method - Google Patents

Gray level and depth information based multi-layer fusion multi-modal face recognition device and method Download PDF

Info

Publication number
WO2016110005A1
WO2016110005A1 PCT/CN2015/074868 CN2015074868W WO2016110005A1 WO 2016110005 A1 WO2016110005 A1 WO 2016110005A1 CN 2015074868 W CN2015074868 W CN 2015074868W WO 2016110005 A1 WO2016110005 A1 WO 2016110005A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
face
face recognition
depth information
dimensional
Prior art date
Application number
PCT/CN2015/074868
Other languages
French (fr)
Chinese (zh)
Inventor
夏春秋
Original Assignee
深圳市唯特视科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市唯特视科技有限公司 filed Critical 深圳市唯特视科技有限公司
Publication of WO2016110005A1 publication Critical patent/WO2016110005A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Definitions

  • the present invention relates to the field of face recognition technology, and in particular to a multi-modal face recognition device and method based on multi-layer fusion of gray and depth information.
  • 3D face recognition Compared with two-dimensional face recognition, 3D face recognition has the advantages of its robustness to illumination, small influence on posture and expression, etc. Therefore, after the rapid development of 3D data acquisition technology and the improvement of the quality and precision of 3D data, Many researchers have invested their research in this field.
  • CN20101025690 proposes related features of three-dimensional bending invariants for performing facial feature description.
  • the method extracts the bending invariant correlation feature by encoding the local features of the bending invariants of the adjacent nodes on the three-dimensional face surface; signing the relevant features of the bending invariant and performing spectral reduction using the spectral regression to obtain the principal component, and
  • the K-nearest neighbor classification method is used to identify three-dimensional faces.
  • the further application of the method is limited in efficiency;
  • CN200910197378 proposes a fully automatic three-dimensional face detection and posture correction method.
  • this method proposes facial region features to detect face surface coarsely, and proposes the tip region feature to accurately locate the tip of the nose, and then further accurately segment the complete Face surface, according to the distance information of the face surface to propose the characteristics of the nasal root region to detect the position of the nose root, a face coordinate system is established, and the face posture correction is automatically applied accordingly.
  • the purpose of this patent is to estimate the pose of three-dimensional face data, which belongs to the data preprocessing stage of the three-dimensional face recognition system.
  • Face grayscale images are susceptible to illumination changes, while face depth images are susceptible to data acquisition accuracy and expression changes. These factors affect the stability and accuracy of face recognition systems to some extent.
  • Multi-modal systems are getting more and more attention.
  • Multi-modal systems can take advantage of each modal data by acquiring multi-modal data, and overcome some inherent weaknesses of the single-mode system through fusion strategies (such as illumination of grayscale images, expressions of depth images). ), effectively improving the performance of the face recognition system.
  • Multi-modal systems can take advantage of the advantages of each modal data by using multi-modal data acquisition, and overcome the inherent weaknesses of single-mode systems (such as the illumination of grayscale images and the expression of depth images) through fusion strategies.
  • the performance of the face recognition system is improved, and the present invention adopts the following technical solutions to solve the above technical problems:
  • a multi-layer fusion multi-modal face recognition device based on gray level and depth information comprising a calculation unit for face recognition of gray information; a calculation unit for face recognition of depth information; based on multi-mode a unit for calculating the fusion of face recognition scores; a classifier calculation unit for classifying data.
  • the calculation unit for performing face recognition on gradation information comprises: a human eye detection unit, two-dimensional A data registration calculation unit, a grayscale face feature extraction unit, and a grayscale face recognition score calculation unit.
  • the calculation unit for performing face recognition on depth information comprises: a nose tip detector unit, and a three-dimensional data registration calculation A unit, a depth face feature extraction unit, and a depth face recognition score calculation unit.
  • the invention also discloses a multi-modal face recognition method based on multi-layer fusion of gray level and depth information, comprising the following steps:
  • the face gray information and depth information are normalized. Based on the normalized matching score, the fusion score is used to obtain the multi-modal fusion matching score to realize multi-modal face recognition.
  • the step A includes the following steps:
  • A1 Feature area localization, using a human eye detector to acquire a human eye region, the human eye detector being a hierarchical classifier H, obtained by the following algorithm:
  • each weak classifier h of the pair operates as follows:
  • the sample space ⁇ is divided to obtain X 1 X 2 ,..., X n ;
  • I a normalization factor such that D t+1 is a probability distribution
  • the final strong classifier H is
  • the LBP algorithm is used to process the human eye position data to obtain the LBP histogram feature, and the value formula is
  • the feature is input to the grayscale image classifier to obtain a grayscale matching score.
  • the step B includes the following steps:
  • the feature area is positioned to determine the position of the face of the face
  • Extraction is a visual dictionary histogram feature vector of a three-dimensional depth image. After the face image is input, after Gabor filtering, any filter vector is compared with all primitive words in the visual sub-dictionary corresponding to its position. By distance matching, it maps to the primitive closest to the distance, extracts the visual dictionary histogram feature of the original depth image, and uses the feature input depth image classifier to obtain the matching score.
  • the multi-modal face recognition based on the multi-layer fusion of gray level and depth information is described above.
  • the step c specifically includes:
  • the two-dimensional gray information and the three-dimensional depth information are subjected to fractional normalization using the principle of maximum and minimum linear normalization, and the formula is as follows
  • the linear discriminant analysis algorithm is used to maximize the objective function by constructing the intra-class scatter matrix SB and the inter-class scatter matrix SW.
  • the step B1 specifically includes
  • Step 1 determining a threshold, determining a threshold of the average average effective energy density of the domain, defined as thr;
  • Step 2 using the depth information to select the data to be processed, and using the depth information of the data to extract the face data in a certain depth range as the data to be processed;
  • Step 3 Calculating the normal vector, and calculating the direction quantity information of the face data selected by the depth information;
  • Step 4 Calculate the average negative effective energy density of the region. According to the definition of the regional average negative effective energy density, find the average negative effective energy density of the connected domains in the data to be processed, and select the most dense value. Large connected domain
  • Step 5 Determine whether the nose tip area is found. When the current area threshold is greater than the predefined thr, the area is the nose tip area, otherwise return to step 1 to restart the cycle.
  • the main steps of the ICP algorithm include:
  • Determining a matching data set pair selecting a reference data point set P from the three-dimensional nose point data in the reference template, and then using the closest distance between the point-to-point to select a data point set Q of the input three-dimensional face that matches the reference data;
  • the European type between the input data and the three-dimensional face model data in the registration library is calculated by the following formula: distance
  • P and Q are respectively a set of feature points to be matched, and the set contains N feature points.
  • step B4 is specifically:
  • the visual dictionary histogram vector is established as the special diagnosis expression of the three-dimensional human face
  • the nearest neighbor classifier is used as the final face recognition, where the L1 distance is chosen as the distance metric.
  • the present invention has the following technical effects:
  • the multi-modal system overcomes some inherent aspects of the single-modal system by using the advantages of two-dimensional gray information and three-dimensional depth information by utilizing the advantages of two-dimensional gray information and three-dimensional depth information.
  • Weaknesses (such as the illumination of grayscale images and the expression of depth images) effectively improve the performance of the face recognition system, making face recognition more accurate and faster.
  • Figure 1 is a flow chart of the present invention
  • FIG. 2 is a block diagram of the system of the present invention.
  • FIG. 3 is a schematic view showing the positioning of a three-dimensional human face tip according to the present invention.
  • FIG. 4 is a schematic diagram of a three-dimensional human face space mapping according to the present invention.
  • FIG. 5 is a schematic diagram of extracting features of a three-dimensional face depth representation of the present invention.
  • FIG. 6 is a schematic diagram of a two-dimensional human face human eye detection according to the present invention.
  • FIG. 7 is a schematic diagram of a two-dimensional face LBP feature of the present invention.
  • FIG. 8 is a schematic diagram of extracting features of a two-dimensional face grayscale representation according to the present invention.
  • FIG. 9 is a schematic diagram of different modal score fusion algorithms according to the present invention.
  • the invention discloses a multi-modality face recognition device based on multi-layer fusion of grayscale and depth information, comprising a calculation unit for face recognition of grayscale information; and a face recognition method for depth information.
  • a computing unit a computing unit that fuses based on multimodal face recognition scores; a classifier computing unit that classifies data.
  • the grayscale information face recognition calculation unit specifically includes a human eye detection unit, a two-dimensional data registration calculation unit, a grayscale face feature extraction unit, and a grayscale face recognition score calculation unit.
  • the depth information face recognition calculation unit specifically includes a nose tip detector unit, a three-dimensional data registration calculation unit, a depth face feature extraction unit, and a depth face recognition score calculation unit.
  • the present invention also discloses a multi-modality face recognition method based on multi-layer fusion of gray and depth information.
  • the multi-modal fusion system disclosed by the present invention includes multiple data sources: Grayscale image, 3D depth image.
  • feature point detection human eye
  • the obtained feature point position is used for registration.
  • the LBP histogram feature is acquired by the LBP algorithm, and the The feature input gray image classifier obtains the matching score; for the three-dimensional depth data, the feature point detection (nose tip) is first performed and the acquired feature points are used for registration, and then the registered three-dimensional spatial data is mapped into the face depth image, and the The visual dictionary algorithm acquires a visual dictionary histogram feature for the data, and uses the feature input depth image classifier to obtain a matching score.
  • the multi-modal system utilizes a decision-making layer fusion strategy. Therefore, after obtaining the matching scores of each data source, these scores need to be normalized, and then based on the normalized matching scores, the fusion strategy can be used to obtain the multi-modal fusion. Match scores to achieve multimodal face recognition.
  • the human eye detector is obtained by a human eye detector, which is a hierarchical classifier, each layer is a strong classifier (such as Adaboost), and each layer filters a part of the non-human.
  • the image area finally obtained is the human eye area.
  • the advantage of the hierarchical classifier is that the first few levels of classifiers contain fewer features, so the calculation speed is faster; after the previous few levels of classifiers, although the level classifier complexity increases, but the rest of the time The image area has been relatively small.
  • Adaboost algorithm can summarize as follows:
  • each weak classifier h of the pair operates as follows:
  • the sample space ⁇ is divided to obtain X 1 , X 2 , ... X n ;
  • Update training sample probability distribution among them Is a normalization factor such that D t+1 is a probability distribution
  • the final strong classifier H is
  • the obtained human eye region position is used for registration, and the LBP algorithm is used to process the human eye position data to obtain the LBP histogram feature.
  • the LBP algorithm compares the pixel point with its neighboring pixel point, and the value thereof is as follows.
  • LBP histogram features LBP histogram features
  • the input two-dimensional face data first extracts a key point by human eye detection, and then adjusts the face image to a positive upright posture according to the position of the human eye through rigid transformation.
  • the LBP histogram features will be extracted from the registered grayscale map.
  • the feature is input to the grayscale image classifier to obtain a grayscale matching score.
  • the detection of the nose tip region of the face is first performed, specifically by the following steps:
  • Determining a threshold determining a threshold of a domain average negative effective energy density, defined as thr;
  • the depth data is used to select the data to be processed, and the depth information of the data is used to extract the face data in a certain depth range as the data to be processed;
  • the calculation of the normal vector calculates the direction quantity information of the face data selected by the depth information
  • the acquired nose region is used for registration.
  • the ICP algorithm is used for data registration, and the reference point data point set P is first selected from the three-dimensional nose data in the reference template, and then the point-to-point point is used. The nearest distance between the two is selected to input the data point set Q in the three-dimensional face that matches the reference data, and the matrix of 3*3 is first calculated.
  • N is the capacity of the data set
  • the distance function is as follows:
  • P and Q are respectively a set of feature points to be matched, and the set contains N feature points.
  • the distance needs to be normalized according to the number of effective feature points.
  • the depth image is first acquired according to the depth information, and then the noise point (data bump or hole point) in the mapped depth image is compensated by the filter. Denoising, finally selecting the robust region of the expression to obtain the final 3D face depth image.
  • any filter vector is compared with all primitive vocabularies in the visual sub-dictionary corresponding to its position, and it is matched by distance matching. Map to the primitive closest to it. In this way, the visual dictionary histogram features of the original depth image can be extracted.
  • the visual dictionary histogram vector is established as the special diagnosis expression of the three-dimensional human face
  • the nearest neighbor classifier is used as the final face recognition, where the L1 distance is chosen as the distance metric.
  • the two-dimensional gray information and the three-dimensional depth information are subjected to fractional normalization using the principle of maximum and minimum linear normalization, and the formula is as follows
  • max represents a position farther away from the space, this value is easily affected by noise (such as 3D human face hair occlusion, etc.), so in the max
  • the value is taken, it is the value of the single modal score set ⁇ S k ⁇ at the 95% position after the ascending order; and since min represents the position closer in the distance space, the value is not affected by the data noise (this The affected value will become larger), so the min value is the minimum value after the single modal score set ⁇ S k ⁇ is sorted in ascending order.
  • S k is the matching score in the modality, Is the normalized matching score in the modality;
  • the matching score after multimodal data fusion is obtained.
  • the weight is obtained by a linear discriminant analysis algorithm (LDA).
  • LDA linear discriminant analysis algorithm
  • the algorithm utilizes the category information of the data to maximize the objective function by constructing an intra-class scatter matrix SB and an inter-class scatter matrix SW.
  • the multi-modal system overcomes some inherent aspects of the single-modal system by using the advantages of two-dimensional gray information and three-dimensional depth information by utilizing the advantages of two-dimensional gray information and three-dimensional depth information.
  • Weaknesses (such as the illumination of grayscale images and the expression of depth images) effectively improve the performance of the face recognition system, making face recognition more accurate and faster.

Abstract

Disclosed in the present invention are a gray level and depth information based multi-layer fusion multi-modal face recognition device and method, the method mainly comprising the steps of: recognizing gray level information of a human face; recognizing depth information of the human face; normalizing the gray level information and the depth information of the human face, and on the basis of normalized matching scores, acquiring a matching score fused multi-modally by adopting a fusion approach to achieve multi-modal face recognition. In the solution of the present invention, a multi-modal system collects two-dimensional gray level information and three-dimensional depth information, takes advantages from the two-dimensional gray level information and the three-dimensional depth information, and overcomes inherent shortcomings of a single-modal system (such as illumination for gray level images and expressions for depth images) via the fusion approach, thereby effectively improving the performance of the face recognition system and bringing more accurate and rapid face recognition.

Description

基于灰度和深度信息的多层融合的多模态人脸识别装置及方法Multi-layer face recognition device and method based on multi-layer fusion of gray level and depth information 技术领域Technical field
本发明涉及人脸识别技术领域,尤其涉一种基于灰度和深度信息的多层融合的多模态人脸识别装置及方法。The present invention relates to the field of face recognition technology, and in particular to a multi-modal face recognition device and method based on multi-layer fusion of gray and depth information.
背景技术Background technique
三维人脸识别相对于二维人脸识别,有着其对光照鲁棒、受姿态以及表情等因素影响较小等优点,因此在三维数据采集技术飞速发展以及三维数据的质量和精度大大提升之后,很多学者都将他们的研究投入到该领域中。Compared with two-dimensional face recognition, 3D face recognition has the advantages of its robustness to illumination, small influence on posture and expression, etc. Therefore, after the rapid development of 3D data acquisition technology and the improvement of the quality and precision of 3D data, Many scholars have invested their research in this field.
CN20101025690提出了三维弯曲不变量的相关特征用来进行人脸特性描述。该方法通过编码三维人脸表面相邻节点的弯曲不变量的局部特征,提取弯曲不变量相关特征;对所述弯曲不变量的相关特征进行签名并采用谱回归进行降维,获得主成分,并运用K最近邻分类方法对三维人脸进行识别。但是由于提取变量相关特征时需要复杂的计算量,因此在效率上限制了该方法的进一步应用;CN20101025690 proposes related features of three-dimensional bending invariants for performing facial feature description. The method extracts the bending invariant correlation feature by encoding the local features of the bending invariants of the adjacent nodes on the three-dimensional face surface; signing the relevant features of the bending invariant and performing spectral reduction using the spectral regression to obtain the principal component, and The K-nearest neighbor classification method is used to identify three-dimensional faces. However, due to the complex calculation amount required to extract the variables related features, the further application of the method is limited in efficiency;
CN200910197378提出了一种全自动三维人脸检测和姿势纠正的方法。该方法通过对人脸三维曲面进行多尺度的矩分析,提出了脸部区域特征来粗糙地检测人脸曲面,及提出鼻尖区域特征来准确地定位鼻尖的位置,然后进一步精确地分割出完整的人脸曲面,根据人脸曲面的距离信息提出鼻根区域特征来检测鼻根的位置后,建立了一个人脸坐标系,并据此自动地进行人脸姿势的纠正应用。该专利目的在于对三维人脸数据的姿态进行估计,属于三维人脸识别系统的数据预处理阶段。 CN200910197378 proposes a fully automatic three-dimensional face detection and posture correction method. By multi-scale moment analysis of human face three-dimensional surface, this method proposes facial region features to detect face surface coarsely, and proposes the tip region feature to accurately locate the tip of the nose, and then further accurately segment the complete Face surface, according to the distance information of the face surface to propose the characteristics of the nasal root region to detect the position of the nose root, a face coordinate system is established, and the face posture correction is automatically applied accordingly. The purpose of this patent is to estimate the pose of three-dimensional face data, which belongs to the data preprocessing stage of the three-dimensional face recognition system.
人脸灰度图像容易受到光照变化的影响,而人脸深度图像容易受到数据采集精度以及表情变化等影响,这些因素在一定程度上影响了人脸识别系统的稳定性及准确性。Face grayscale images are susceptible to illumination changes, while face depth images are susceptible to data acquisition accuracy and expression changes. These factors affect the stability and accuracy of face recognition systems to some extent.
因此多模态融合系统越来越受到人们的关注。多模态系统通过进行多模态数据的采集,可以利用每个模态数据的优点,并通过融合策略来克服单模态系统的某些内在弱点(如灰度图像的光照,深度图像的表情),有效地提升了人脸识别系统的性能。Therefore, multimodal fusion systems are getting more and more attention. Multi-modal systems can take advantage of each modal data by acquiring multi-modal data, and overcome some inherent weaknesses of the single-mode system through fusion strategies (such as illumination of grayscale images, expressions of depth images). ), effectively improving the performance of the face recognition system.
发明内容Summary of the invention
为了解决上述技术问题,多模态融合系统越来越受到人们的关注。多模态系统通过进行多模态数据的采集,可以利用每个模态数据的优点,通过融合策略来克服单模态系统的内在弱点(如灰度图像的光照,深度图像的表情),有效的提升了人脸识别系统的性能,本发明采用如下技术方案来解决上述技术问题:In order to solve the above technical problems, multi-modal fusion systems have attracted more and more attention. Multi-modal systems can take advantage of the advantages of each modal data by using multi-modal data acquisition, and overcome the inherent weaknesses of single-mode systems (such as the illumination of grayscale images and the expression of depth images) through fusion strategies. The performance of the face recognition system is improved, and the present invention adopts the following technical solutions to solve the above technical problems:
一种基于灰度和深度信息的多层融合的多模态人脸识别装置,包括对于灰度信息进行人脸识别的计算单元;用于对深度信息进行人脸识别的计算单元;基于多模态人脸识别分数进行融合的计算单元;对数据进行分类的分类器计算单元。A multi-layer fusion multi-modal face recognition device based on gray level and depth information, comprising a calculation unit for face recognition of gray information; a calculation unit for face recognition of depth information; based on multi-mode a unit for calculating the fusion of face recognition scores; a classifier calculation unit for classifying data.
优选的,在上述的一种基于灰度和深度信息的多层融合的多模态人脸识别装置中,所述对于灰度信息进行人脸识别的计算单元包括:人眼检测单元、二维数据配准计算单元、灰度人脸特征提取单元和灰度人脸识别分数计算单元。Preferably, in the above-described multi-modality face recognition device based on gradation and depth information, the calculation unit for performing face recognition on gradation information comprises: a human eye detection unit, two-dimensional A data registration calculation unit, a grayscale face feature extraction unit, and a grayscale face recognition score calculation unit.
优选的,在上述的一种基于灰度和深度信息的多层融合的多模态人脸识别装置中,对深度信息进行人脸识别的计算单元包括:鼻尖检测器单元、三维数据配准计算单元、深度人脸特征提取单元和深度人脸识别分数计算单元。 Preferably, in the above-mentioned multi-modal face recognition device based on gradation and depth information, the calculation unit for performing face recognition on depth information comprises: a nose tip detector unit, and a three-dimensional data registration calculation A unit, a depth face feature extraction unit, and a depth face recognition score calculation unit.
本发明还公开一种一种基于灰度和深度信息的多层融合的多模态人脸识别方法,包括如下步骤:The invention also discloses a multi-modal face recognition method based on multi-layer fusion of gray level and depth information, comprising the following steps:
A.对人脸灰度信息进行识别;A. Identify the face grayscale information;
B.对人脸深度信息进行识别;B. Identify face depth information;
C.对人脸灰度信息及深度信息进行归一化,基于归一化的匹配分数,采用融合策略得到多模态融合后的匹配分数,实现多模态人脸识别。C. The face gray information and depth information are normalized. Based on the normalized matching score, the fusion score is used to obtain the multi-modal fusion matching score to realize multi-modal face recognition.
优选的,在上述的一种基于灰度和深度信息的多层融合的多模态脸识别方法中,所述步骤A包括如下步骤:Preferably, in the above-mentioned multi-modality face recognition method based on gradation and depth information, the step A includes the following steps:
A1.特征区域定位,使用人眼检测器获取人眼区域,所述人眼检测器为层级分类器H,经如下算法得到:A1. Feature area localization, using a human eye detector to acquire a human eye region, the human eye detector being a hierarchical classifier H, obtained by the following algorithm:
给定训练样本集合S={(x1,y1),...,(xm,ym)},弱空间分类器
Figure PCTCN2015074868-appb-000001
其中xi∈χ,为样本向量,yi=±1,为分类标签,m为样本总数;初始化样本概率分布
Figure PCTCN2015074868-appb-000002
Given a set of training samples S = {(x 1 , y 1 ), ..., (x m , y m )}, weak space classifier
Figure PCTCN2015074868-appb-000001
Where x i ∈χ is the sample vector, y i =±1, is the classification label, m is the total number of samples; initial sample probability distribution
Figure PCTCN2015074868-appb-000002
t=1,...,T,对中的每个弱分类器h作如下操作:t=1,...,T, each weak classifier h of the pair operates as follows:
对样本空间χ进行划分,得到X1X2,...,XnThe sample space χ is divided to obtain X 1 X 2 ,..., X n ;
Figure PCTCN2015074868-appb-000003
其中ε为一小正常数;
Figure PCTCN2015074868-appb-000003
Where ε is a small normal number;
计算归一化因子,
Figure PCTCN2015074868-appb-000004
Calculate the normalization factor,
Figure PCTCN2015074868-appb-000004
在弱分类器空间中选择一个ht,使得Z最小化Select an h t in the weak classifier space to minimize Z
Figure PCTCN2015074868-appb-000005
Figure PCTCN2015074868-appb-000005
更新训练样本概率分布
Figure PCTCN2015074868-appb-000006
其中
Update training sample probability distribution
Figure PCTCN2015074868-appb-000006
among them
Figure PCTCN2015074868-appb-000007
Figure PCTCN2015074868-appb-000007
Figure PCTCN2015074868-appb-000008
为归一化因子,使得Dt+1为一个概率分布;
Figure PCTCN2015074868-appb-000008
Is a normalization factor such that D t+1 is a probability distribution;
最终强分类器H为
Figure PCTCN2015074868-appb-000009
The final strong classifier H is
Figure PCTCN2015074868-appb-000009
A2.使用获得的人眼区域位置进行配准,利用LBP算法处理人眼位置数据获得LBP直方图特征,取值公式为A2. Using the obtained position of the human eye region for registration, the LBP algorithm is used to process the human eye position data to obtain the LBP histogram feature, and the value formula is
Figure PCTCN2015074868-appb-000010
Figure PCTCN2015074868-appb-000010
将该特征输入灰度图像分类器获取灰度匹配分数。The feature is input to the grayscale image classifier to obtain a grayscale matching score.
优选的,在上述的一种基于灰度和深度信息的多层融合的多模态人脸识别方法中,所述步骤B包括如下步骤:Preferably, in the above-mentioned multi-modal face recognition method based on gradation and depth information, the step B includes the following steps:
B1.特征区域定位,判定人脸鼻尖区域位置;B1. The feature area is positioned to determine the position of the face of the face;
B2.对于不同姿态的三维数据,得到配准的参考区域后,按照ICP算法进行数据的配准,配准完成后计算输入数据与注册库中的三维人脸模型数据之间的欧式距离;B2. For the three-dimensional data of different postures, after the registration reference area is obtained, the data is registered according to the ICP algorithm, and the Euclidean distance between the input data and the three-dimensional face model data in the registration library is calculated after the registration is completed;
B3.按照深度信息进行深度图像的获取,利用滤波器对于映射后的深度图像中的噪音点进行补偿去噪,最后对表情鲁棒区域进行选择,得到最终的三维人脸深度图像;B3. Obtain the depth image according to the depth information, and use the filter to compensate and denoise the noise points in the mapped depth image, and finally select the robust region of the expression to obtain the final 3D face depth image;
B4.提取是三维深度图像的视觉词典直方图特征向量,当测试人脸图像输入后,经过Gabor滤波后,将任一滤波向量都与其所在位置相对应的视觉分词典中的所有基元词汇比较,通过距离匹配的方式,把它映射到与之距离最为接近的基元上,提取出原始深度图像的视觉词典直方图特征,利用该特征输入深度图像分类器获取匹配分数。B4. Extraction is a visual dictionary histogram feature vector of a three-dimensional depth image. After the face image is input, after Gabor filtering, any filter vector is compared with all primitive words in the visual sub-dictionary corresponding to its position. By distance matching, it maps to the primitive closest to the distance, extracts the visual dictionary histogram feature of the original depth image, and uses the feature input depth image classifier to obtain the matching score.
优选的,在上述的一种基于灰度和深度信息的多层融合的多模态人脸识别 方法中,所述步骤c具体包括:Preferably, the multi-modal face recognition based on the multi-layer fusion of gray level and depth information is described above. In the method, the step c specifically includes:
对二维灰度信息和三维深度信息采用最大最小线性归一化原则进行分数归一化,公式如下The two-dimensional gray information and the three-dimensional depth information are subjected to fractional normalization using the principle of maximum and minimum linear normalization, and the formula is as follows
Figure PCTCN2015074868-appb-000011
Figure PCTCN2015074868-appb-000011
分数归一化之后,采用比较鲁棒的加权加法原则对不同模态的匹配分数进行融合,公式如下After the score is normalized, the matching scores of different modes are merged by using the relatively robust weighted addition principle. The formula is as follows
Figure PCTCN2015074868-appb-000012
Figure PCTCN2015074868-appb-000012
得到多模态数据融合后的匹配分数后采用线性判别分析算法通过构建类内散布矩阵SB和类间散布矩阵SW,最大化目标函数After obtaining the matching scores of the multi-modal data fusion, the linear discriminant analysis algorithm is used to maximize the objective function by constructing the intra-class scatter matrix SB and the inter-class scatter matrix SW.
Figure PCTCN2015074868-appb-000013
Figure PCTCN2015074868-appb-000013
获取LDA映射矩阵W,即为权值。Obtain the LDA mapping matrix W, which is the weight.
优选的,在上述的一种基于灰度和深度信息的多层融合的多模态人脸识别方法中,所述步骤B1具体包括Preferably, in the above-mentioned multi-modal face recognition method based on gradation and depth information, the step B1 specifically includes
步骤1:确定阈值,确定域平均负有效能量密度的阈值,定义为thr;Step 1: determining a threshold, determining a threshold of the average average effective energy density of the domain, defined as thr;
步骤2:利用深度信息选取待处理数据,利用数据的深度信息,提取在一定深度范围内的人脸数据作为待处理数据;Step 2: using the depth information to select the data to be processed, and using the depth information of the data to extract the face data in a certain depth range as the data to be processed;
步骤3:法向量的计算,计算由深度信息选取出的人脸数据的方向量信息;Step 3: Calculating the normal vector, and calculating the direction quantity information of the face data selected by the depth information;
步骤4:区域平均负有效能量密度的计算,按照区域平均负有效能量密度的定义,求出待处理数据中个连通域的平均负有效能量密度,选择其中密度值最 大的连通域;Step 4: Calculate the average negative effective energy density of the region. According to the definition of the regional average negative effective energy density, find the average negative effective energy density of the connected domains in the data to be processed, and select the most dense value. Large connected domain
步骤5:判定是否找到鼻尖区域,当前区域阈值大于预定义的thr时,该区域即为鼻尖区域,否则回到步骤1重新开始循环。Step 5: Determine whether the nose tip area is found. When the current area threshold is greater than the predefined thr, the area is the nose tip area, otherwise return to step 1 to restart the cycle.
优选的,在上述的一种基于灰度和深度信息的多层融合的多模态人脸识别方法中,所述ICP算法主要步骤包括:Preferably, in the multi-modal face recognition method based on the multi-layer fusion of gray level and depth information, the main steps of the ICP algorithm include:
确定匹配数据集合对,从参考模板中的三维鼻尖数据选取参考数据点集P,再利用点对点之间的最近的距离来选择输入三维人脸中与参考数据相匹配的数据点集Q;Determining a matching data set pair, selecting a reference data point set P from the three-dimensional nose point data in the reference template, and then using the closest distance between the point-to-point to select a data point set Q of the input three-dimensional face that matches the reference data;
计算刚性运动参数,计算旋转矩阵R和平移向量tCalculate the rigid motion parameters and calculate the rotation matrix R and the translation vector t
当X行列式值为1时,R=X;When the X determinant value is 1, R = X;
t=P-R*Qt=P-R*Q
根据刚性变换后的数据集RQ+t和参考数据集P之间的误差判断三维数据集是否配准,配准之后通过下式计算输入数据与注册库中的三维人脸模型数据之间的欧式距离According to the error between the rigid transformed data set RQ+t and the reference data set P, whether the three-dimensional data set is registered or not, after registration, the European type between the input data and the three-dimensional face model data in the registration library is calculated by the following formula: distance
Figure PCTCN2015074868-appb-000014
Figure PCTCN2015074868-appb-000014
其中P,Q分别是待匹配的特征点集合,集合中含有N个特征点。Where P and Q are respectively a set of feature points to be matched, and the set contains N feature points.
优选的,在上述的一种基于灰度和深度信息的多层融合的多模态人脸识别方法中,步骤B4具体为:Preferably, in the multi-modal face recognition method based on the multi-layer fusion of gray level and depth information, step B4 is specifically:
将三维人脸深度图像分割成一些局部纹理区域;Dividing the 3D face depth image into some local texture regions;
对于每个GaBor滤波响应向量,按照位置的不同将其映射到其对应的视觉分词典的词汇中,并依此为基础建立视觉词典直方图向量作为三维人脸的特诊表达;For each GaBor filter response vector, it is mapped to the vocabulary of its corresponding visual sub-dictionary according to the position, and based on this, the visual dictionary histogram vector is established as the special diagnosis expression of the three-dimensional human face;
将最近邻分类器用来作为最后的人脸识别,其中L1距离被选作为距离度量。 The nearest neighbor classifier is used as the final face recognition, where the L1 distance is chosen as the distance metric.
与现有技术相比,本发明具有如下技术效果:Compared with the prior art, the present invention has the following technical effects:
采用本发明的方案,多模态系统通过进行二维灰度信息和三维深度信息的采集,利用二维灰度信息和三维深度信息的优点,通过融合策略来克服单模态系统的某些内在弱点(如灰度图像的光照,深度图像的表情),有效地提升了人脸识别系统的性能,使得人脸识别更加准确快捷。By adopting the scheme of the invention, the multi-modal system overcomes some inherent aspects of the single-modal system by using the advantages of two-dimensional gray information and three-dimensional depth information by utilizing the advantages of two-dimensional gray information and three-dimensional depth information. Weaknesses (such as the illumination of grayscale images and the expression of depth images) effectively improve the performance of the face recognition system, making face recognition more accurate and faster.
附图说明DRAWINGS
图1为本发明流程框图;Figure 1 is a flow chart of the present invention;
图2为本发明系统框图;Figure 2 is a block diagram of the system of the present invention;
图3为本发明三维人脸鼻尖定位示意图;3 is a schematic view showing the positioning of a three-dimensional human face tip according to the present invention;
图4为本发明三维人脸空间映射示意图;4 is a schematic diagram of a three-dimensional human face space mapping according to the present invention;
图5为本发明人三维人脸深度表象特征提取示意图;FIG. 5 is a schematic diagram of extracting features of a three-dimensional face depth representation of the present invention; FIG.
图6为本发明二维人脸人眼检测示意图;6 is a schematic diagram of a two-dimensional human face human eye detection according to the present invention;
图7为本发明二维人脸LBP特征示意图;7 is a schematic diagram of a two-dimensional face LBP feature of the present invention;
图8为本发明二维人脸灰度表象特征提取示意图;8 is a schematic diagram of extracting features of a two-dimensional face grayscale representation according to the present invention;
图9为本发明不同模态分数融合算法示意图。FIG. 9 is a schematic diagram of different modal score fusion algorithms according to the present invention.
具体实施方式detailed description
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention are clearly and completely described in the following with reference to the accompanying drawings in the embodiments of the present invention. It is obvious that the described embodiments are only a part of the embodiments of the present invention, but not all embodiments. All other embodiments obtained by those skilled in the art based on the embodiments of the present invention without creative efforts are within the scope of the present invention.
本发明公开一种基于灰度和深度信息的多层融合的多模态人脸识别装置,包括对于灰度信息进行人脸识别的计算单元;用于对深度信息进行人脸识别的 计算单元;基于多模态人脸识别分数进行融合的计算单元;对数据进行分类的分类器计算单元。The invention discloses a multi-modality face recognition device based on multi-layer fusion of grayscale and depth information, comprising a calculation unit for face recognition of grayscale information; and a face recognition method for depth information. a computing unit; a computing unit that fuses based on multimodal face recognition scores; a classifier computing unit that classifies data.
其中上述灰度信息人脸识别计算单元具体包括人眼检测单元、二维数据配准计算单元、灰度人脸特征提取单元和灰度人脸识别分数计算单元。The grayscale information face recognition calculation unit specifically includes a human eye detection unit, a two-dimensional data registration calculation unit, a grayscale face feature extraction unit, and a grayscale face recognition score calculation unit.
上述深度信息人脸识别计算单元具体包括鼻尖检测器单元、三维数据配准计算单元、深度人脸特征提取单元和深度人脸识别分数计算单元。The depth information face recognition calculation unit specifically includes a nose tip detector unit, a three-dimensional data registration calculation unit, a depth face feature extraction unit, and a depth face recognition score calculation unit.
同时本发明还公开一种基于灰度和深度信息的多层融合的多模态人脸识别方法,如图9所示,本发明公开的多模态融合系统包括多个数据源:如二维灰度图像,三维深度图像。对于二维灰度图像,首先进行特征点检测(人眼),然后利用获得的特征点位置进行配准,在灰度图像配准后,利用LBP算法对该数据获取LBP直方图特征,利用该特征输入灰度图像分类器获取匹配分数;对于三维深度数据,首先进行特征点检测(鼻尖)并利用获取的特征点进行配准,然后将配准后的三维空间数据映射为人脸深度图像,利用视觉词典算法对该数据获取视觉词典直方图特征,利用该特征输入深度图像分类器获取匹配分数。该多模态系统利用决策层融合策略,因此在获取各数据源匹配分数后,需要对这些分数进行归一化,然后基于归一化的匹配分数,可以采用融合策略得到多模态融合后的匹配分数,以此实现多模态人脸识别。At the same time, the present invention also discloses a multi-modality face recognition method based on multi-layer fusion of gray and depth information. As shown in FIG. 9, the multi-modal fusion system disclosed by the present invention includes multiple data sources: Grayscale image, 3D depth image. For the two-dimensional gray image, feature point detection (human eye) is first performed, and then the obtained feature point position is used for registration. After the gray image registration, the LBP histogram feature is acquired by the LBP algorithm, and the The feature input gray image classifier obtains the matching score; for the three-dimensional depth data, the feature point detection (nose tip) is first performed and the acquired feature points are used for registration, and then the registered three-dimensional spatial data is mapped into the face depth image, and the The visual dictionary algorithm acquires a visual dictionary histogram feature for the data, and uses the feature input depth image classifier to obtain a matching score. The multi-modal system utilizes a decision-making layer fusion strategy. Therefore, after obtaining the matching scores of each data source, these scores need to be normalized, and then based on the normalized matching scores, the fusion strategy can be used to obtain the multi-modal fusion. Match scores to achieve multimodal face recognition.
如图6所示,通过人眼检测器获取人眼区域,,该人眼检测器由为层级分类器,每一层都是一个强分类器(如Adaboost),每一层都会过滤一部分非人眼区域,最后获得的图像区域就是人眼区域。层级分类器的好处在于,前几层分类器所包含特征比较少,因此计算速度比较快;在经过前面几层的分类器之后,虽然层级分类器复杂度升高,但是此时所剩下的图像区域已经比较少。通过上述机理,该层级分类器可以达到实时的检测性能。Adaboost算法可以总结 如下:As shown in FIG. 6, the human eye detector is obtained by a human eye detector, which is a hierarchical classifier, each layer is a strong classifier (such as Adaboost), and each layer filters a part of the non-human. In the eye area, the image area finally obtained is the human eye area. The advantage of the hierarchical classifier is that the first few levels of classifiers contain fewer features, so the calculation speed is faster; after the previous few levels of classifiers, although the level classifier complexity increases, but the rest of the time The image area has been relatively small. Through the above mechanism, the hierarchical classifier can achieve real-time detection performance. Adaboost algorithm can summarize as follows:
给定训练样本集合S={(x1,y1),...,(xm,ym)},弱空间分类器
Figure PCTCN2015074868-appb-000015
其中xi∈χ,为样本向量,yi=±1,为分类标签,m为样本总数;初始化样本概率分布
Figure PCTCN2015074868-appb-000016
Given a set of training samples S = {(x 1 , y 1 ), ..., (x m , y m )}, weak space classifier
Figure PCTCN2015074868-appb-000015
Where x i ∈χ is the sample vector, y i =±1, is the classification label, m is the total number of samples; initial sample probability distribution
Figure PCTCN2015074868-appb-000016
t=1,...,T,对中的每个弱分类器h作如下操作:t=1,...,T, each weak classifier h of the pair operates as follows:
对样本空间χ进行划分,得到X1,X2,...XnThe sample space χ is divided to obtain X 1 , X 2 , ... X n ;
Figure PCTCN2015074868-appb-000017
其中ε为一小正常数;
Figure PCTCN2015074868-appb-000017
Where ε is a small normal number;
计算归一化因子,
Figure PCTCN2015074868-appb-000018
Calculate the normalization factor,
Figure PCTCN2015074868-appb-000018
在弱分类器空间中选择一个ht,使得Z最小化Select an h t in the weak classifier space to minimize Z
Figure PCTCN2015074868-appb-000019
Figure PCTCN2015074868-appb-000019
更新训练样本概率分布
Figure PCTCN2015074868-appb-000020
其中
Figure PCTCN2015074868-appb-000021
为归一化因子,使得Dt+1为一个概率分布;
Update training sample probability distribution
Figure PCTCN2015074868-appb-000020
among them
Figure PCTCN2015074868-appb-000021
Is a normalization factor such that D t+1 is a probability distribution;
最终强分类器H为
Figure PCTCN2015074868-appb-000022
The final strong classifier H is
Figure PCTCN2015074868-appb-000022
如图7、8所示,使用获得的人眼区域位置进行配准,利用LBP算法处理人眼位置数据获得LBP直方图特征,LBP算法将像素点与其邻域像素点做对比,其取值如公式: As shown in FIGS. 7 and 8, the obtained human eye region position is used for registration, and the LBP algorithm is used to process the human eye position data to obtain the LBP histogram feature. The LBP algorithm compares the pixel point with its neighboring pixel point, and the value thereof is as follows. Formula:
Figure PCTCN2015074868-appb-000023
Figure PCTCN2015074868-appb-000023
如果取P=8,R=1,则一些具有纹理特性的意义的LBP值如图(c)所示。其中第一幅图代表的是纹理亮点,第二幅图代表纹理边界,第三幅图代表纹理暗点或是平滑纹理区域。按照纹理的统计分布规律将所得LBP值归为59类,并把这59类作为直方图的基础构造统计特征向量(LBP直方图特征)。通过这种形式把局部纹理信息的描述性和直方图的鲁棒性有效结合,在人脸识别领域取得了不错的识别性能。If P = 8 and R = 1, some LBP values with the meaning of texture characteristics are shown in Figure (c). The first image represents texture highlights, the second image represents texture boundaries, and the third image represents texture dark spots or smooth texture regions. According to the statistical distribution law of the texture, the obtained LBP values are classified into 59 categories, and these 59 categories are used as the basic structural statistical feature vectors (LBP histogram features) of the histogram. This form effectively combines the descriptiveness of local texture information with the robustness of histograms, and has achieved good recognition performance in the field of face recognition.
输入的二维人脸数据,先通过人眼检测提取出关键点,然后根据人眼位置将该人脸图像通过刚性变换调整为正向直立姿态。将通过配准的灰度图提取出LBP直方图特征。The input two-dimensional face data first extracts a key point by human eye detection, and then adjusts the face image to a positive upright posture according to the position of the human eye through rigid transformation. The LBP histogram features will be extracted from the registered grayscale map.
将该特征输入灰度图像分类器获取灰度匹配分数。The feature is input to the grayscale image classifier to obtain a grayscale matching score.
如图3所示,对于三维深度数据,首先进行人脸鼻尖区域的检测,具体通过如下步骤:As shown in FIG. 3, for the three-dimensional depth data, the detection of the nose tip region of the face is first performed, specifically by the following steps:
确定阈值,确定域平均负有效能量密度的阈值,定义为thr;Determining a threshold, determining a threshold of a domain average negative effective energy density, defined as thr;
利用深度信息选取待处理数据,利用数据的深度信息,提取在一定深度范围内的人脸数据作为待处理数据;The depth data is used to select the data to be processed, and the depth information of the data is used to extract the face data in a certain depth range as the data to be processed;
法向量的计算,计算由深度信息选取出的人脸数据的方向量信息;The calculation of the normal vector calculates the direction quantity information of the face data selected by the depth information;
区域平均负有效能量密度的计算,按照区域平均负有效能量密度的定义,求出待处理数据中个连通域的平均负有效能量密度,选择其中密度值最大的连通域;Calculating the average negative effective energy density of the region, according to the definition of the regional average negative effective energy density, finding the average negative effective energy density of the connected domains in the data to be processed, and selecting the connected domain with the largest density value;
判定是否找到鼻尖区域,当前区域阈值大于预定义的thr时,该区域即为鼻尖区域,否则重新开始选取。 It is determined whether the nose tip region is found. When the current region threshold is greater than the predefined thr, the region is the nose tip region, otherwise the selection is restarted.
如图4所示,利用获取的鼻尖区域进行配准,在本发明中使用ICP算法进行数据的配准,先从参考模板中的三维鼻尖数据选取参考点数据点集P,然后再利用点对点之间的最近的距离来选择输入三维人脸中与参考数据相匹配的数据点集Q,首先计算3*3的矩阵As shown in FIG. 4, the acquired nose region is used for registration. In the present invention, the ICP algorithm is used for data registration, and the reference point data point set P is first selected from the three-dimensional nose data in the reference template, and then the point-to-point point is used. The nearest distance between the two is selected to input the data point set Q in the three-dimensional face that matches the reference data, and the matrix of 3*3 is first calculated.
Figure PCTCN2015074868-appb-000024
Figure PCTCN2015074868-appb-000024
其中N是数据集合的容量,再做H矩阵的SVD分解Where N is the capacity of the data set, and then the SVD decomposition of the H matrix
H=U∧VT H=U∧V T
X=VUT X=VU T
计算旋转矩阵R和平移矩阵tCalculate the rotation matrix R and the translation matrix t
当X行列式值为1时,R=X;When the X determinant value is 1, R = X;
t=P-R*Qt=P-R*Q
判断刚性变换后的数据集RQ+t和参考数据集P之间的误差是否足够小。当该误差小于某一阈值时,则这两个三维数据集合已经实现配准;否则从第一步重新开始直到数据集合对实现配准。It is judged whether the error between the rigid transformed data set RQ+t and the reference data set P is sufficiently small. When the error is less than a certain threshold, then the two three-dimensional data sets have been registered; otherwise, the first step is restarted until the data set pair is registered.
根据上述自适应特征点采样和ICP配准算法,则距离函数如下:According to the above adaptive feature point sampling and ICP registration algorithm, the distance function is as follows:
Figure PCTCN2015074868-appb-000025
Figure PCTCN2015074868-appb-000025
其中P,Q分别是待匹配的特征点集合,集合中含有N个特征点。Where P and Q are respectively a set of feature points to be matched, and the set contains N feature points.
由于特征点采样密度的不同,因此在配准完成后计算输入数据与注册库中的三维人脸模型数据之间的欧式距离时,需要根据有效特征点的数目对该距离进行归一化。Since the feature point sampling density is different, when calculating the Euclidean distance between the input data and the 3D face model data in the registration library after the registration is completed, the distance needs to be normalized according to the number of effective feature points.
如图4所示,配准后,首先按照深度信息进行深度图像的获取,然后利用滤波器对于映射后的深度图像中的噪音点(数据突起点或者空洞点)进行补偿 去噪,最后对表情鲁棒区域进行选择,得到最终的三维人脸深度图像。As shown in FIG. 4, after registration, the depth image is first acquired according to the depth information, and then the noise point (data bump or hole point) in the mapped depth image is compensated by the filter. Denoising, finally selecting the robust region of the expression to obtain the final 3D face depth image.
如图5所示,当测试人脸图像输入后,经过Gabor滤波后,将任一滤波向量都与其所在位置相对应的视觉分词典中的所有基元词汇比较,通过距离匹配的方式,把它映射到与之距离最为接近的基元上。通过这种方式,就可以提取出原始深度图像的视觉词典直方图特征。其大致流程总结如下:As shown in FIG. 5, after the face image is input, after Gabor filtering, any filter vector is compared with all primitive vocabularies in the visual sub-dictionary corresponding to its position, and it is matched by distance matching. Map to the primitive closest to it. In this way, the visual dictionary histogram features of the original depth image can be extracted. The general process is summarized as follows:
将三维人脸深度图像分割成一些局部纹理区域;Dividing the 3D face depth image into some local texture regions;
对于每个Gabor滤波响应向量,按照位置的不同将其映射到其对应的视觉分词典的词汇中,并依此为基础建立视觉词典直方图向量作为三维人脸的特诊表达;For each Gabor filter response vector, it is mapped to the vocabulary of its corresponding visual sub-dictionary according to the position, and based on this, the visual dictionary histogram vector is established as the special diagnosis expression of the three-dimensional human face;
将最近邻分类器用来作为最后的人脸识别,其中L1距离被选作为距离度量。The nearest neighbor classifier is used as the final face recognition, where the L1 distance is chosen as the distance metric.
如图9所示,本发明中对二维灰度信息和三维深度信息采用最大最小线性归一化原则进行分数归一化,公式如下As shown in FIG. 9, in the present invention, the two-dimensional gray information and the three-dimensional depth information are subjected to fractional normalization using the principle of maximum and minimum linear normalization, and the formula is as follows
Figure PCTCN2015074868-appb-000026
Figure PCTCN2015074868-appb-000026
不同于传统的最大最小线性归一化原则,由于max代表的是在距离空间中较远的位置,因此该数值很容易受到噪音(如三维人脸的头发遮挡等)的影响,因此在对max进行取值时,是单一模态分数集合{Sk}升序排序后位于95%位置的值;而由于min代表的是在距离空间中较近的位置,因此该数值未受到数据噪音影响(此处受影响值会变大),因此min取值是单一模态分数集合{Sk}升序排序后的最小值。Sk是该模态中的匹配分数,
Figure PCTCN2015074868-appb-000027
是该模态中归一化的匹配分数;
Unlike the traditional principle of maximum and minimum linear normalization, since max represents a position farther away from the space, this value is easily affected by noise (such as 3D human face hair occlusion, etc.), so in the max When the value is taken, it is the value of the single modal score set {S k } at the 95% position after the ascending order; and since min represents the position closer in the distance space, the value is not affected by the data noise (this The affected value will become larger), so the min value is the minimum value after the single modal score set {S k } is sorted in ascending order. S k is the matching score in the modality,
Figure PCTCN2015074868-appb-000027
Is the normalized matching score in the modality;
分数归一化之后,采用比较鲁棒的加权加法原则对不同模态的匹配分数进行融合,公式如下After the score is normalized, the matching scores of different modes are merged by using the relatively robust weighted addition principle. The formula is as follows
Figure PCTCN2015074868-appb-000028
Figure PCTCN2015074868-appb-000028
得到多模态数据融合后的匹配分数。此处权值的获取采用线性判别分析算法(LDA)。该算法利用了数据的类别信息,通过构建类内散布矩阵SB和类间散布矩阵SW,最大化目标函数The matching score after multimodal data fusion is obtained. Here, the weight is obtained by a linear discriminant analysis algorithm (LDA). The algorithm utilizes the category information of the data to maximize the objective function by constructing an intra-class scatter matrix SB and an inter-class scatter matrix SW.
Figure PCTCN2015074868-appb-000029
Figure PCTCN2015074868-appb-000029
获取LDA映射矩阵W,即为权值。Obtain the LDA mapping matrix W, which is the weight.
采用本发明的方案,多模态系统通过进行二维灰度信息和三维深度信息的采集,利用二维灰度信息和三维深度信息的优点,通过融合策略来克服单模态系统的某些内在弱点(如灰度图像的光照,深度图像的表情),有效地提升了人脸识别系统的性能,使得人脸识别更加准确快捷。By adopting the scheme of the invention, the multi-modal system overcomes some inherent aspects of the single-modal system by using the advantages of two-dimensional gray information and three-dimensional depth information by utilizing the advantages of two-dimensional gray information and three-dimensional depth information. Weaknesses (such as the illumination of grayscale images and the expression of depth images) effectively improve the performance of the face recognition system, making face recognition more accurate and faster.
对于本领域技术人员而言,显然本发明不限于上述示范性实施例的细节,而且在不背离本发明的精神或基本特征的情况下,能够以其他的具体形式实现本发明。因此,无论从哪一点来看,均应将实施例看作是示范性的,而且是非限制性的,本发明的范围由所附权利要求而不是上述说明限定,因此旨在将落在权利要求的等同要件的含义和范围内的所有变化囊括在本发明内。不应将权利要求中的任何附图标记视为限制所涉及的权利要求。It is apparent to those skilled in the art that the present invention is not limited to the details of the above-described exemplary embodiments, and the present invention can be embodied in other specific forms without departing from the spirit or essential characteristics of the invention. Therefore, the present embodiments are to be considered as illustrative and not restrictive, and the scope of the invention is defined by the appended claims instead All changes in the meaning and scope of equivalent elements are included in the present invention. Any reference signs in the claims should not be construed as limiting the claim.
此外,应当理解,虽然本说明书按照实施方式加以描述,但并非每个实施方式仅包含一个独立的技术方案,说明书的这种叙述方式仅仅是为清楚起见,本领域技术人员应当将说明书作为一个整体,各实施例中的技术方案也可以经适当组合,形成本领域技术人员可以理解的其他实施方式。 In addition, it should be understood that although the description is described in terms of embodiments, not every embodiment includes only one independent technical solution. The description of the specification is merely for the sake of clarity, and those skilled in the art should regard the specification as a whole. The technical solutions in the respective embodiments may also be combined as appropriate to form other embodiments that can be understood by those skilled in the art.

Claims (10)

  1. 一种基于灰度和深度信息的多层融合的多模态人脸识别装置,其特征在于:包括对于灰度信息进行人脸识别的计算单元;用于对深度信息进行人脸识别的计算单元;基于多模态人脸识别分数进行融合的计算单元;对数据进行分类的分类器计算单元。A multi-layer fusion multi-modal face recognition device based on gray level and depth information, comprising: a calculation unit for performing face recognition on gray information; and a calculation unit for performing face recognition on depth information a computing unit that fuses based on multimodal face recognition scores; a classifier computing unit that classifies data.
  2. 根据权利要求1所述的一种基于灰度和深度信息的多层融合的多模态人脸识别装置,其特征在于,所述对于灰度信息进行人脸识别的计算单元包括:人眼检测单元、二维数据配准计算单元、灰度人脸特征提取单元和灰度人脸识别分数计算单元。A multi-modal face recognition apparatus based on gradation and depth information based on multi-layer fusion according to claim 1, wherein the calculation unit for performing face recognition on gradation information comprises: human eye detection The unit, the two-dimensional data registration calculation unit, the grayscale face feature extraction unit, and the grayscale face recognition score calculation unit.
  3. 根据权利要求1所述的一种基于灰度和深度信息的多层融合的多模态人脸识别装置,其特征在于,多数对深度信息进行人脸识别的计算单元包括:鼻尖检测器单元、三维数据配准计算单元、深度人脸特征提取单元和深度人脸识别分数计算单元。A multi-modal face recognition apparatus based on gradation and depth information based on multi-layer fusion according to claim 1, wherein a plurality of calculation units for performing face recognition on depth information include: a nose tip detector unit, A three-dimensional data registration calculation unit, a deep face feature extraction unit, and a deep face recognition score calculation unit.
  4. 一种基于灰度和深度信息的多层融合的多模态人脸识别方法,其特征在于,包括如下步骤:A multi-layer fusion multi-modal face recognition method based on gray level and depth information, comprising the following steps:
    A.对人脸灰度信息进行识别;A. Identify the face grayscale information;
    B.对人脸深度信息进行识别;B. Identify face depth information;
    C.对人脸灰度信息及深度信息进行归一化,基于归一化的匹配分数,采用融合策略得到多模态融合后的匹配分数,实现多模态人脸识别。C. The face gray information and depth information are normalized. Based on the normalized matching score, the fusion score is used to obtain the multi-modal fusion matching score to realize multi-modal face recognition.
  5. 根据权利要求4所述的一种基于灰度和深度信息的多层融合的多模态 人脸识别方法,其特征在于,所述步骤A包括如下步骤:Multi-mode fusion multi-modality based on gray level and depth information according to claim 4 The face recognition method is characterized in that the step A includes the following steps:
    A1.特征区域定位,使用人眼检测器获取人眼区域,所述人眼检测器为层级分类器H,经如下算法得到:A1. Feature area localization, using a human eye detector to acquire a human eye region, the human eye detector being a hierarchical classifier H, obtained by the following algorithm:
    给定训练样本集合S={(x1,y1),...,(xm,ym)},弱空间分类器
    Figure PCTCN2015074868-appb-100001
    其中xi∈χ,为样本向量,yi=±1,为分类标签,m为样本总数;初始化样本概率分布
    Figure PCTCN2015074868-appb-100002
    Given a set of training samples S = {(x 1 , y 1 ), ..., (x m , y m )}, weak space classifier
    Figure PCTCN2015074868-appb-100001
    Where x i ∈χ is the sample vector, y i =±1, is the classification label, m is the total number of samples; initial sample probability distribution
    Figure PCTCN2015074868-appb-100002
    t-1,...,T,对中的每个弱分类器h作如下操作:T-1,...,T, each weak classifier h of the pair performs the following operations:
    对样本空间χ进行划分,得到X1,X2,...,XnThe sample space χ is divided to obtain X 1 , X 2 ,..., X n ;
    Figure PCTCN2015074868-appb-100003
    其中ε为一小正常数;
    Figure PCTCN2015074868-appb-100003
    Where ε is a small normal number;
    计算归一化因子,
    Figure PCTCN2015074868-appb-100004
    Calculate the normalization factor,
    Figure PCTCN2015074868-appb-100004
    在弱分类器空间中选择一个ht,使得Z最小化Select an h t in the weak classifier space to minimize Z
    Figure PCTCN2015074868-appb-100005
    Figure PCTCN2015074868-appb-100005
    更新训练样本概率分布
    Figure PCTCN2015074868-appb-100006
    i=1,...,m,其中
    Figure PCTCN2015074868-appb-100007
    为归一化因子,使得Dt+1为一个概率分布;
    Update training sample probability distribution
    Figure PCTCN2015074868-appb-100006
    i=1,...,m, where
    Figure PCTCN2015074868-appb-100007
    Is a normalization factor such that D t+1 is a probability distribution;
    最终强分类器H为
    Figure PCTCN2015074868-appb-100008
    The final strong classifier H is
    Figure PCTCN2015074868-appb-100008
    A2.使用获得的人眼区域位置进行配准,利用LBP算法处理人眼位置数据获得LBP直方图特征,取值公式为 A2. Using the obtained position of the human eye region for registration, the LBP algorithm is used to process the human eye position data to obtain the LBP histogram feature, and the value formula is
    Figure PCTCN2015074868-appb-100009
    Figure PCTCN2015074868-appb-100009
    将该特征输入灰度图像分类器获取灰度匹配分数。The feature is input to the grayscale image classifier to obtain a grayscale matching score.
  6. 根据权利要求4所述的一种基于弧度和深度信息的多层融合的多模态人脸识别方法,其特征在于,所述步骤B包括如下步骤:The method of claim 4, wherein the step B comprises the following steps:
    B1.特征区域定位,判定人脸鼻尖区域位置;B1. The feature area is positioned to determine the position of the face of the face;
    B2.对于不同姿态的三维数据,得到配准的参考区域后,按照ICP算法进行数据的配准,配准完成后计算输入数据与注册库中的三维人脸模型数据之间的欧式距离;B2. For the three-dimensional data of different postures, after the registration reference area is obtained, the data is registered according to the ICP algorithm, and the Euclidean distance between the input data and the three-dimensional face model data in the registration library is calculated after the registration is completed;
    B3.按照深度信息进行深度图像的获取,利用滤波器对于映射后的深度图像中的噪音点进行补偿去噪,最后对表情鲁棒区域进行选择,得到最终的三维人脸深度图像;B3. Obtain the depth image according to the depth information, and use the filter to compensate and denoise the noise points in the mapped depth image, and finally select the robust region of the expression to obtain the final 3D face depth image;
    B4.提取是三维深度图像的视觉词典直方图特征向量,当测试人脸图像输入后,经过Gabor滤波后,将任一滤波向量都与其所在位置相对应的视觉分词典中的所有基元词汇比较,通过距离匹配的方式,把它映射到与之距离最为接近的基元上,提取出原始深度图像的视觉词典直方图特征,利用该特征输入深度图像分类器获取匹配分数。B4. Extraction is a visual dictionary histogram feature vector of a three-dimensional depth image. After the face image is input, after Gabor filtering, any filter vector is compared with all primitive words in the visual sub-dictionary corresponding to its position. By distance matching, it maps to the primitive closest to the distance, extracts the visual dictionary histogram feature of the original depth image, and uses the feature input depth image classifier to obtain the matching score.
  7. 根据权利要求4所述的一种基于灰度和深度信息的多层融合的多模态人脸识别方法,其特征在于,所述步骤c具体包括:The multi-modal face recognition method based on the multi-layer fusion of the gradation and the depth information according to claim 4, wherein the step c specifically includes:
    对二维灰度信息和三维深度信息采用最大最小线性归一化原则进行分数 归一化,公式如下Fractionation of two-dimensional gray information and three-dimensional depth information using the principle of maximum and minimum linear normalization Normalized, the formula is as follows
    Figure PCTCN2015074868-appb-100010
    Figure PCTCN2015074868-appb-100010
    分数归一化之后,采用比较鲁棒的加权加法原则对不同模态的匹配分数进行融合,公式如下After the score is normalized, the matching scores of different modes are merged by using the relatively robust weighted addition principle. The formula is as follows
    Figure PCTCN2015074868-appb-100011
    Figure PCTCN2015074868-appb-100011
    得到多模态数据融合后的匹配分数后采用线性判别分析算法通过构建类内散布矩阵SB和类间散布矩阵SW,最大化目标函数After obtaining the matching scores of the multi-modal data fusion, the linear discriminant analysis algorithm is used to maximize the objective function by constructing the intra-class scatter matrix SB and the inter-class scatter matrix SW.
    Figure PCTCN2015074868-appb-100012
    Figure PCTCN2015074868-appb-100012
    获取LDA映射矩阵W,即为权值。Obtain the LDA mapping matrix W, which is the weight.
  8. 根据权利要求6所述的一种基于灰度和深度信息的多层融合的多模态人脸识别方法,其特征在于,所述步骤B1具体包括A multi-modal face recognition method based on gradation and depth information based on multi-layer fusion according to claim 6, wherein the step B1 specifically comprises
    步骤1:确定阈值,确定域平均负有效能量密度的阈值,定义为thr;Step 1: determining a threshold, determining a threshold of the average average effective energy density of the domain, defined as thr;
    步骤2:利用深度信息选取待处理数据,利用数据的深度信息,提取在一定深度范围内的人脸数据作为待处理数据;Step 2: using the depth information to select the data to be processed, and using the depth information of the data to extract the face data in a certain depth range as the data to be processed;
    步骤3:法向量的计算,计算由深度信息选取出的人脸数据的方向量信息;Step 3: Calculating the normal vector, and calculating the direction quantity information of the face data selected by the depth information;
    步骤4:区域平均负有效能量密度的计算,按照区域平均负有效能量密度 的定义,求出待处理数据中个连通域的平均负有效能量密度,选择其中密度值最大的连通域;Step 4: Calculate the average negative effective energy density of the region, according to the regional average negative effective energy density The definition, find the average negative effective energy density of the connected domains in the data to be processed, and select the connected domain with the largest density value;
    步骤5:判定是否找到鼻尖区域,当前区域阈值大于预定义的thr时,该区域即为鼻尖区域,否则回到步骤1重新开始循环。Step 5: Determine whether the nose tip area is found. When the current area threshold is greater than the predefined thr, the area is the nose tip area, otherwise return to step 1 to restart the cycle.
  9. 根据权利要求6所述的一种基于灰度和深度信息的多层融合的多模态人脸识别方法,其特征在于,所述ICP算法主要步骤包括:The multi-modal face recognition method based on gradation and depth information based on multi-layer fusion according to claim 6, wherein the main steps of the ICP algorithm include:
    确定匹配数据集合对,从参考模板中的三维鼻尖数据选取参考数据点集P,再利用点对点之间的最近的距离来选择输入三维人脸中与参考数据相匹配的数据点集Q;Determining a matching data set pair, selecting a reference data point set P from the three-dimensional nose point data in the reference template, and then using the closest distance between the point-to-point to select a data point set Q of the input three-dimensional face that matches the reference data;
    计算刚性运动参数,计算旋转矩阵R和平移向量tCalculate the rigid motion parameters and calculate the rotation matrix R and the translation vector t
    当X行列式值为1时,R=X;When the X determinant value is 1, R = X;
    t=P-R*Qt=P-R*Q
    根据刚性变换后的数据集RQ+t和参考数据集P之间的误差判断三维数据集是否配准,配准之后通过下式计算输入数据与注册库中的三维人脸模型数据之间的欧式距离According to the error between the rigid transformed data set RQ+t and the reference data set P, whether the three-dimensional data set is registered or not, after registration, the European type between the input data and the three-dimensional face model data in the registration library is calculated by the following formula: distance
    Figure PCTCN2015074868-appb-100013
    Figure PCTCN2015074868-appb-100013
    其中P,Q分别是待匹配的特征点集合,集合中含有N个特征点。Where P and Q are respectively a set of feature points to be matched, and the set contains N feature points.
  10. 根据权利要求6所述的一种基于灰度和深度信息的多层融合的多模态人脸识别方法,其特征在于,步骤B4具体为: The multi-modal face recognition method based on the gradation and depth information of the multi-layer fusion according to claim 6, wherein the step B4 is specifically:
    将三维人脸深度图像分割成一些局部纹理区域;Dividing the 3D face depth image into some local texture regions;
    对于每个Gabor滤波响应向量,按照位置的不同将其映射到其对应的视觉分词典的词汇中,并依此为基础建立视觉词典直方图向量作为三维人脸的特诊表达;For each Gabor filter response vector, it is mapped to the vocabulary of its corresponding visual sub-dictionary according to the position, and based on this, the visual dictionary histogram vector is established as the special diagnosis expression of the three-dimensional human face;
    将最近邻分类器用来获取三维人脸识别的识别分数,其中L1距离被选作为距离度量。 The nearest neighbor classifier is used to obtain the recognition score of the three-dimensional face recognition, wherein the L1 distance is selected as the distance metric.
PCT/CN2015/074868 2015-01-07 2015-03-23 Gray level and depth information based multi-layer fusion multi-modal face recognition device and method WO2016110005A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201510006214.4A CN104598878A (en) 2015-01-07 2015-01-07 Multi-modal face recognition device and method based on multi-layer fusion of gray level and depth information
CN2015100062144 2015-01-07

Publications (1)

Publication Number Publication Date
WO2016110005A1 true WO2016110005A1 (en) 2016-07-14

Family

ID=53124652

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/074868 WO2016110005A1 (en) 2015-01-07 2015-03-23 Gray level and depth information based multi-layer fusion multi-modal face recognition device and method

Country Status (2)

Country Link
CN (1) CN104598878A (en)
WO (1) WO2016110005A1 (en)

Cited By (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107392137A (en) * 2017-07-18 2017-11-24 艾普柯微电子(上海)有限公司 Face identification method and device
CN109035388A (en) * 2018-06-28 2018-12-18 北京的卢深视科技有限公司 Three-dimensional face model method for reconstructing and device
CN109191611A (en) * 2018-10-30 2019-01-11 惠州学院 A kind of Time Attendance Device and method based on recognition of face
CN109902590A (en) * 2019-01-30 2019-06-18 西安理工大学 Pedestrian's recognition methods again of depth multiple view characteristic distance study
CN110033007A (en) * 2019-04-19 2019-07-19 福州大学 Attribute recognition approach is worn clothes based on the pedestrian of depth attitude prediction and multiple features fusion
CN110046587A (en) * 2019-04-22 2019-07-23 安徽理工大学 Human face expression feature extracting method based on Gabor difference weight
CN110046559A (en) * 2019-03-28 2019-07-23 广东工业大学 A kind of face identification method
CN110084266A (en) * 2019-03-11 2019-08-02 中国地质大学(武汉) A kind of dynamic emotion identification method based on audiovisual features depth integration
CN110163049A (en) * 2018-07-18 2019-08-23 腾讯科技(深圳)有限公司 A kind of face character prediction technique, device and storage medium
CN110232378A (en) * 2019-05-30 2019-09-13 苏宁易购集团股份有限公司 A kind of image interest point detecting method, system and readable storage medium storing program for executing
CN110472479A (en) * 2019-06-28 2019-11-19 广州中国科学院先进技术研究所 A kind of finger vein identification method based on SURF feature point extraction and part LBP coding
CN110472582A (en) * 2019-08-16 2019-11-19 腾讯科技(深圳)有限公司 3D face identification method, device and terminal based on eye recognition
CN110517185A (en) * 2019-07-23 2019-11-29 北京达佳互联信息技术有限公司 Image processing method, device, electronic equipment and storage medium
CN110827394A (en) * 2018-08-10 2020-02-21 宏达国际电子股份有限公司 Facial expression construction method and device and non-transitory computer readable recording medium
CN110969089A (en) * 2019-11-01 2020-04-07 北京交通大学 Lightweight face recognition system and recognition method under noise environment
CN111046770A (en) * 2019-12-05 2020-04-21 上海中信信息发展股份有限公司 Automatic annotation method for photo file figures
CN111046708A (en) * 2018-10-15 2020-04-21 天津大学青岛海洋技术研究院 Human face gender discrimination algorithm based on Wasserstein distance
CN111062324A (en) * 2019-12-17 2020-04-24 上海眼控科技股份有限公司 Face detection method and device, computer equipment and storage medium
CN111079684A (en) * 2019-12-24 2020-04-28 河南中原大数据研究院有限公司 Three-dimensional face detection method based on rough-fine fitting
CN111079700A (en) * 2019-12-30 2020-04-28 河南中原大数据研究院有限公司 Three-dimensional face recognition method based on fusion of multiple data types
CN111160292A (en) * 2019-12-31 2020-05-15 上海易维视科技有限公司 Human eye detection method
CN111160208A (en) * 2019-12-24 2020-05-15 河南中原大数据研究院有限公司 Three-dimensional face super-resolution method based on multi-frame point cloud fusion and variable model
CN111178129A (en) * 2019-11-25 2020-05-19 浙江工商大学 Multi-modal personnel identification method based on face and posture
CN111242992A (en) * 2020-01-13 2020-06-05 洛阳理工学院 Image registration method
CN111460864A (en) * 2019-01-22 2020-07-28 天津大学青岛海洋技术研究院 Animal disease detection method based on image recognition
CN111476175A (en) * 2020-04-09 2020-07-31 上海看看智能科技有限公司 Adaptive topological graph matching method and system suitable for old people face comparison
CN111488856A (en) * 2020-04-28 2020-08-04 江西吉为科技有限公司 Multimode 2D and 3D facial expression recognition based on orthogonal guide learning
CN111695498A (en) * 2020-06-10 2020-09-22 西南林业大学 Wood identity detection method
CN111753877A (en) * 2020-05-19 2020-10-09 海克斯康制造智能技术(青岛)有限公司 Product quality detection method based on deep neural network transfer learning
CN111950389A (en) * 2020-07-22 2020-11-17 重庆邮电大学 Depth binary feature facial expression recognition method based on lightweight network
CN112001219A (en) * 2020-06-19 2020-11-27 国家电网有限公司技术学院分公司 Multi-angle multi-face recognition attendance checking method and system
CN112016524A (en) * 2020-09-25 2020-12-01 北京百度网讯科技有限公司 Model training method, face recognition device, face recognition equipment and medium
CN112069981A (en) * 2020-09-03 2020-12-11 Oppo广东移动通信有限公司 Image classification method and device, electronic equipment and storage medium
CN112215064A (en) * 2020-09-03 2021-01-12 广州市标准化研究院 Face recognition method and system for public safety precaution
CN112329683A (en) * 2020-11-16 2021-02-05 常州大学 Attention mechanism fusion-based multi-channel convolutional neural network facial expression recognition method
CN112364711A (en) * 2020-10-20 2021-02-12 盛视科技股份有限公司 3D face recognition method, device and system
CN112528902A (en) * 2020-12-17 2021-03-19 四川大学 Video monitoring dynamic face recognition method and device based on 3D face model
CN112634269A (en) * 2021-01-14 2021-04-09 华东交通大学 Rail vehicle body detection method
CN112652002A (en) * 2020-12-25 2021-04-13 江苏集萃复合材料装备研究所有限公司 Medical image registration method based on IDC algorithm
CN112766173A (en) * 2021-01-21 2021-05-07 福建天泉教育科技有限公司 Multi-mode emotion analysis method and system based on AI deep learning
CN112883914A (en) * 2021-03-19 2021-06-01 西安科技大学 Mining robot idea perception and decision method combining multiple classifiers
CN113449674A (en) * 2021-07-12 2021-09-28 江苏商贸职业学院 Pig face identification method and system
CN113616184A (en) * 2021-06-30 2021-11-09 北京师范大学 Brain network modeling and individual prediction method based on multi-mode magnetic resonance image
CN113643348A (en) * 2020-04-23 2021-11-12 杭州海康威视数字技术股份有限公司 Face attribute analysis method and device
CN113705457A (en) * 2021-08-30 2021-11-26 支付宝(杭州)信息技术有限公司 Service processing method and device based on human face
CN114283471A (en) * 2021-12-16 2022-04-05 武汉大学 Multi-modal sequencing optimization method for heterogeneous face image re-recognition
CN115147578A (en) * 2022-06-30 2022-10-04 北京百度网讯科技有限公司 Stylized three-dimensional face generation method and device, electronic equipment and storage medium
CN115496975A (en) * 2022-08-29 2022-12-20 锋睿领创(珠海)科技有限公司 Auxiliary weighted data fusion method, device, equipment and storage medium
CN116362933A (en) * 2023-05-30 2023-06-30 南京农业大学 Intelligent campus management method and system based on big data
CN116403270A (en) * 2023-06-07 2023-07-07 南昌航空大学 Facial expression recognition method and system based on multi-feature fusion
CN112652002B (en) * 2020-12-25 2024-05-03 江苏集萃复合材料装备研究所有限公司 Medical image registration method based on IDC algorithm

Families Citing this family (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106909873B (en) * 2016-06-21 2019-11-05 湖南拓视觉信息技术有限公司 The method and apparatus of recognition of face
CN106156740B (en) * 2016-07-05 2019-06-28 张宁 Civil Aviation Airport terminal face system for rapidly identifying
CN106326867B (en) * 2016-08-26 2019-06-07 维沃移动通信有限公司 A kind of method and mobile terminal of recognition of face
CN106469465A (en) * 2016-08-31 2017-03-01 深圳市唯特视科技有限公司 A kind of three-dimensional facial reconstruction method based on gray scale and depth information
CN106407914B (en) * 2016-08-31 2019-12-10 北京旷视科技有限公司 Method and device for detecting human face and remote teller machine system
CN106650693B (en) * 2016-12-30 2020-07-03 河北三川科技有限公司 Multi-feature fusion recognition algorithm for face comparison
CN107301377B (en) * 2017-05-26 2020-08-18 浙江大学 Face and pedestrian sensing system based on depth camera
CN109299639B (en) * 2017-07-25 2021-03-16 虹软科技股份有限公司 Method and device for facial expression recognition
CN110033291A (en) * 2018-01-12 2019-07-19 北京京东金融科技控股有限公司 Information object method for pushing, device and system
CN108197587B (en) * 2018-01-18 2021-08-03 中科视拓(北京)科技有限公司 Method for performing multi-mode face recognition through face depth prediction
CN108960173A (en) * 2018-07-12 2018-12-07 芜湖博高光电科技股份有限公司 A kind of millimeter wave and camera merge face identification method
CN109143260A (en) * 2018-09-29 2019-01-04 北京理工大学 A kind of three-dimensional solid-state face battle array laser radar face identification device and method
CN110287776B (en) * 2019-05-15 2020-06-26 北京邮电大学 Face recognition method and device and computer readable storage medium
CN110532907B (en) * 2019-08-14 2022-01-21 中国科学院自动化研究所 Traditional Chinese medicine human body constitution classification method based on face image and tongue image bimodal feature extraction
CN110781856B (en) * 2019-11-04 2023-12-19 浙江大华技术股份有限公司 Heterogeneous face recognition model training method, face recognition method and related device
CN110967678A (en) * 2019-12-20 2020-04-07 安徽博微长安电子有限公司 Data fusion algorithm and system for multiband radar target identification
CN112767303B (en) * 2020-08-12 2023-11-28 腾讯科技(深圳)有限公司 Image detection method, device, equipment and computer readable storage medium
CN112102409B (en) * 2020-09-21 2023-09-01 杭州海康威视数字技术股份有限公司 Target detection method, device, equipment and storage medium
CN112215136B (en) * 2020-10-10 2023-09-05 北京奇艺世纪科技有限公司 Target person identification method and device, electronic equipment and storage medium
CN112364825B (en) * 2020-11-30 2022-10-25 支付宝(杭州)信息技术有限公司 Method, apparatus and computer-readable storage medium for face recognition
CN112926543A (en) * 2021-04-09 2021-06-08 商汤集团有限公司 Image generation method, three-dimensional model generation method, image generation device, three-dimensional model generation device, electronic device, and medium
CN113780222B (en) * 2021-09-17 2024-02-27 深圳市繁维科技有限公司 Face living body detection method and device, electronic equipment and readable storage medium
CN113837105A (en) * 2021-09-26 2021-12-24 北京的卢深视科技有限公司 Face recognition method, face recognition system, electronic equipment and storage medium
CN114093012B (en) * 2022-01-18 2022-06-10 荣耀终端有限公司 Face shielding detection method and detection device
CN116386121B (en) * 2023-05-30 2023-08-11 湖北华中电力科技开发有限责任公司 Personnel identification method and device based on power grid safety production

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101339607A (en) * 2008-08-15 2009-01-07 北京中星微电子有限公司 Human face recognition method and system, human face recognition model training method and system
US20130322708A1 (en) * 2012-06-04 2013-12-05 Sony Mobile Communications Ab Security by z-face detection
CN103679151A (en) * 2013-12-19 2014-03-26 成都品果科技有限公司 LBP and Gabor characteristic fused face clustering method
CN103810491A (en) * 2014-02-19 2014-05-21 北京工业大学 Head posture estimation interest point detection method fusing depth and gray scale image characteristic points
US20140294295A1 (en) * 2009-06-05 2014-10-02 Samsung Electronics Co., Ltd. Apparatus and method for video sensor-based human activity and facial expression modeling and recognition

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8285006B2 (en) * 2007-04-13 2012-10-09 Mira Electronics Co., Ltd. Human face recognition and user interface system for digital camera and video camera
CN103390164B (en) * 2012-05-10 2017-03-29 南京理工大学 Method for checking object based on depth image and its realize device
CN103164695B (en) * 2013-02-26 2016-02-17 中国农业大学 A kind of fruit identification method based on multi-source image information fusion
CN103198605A (en) * 2013-03-11 2013-07-10 成都百威讯科技有限责任公司 Indoor emergent abnormal event alarm system
CN103971122B (en) * 2014-04-30 2018-04-17 深圳市唯特视科技有限公司 Three-dimensional face based on depth image describes method
CN104143080B (en) * 2014-05-21 2017-06-23 深圳市唯特视科技有限公司 Three-dimensional face identifying device and method based on three-dimensional point cloud

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101339607A (en) * 2008-08-15 2009-01-07 北京中星微电子有限公司 Human face recognition method and system, human face recognition model training method and system
US20140294295A1 (en) * 2009-06-05 2014-10-02 Samsung Electronics Co., Ltd. Apparatus and method for video sensor-based human activity and facial expression modeling and recognition
US20130322708A1 (en) * 2012-06-04 2013-12-05 Sony Mobile Communications Ab Security by z-face detection
CN103679151A (en) * 2013-12-19 2014-03-26 成都品果科技有限公司 LBP and Gabor characteristic fused face clustering method
CN103810491A (en) * 2014-02-19 2014-05-21 北京工业大学 Head posture estimation interest point detection method fusing depth and gray scale image characteristic points

Cited By (86)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107392137B (en) * 2017-07-18 2020-09-08 艾普柯微电子(上海)有限公司 Face recognition method and device
CN107392137A (en) * 2017-07-18 2017-11-24 艾普柯微电子(上海)有限公司 Face identification method and device
CN109035388A (en) * 2018-06-28 2018-12-18 北京的卢深视科技有限公司 Three-dimensional face model method for reconstructing and device
CN109035388B (en) * 2018-06-28 2023-12-05 合肥的卢深视科技有限公司 Three-dimensional face model reconstruction method and device
CN110163049A (en) * 2018-07-18 2019-08-23 腾讯科技(深圳)有限公司 A kind of face character prediction technique, device and storage medium
CN110163049B (en) * 2018-07-18 2023-08-29 腾讯科技(深圳)有限公司 Face attribute prediction method, device and storage medium
CN110827394B (en) * 2018-08-10 2024-04-02 宏达国际电子股份有限公司 Facial expression construction method, device and non-transitory computer readable recording medium
CN110827394A (en) * 2018-08-10 2020-02-21 宏达国际电子股份有限公司 Facial expression construction method and device and non-transitory computer readable recording medium
CN111046708A (en) * 2018-10-15 2020-04-21 天津大学青岛海洋技术研究院 Human face gender discrimination algorithm based on Wasserstein distance
CN109191611A (en) * 2018-10-30 2019-01-11 惠州学院 A kind of Time Attendance Device and method based on recognition of face
CN111460864A (en) * 2019-01-22 2020-07-28 天津大学青岛海洋技术研究院 Animal disease detection method based on image recognition
CN111460864B (en) * 2019-01-22 2023-10-17 天津大学青岛海洋技术研究院 Animal disease detection method based on image recognition
CN109902590B (en) * 2019-01-30 2022-09-16 西安理工大学 Pedestrian re-identification method for deep multi-view characteristic distance learning
CN109902590A (en) * 2019-01-30 2019-06-18 西安理工大学 Pedestrian's recognition methods again of depth multiple view characteristic distance study
CN110084266A (en) * 2019-03-11 2019-08-02 中国地质大学(武汉) A kind of dynamic emotion identification method based on audiovisual features depth integration
CN110046559A (en) * 2019-03-28 2019-07-23 广东工业大学 A kind of face identification method
CN110033007A (en) * 2019-04-19 2019-07-19 福州大学 Attribute recognition approach is worn clothes based on the pedestrian of depth attitude prediction and multiple features fusion
CN110033007B (en) * 2019-04-19 2022-08-09 福州大学 Pedestrian clothing attribute identification method based on depth attitude estimation and multi-feature fusion
CN110046587A (en) * 2019-04-22 2019-07-23 安徽理工大学 Human face expression feature extracting method based on Gabor difference weight
CN110046587B (en) * 2019-04-22 2022-11-25 安徽理工大学 Facial expression feature extraction method based on Gabor differential weight
CN110232378B (en) * 2019-05-30 2023-01-20 苏宁易购集团股份有限公司 Image interest point detection method and system and readable storage medium
CN110232378A (en) * 2019-05-30 2019-09-13 苏宁易购集团股份有限公司 A kind of image interest point detecting method, system and readable storage medium storing program for executing
CN110472479A (en) * 2019-06-28 2019-11-19 广州中国科学院先进技术研究所 A kind of finger vein identification method based on SURF feature point extraction and part LBP coding
CN110472479B (en) * 2019-06-28 2022-11-22 广州中国科学院先进技术研究所 Finger vein identification method based on SURF feature point extraction and local LBP coding
CN110517185A (en) * 2019-07-23 2019-11-29 北京达佳互联信息技术有限公司 Image processing method, device, electronic equipment and storage medium
CN110517185B (en) * 2019-07-23 2024-02-09 北京达佳互联信息技术有限公司 Image processing method, device, electronic equipment and storage medium
CN110472582A (en) * 2019-08-16 2019-11-19 腾讯科技(深圳)有限公司 3D face identification method, device and terminal based on eye recognition
CN110472582B (en) * 2019-08-16 2023-07-21 腾讯科技(深圳)有限公司 3D face recognition method and device based on eye recognition and terminal
CN110969089B (en) * 2019-11-01 2023-08-18 北京交通大学 Lightweight face recognition system and recognition method in noise environment
CN110969089A (en) * 2019-11-01 2020-04-07 北京交通大学 Lightweight face recognition system and recognition method under noise environment
CN111178129B (en) * 2019-11-25 2023-07-14 浙江工商大学 Multi-mode personnel identification method based on human face and gesture
CN111178129A (en) * 2019-11-25 2020-05-19 浙江工商大学 Multi-modal personnel identification method based on face and posture
CN111046770B (en) * 2019-12-05 2023-08-01 上海信联信息发展股份有限公司 Automatic labeling method for photo archive characters
CN111046770A (en) * 2019-12-05 2020-04-21 上海中信信息发展股份有限公司 Automatic annotation method for photo file figures
CN111062324A (en) * 2019-12-17 2020-04-24 上海眼控科技股份有限公司 Face detection method and device, computer equipment and storage medium
CN111160208A (en) * 2019-12-24 2020-05-15 河南中原大数据研究院有限公司 Three-dimensional face super-resolution method based on multi-frame point cloud fusion and variable model
CN111160208B (en) * 2019-12-24 2023-04-07 陕西西图数联科技有限公司 Three-dimensional face super-resolution method based on multi-frame point cloud fusion and variable model
CN111079684B (en) * 2019-12-24 2023-04-07 陕西西图数联科技有限公司 Three-dimensional face detection method based on rough-fine fitting
CN111079684A (en) * 2019-12-24 2020-04-28 河南中原大数据研究院有限公司 Three-dimensional face detection method based on rough-fine fitting
CN111079700A (en) * 2019-12-30 2020-04-28 河南中原大数据研究院有限公司 Three-dimensional face recognition method based on fusion of multiple data types
CN111079700B (en) * 2019-12-30 2023-04-07 陕西西图数联科技有限公司 Three-dimensional face recognition method based on fusion of multiple data types
CN111160292A (en) * 2019-12-31 2020-05-15 上海易维视科技有限公司 Human eye detection method
CN111160292B (en) * 2019-12-31 2023-09-22 上海易维视科技有限公司 Human eye detection method
CN111242992A (en) * 2020-01-13 2020-06-05 洛阳理工学院 Image registration method
CN111476175A (en) * 2020-04-09 2020-07-31 上海看看智能科技有限公司 Adaptive topological graph matching method and system suitable for old people face comparison
CN113643348A (en) * 2020-04-23 2021-11-12 杭州海康威视数字技术股份有限公司 Face attribute analysis method and device
CN113643348B (en) * 2020-04-23 2024-02-06 杭州海康威视数字技术股份有限公司 Face attribute analysis method and device
CN111488856A (en) * 2020-04-28 2020-08-04 江西吉为科技有限公司 Multimode 2D and 3D facial expression recognition based on orthogonal guide learning
CN111488856B (en) * 2020-04-28 2023-04-18 江西吉为科技有限公司 Multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning
CN111753877A (en) * 2020-05-19 2020-10-09 海克斯康制造智能技术(青岛)有限公司 Product quality detection method based on deep neural network transfer learning
CN111753877B (en) * 2020-05-19 2024-03-05 海克斯康制造智能技术(青岛)有限公司 Product quality detection method based on deep neural network migration learning
CN111695498A (en) * 2020-06-10 2020-09-22 西南林业大学 Wood identity detection method
CN112001219B (en) * 2020-06-19 2024-02-09 国家电网有限公司技术学院分公司 Multi-angle multi-face recognition attendance checking method and system
CN112001219A (en) * 2020-06-19 2020-11-27 国家电网有限公司技术学院分公司 Multi-angle multi-face recognition attendance checking method and system
CN111950389A (en) * 2020-07-22 2020-11-17 重庆邮电大学 Depth binary feature facial expression recognition method based on lightweight network
CN112069981A (en) * 2020-09-03 2020-12-11 Oppo广东移动通信有限公司 Image classification method and device, electronic equipment and storage medium
CN112215064A (en) * 2020-09-03 2021-01-12 广州市标准化研究院 Face recognition method and system for public safety precaution
CN112016524A (en) * 2020-09-25 2020-12-01 北京百度网讯科技有限公司 Model training method, face recognition device, face recognition equipment and medium
CN112016524B (en) * 2020-09-25 2023-08-08 北京百度网讯科技有限公司 Model training method, face recognition device, equipment and medium
CN112364711A (en) * 2020-10-20 2021-02-12 盛视科技股份有限公司 3D face recognition method, device and system
CN112329683B (en) * 2020-11-16 2024-01-26 常州大学 Multi-channel convolutional neural network facial expression recognition method
CN112329683A (en) * 2020-11-16 2021-02-05 常州大学 Attention mechanism fusion-based multi-channel convolutional neural network facial expression recognition method
CN112528902A (en) * 2020-12-17 2021-03-19 四川大学 Video monitoring dynamic face recognition method and device based on 3D face model
CN112652002A (en) * 2020-12-25 2021-04-13 江苏集萃复合材料装备研究所有限公司 Medical image registration method based on IDC algorithm
CN112652002B (en) * 2020-12-25 2024-05-03 江苏集萃复合材料装备研究所有限公司 Medical image registration method based on IDC algorithm
CN112634269B (en) * 2021-01-14 2023-12-26 华东交通大学 Railway vehicle body detection method
CN112634269A (en) * 2021-01-14 2021-04-09 华东交通大学 Rail vehicle body detection method
CN112766173B (en) * 2021-01-21 2023-08-04 福建天泉教育科技有限公司 Multi-mode emotion analysis method and system based on AI deep learning
CN112766173A (en) * 2021-01-21 2021-05-07 福建天泉教育科技有限公司 Multi-mode emotion analysis method and system based on AI deep learning
CN112883914A (en) * 2021-03-19 2021-06-01 西安科技大学 Mining robot idea perception and decision method combining multiple classifiers
CN112883914B (en) * 2021-03-19 2024-03-19 西安科技大学 Multi-classifier combined mining robot idea sensing and decision making method
CN113616184B (en) * 2021-06-30 2023-10-24 北京师范大学 Brain network modeling and individual prediction method based on multi-mode magnetic resonance image
CN113616184A (en) * 2021-06-30 2021-11-09 北京师范大学 Brain network modeling and individual prediction method based on multi-mode magnetic resonance image
CN113449674B (en) * 2021-07-12 2022-09-30 江苏商贸职业学院 Pig face identification method and system
CN113449674A (en) * 2021-07-12 2021-09-28 江苏商贸职业学院 Pig face identification method and system
CN113705457A (en) * 2021-08-30 2021-11-26 支付宝(杭州)信息技术有限公司 Service processing method and device based on human face
CN114283471B (en) * 2021-12-16 2024-04-02 武汉大学 Multi-mode ordering optimization method for heterogeneous face image re-recognition
CN114283471A (en) * 2021-12-16 2022-04-05 武汉大学 Multi-modal sequencing optimization method for heterogeneous face image re-recognition
CN115147578B (en) * 2022-06-30 2023-10-27 北京百度网讯科技有限公司 Stylized three-dimensional face generation method and device, electronic equipment and storage medium
CN115147578A (en) * 2022-06-30 2022-10-04 北京百度网讯科技有限公司 Stylized three-dimensional face generation method and device, electronic equipment and storage medium
CN115496975B (en) * 2022-08-29 2023-08-18 锋睿领创(珠海)科技有限公司 Auxiliary weighted data fusion method, device, equipment and storage medium
CN115496975A (en) * 2022-08-29 2022-12-20 锋睿领创(珠海)科技有限公司 Auxiliary weighted data fusion method, device, equipment and storage medium
CN116362933A (en) * 2023-05-30 2023-06-30 南京农业大学 Intelligent campus management method and system based on big data
CN116362933B (en) * 2023-05-30 2023-09-26 南京农业大学 Intelligent campus management method and system based on big data
CN116403270A (en) * 2023-06-07 2023-07-07 南昌航空大学 Facial expression recognition method and system based on multi-feature fusion
CN116403270B (en) * 2023-06-07 2023-09-05 南昌航空大学 Facial expression recognition method and system based on multi-feature fusion

Also Published As

Publication number Publication date
CN104598878A (en) 2015-05-06

Similar Documents

Publication Publication Date Title
WO2016110005A1 (en) Gray level and depth information based multi-layer fusion multi-modal face recognition device and method
Zhan et al. Face detection using representation learning
Keskin et al. Hand pose estimation and hand shape classification using multi-layered randomized decision forests
WO2017219391A1 (en) Face recognition system based on three-dimensional data
KR102174595B1 (en) System and method for identifying faces in unconstrained media
JP5845365B2 (en) Improvements in or related to 3D proximity interaction
Zhang et al. Adaptive facial point detection and emotion recognition for a humanoid robot
CN108256421A (en) A kind of dynamic gesture sequence real-time identification method, system and device
US8837773B2 (en) Apparatus which detects moving object from image and method thereof
CN106407958B (en) Face feature detection method based on double-layer cascade
CN111126240B (en) Three-channel feature fusion face recognition method
CN111460976B (en) Data-driven real-time hand motion assessment method based on RGB video
CN109086659B (en) Human behavior recognition method and device based on multi-channel feature fusion
CN104778441A (en) Multi-mode face identification device and method fusing grey information and depth information
Pandey et al. Hand gesture recognition for sign language recognition: A review
Yang et al. Facial expression recognition based on dual-feature fusion and improved random forest classifier
CN106599785A (en) Method and device for building human body 3D feature identity information database
Hsu et al. Fast landmark localization with 3D component reconstruction and CNN for cross-pose recognition
Galdámez et al. Ear recognition using a hybrid approach based on neural networks
Xia et al. Face occlusion detection using deep convolutional neural networks
Pathak et al. Multimodal eye biometric system based on contour based E-CNN and multi algorithmic feature extraction using SVBF matching
Vezzetti et al. Application of geometry to rgb images for facial landmark localisation-a preliminary approach
Galdámez et al. A small look at the ear recognition process using a hybrid approach
Xu et al. A novel mid-level distinctive feature learning for action recognition via diffusion map
Ayodele et al. Development of a modified local Binary Pattern-Gabor Wavelet transform aging invariant face recognition system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15876513

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15876513

Country of ref document: EP

Kind code of ref document: A1