CN105550657A - Key point based improved SIFT human face feature extraction method - Google Patents

Key point based improved SIFT human face feature extraction method Download PDF

Info

Publication number
CN105550657A
CN105550657A CN201510977092.3A CN201510977092A CN105550657A CN 105550657 A CN105550657 A CN 105550657A CN 201510977092 A CN201510977092 A CN 201510977092A CN 105550657 A CN105550657 A CN 105550657A
Authority
CN
China
Prior art keywords
face
pixel
crucial
face picture
convolutional network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510977092.3A
Other languages
Chinese (zh)
Other versions
CN105550657B (en
Inventor
李伟
王璐
冯复标
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Chemical Technology
Original Assignee
Beijing University of Chemical Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Chemical Technology filed Critical Beijing University of Chemical Technology
Priority to CN201510977092.3A priority Critical patent/CN105550657B/en
Publication of CN105550657A publication Critical patent/CN105550657A/en
Application granted granted Critical
Publication of CN105550657B publication Critical patent/CN105550657B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a key point based improved SIFT human face feature extraction method. Five key points are described by utilizing a directional histogram through positioning five key pixel points in a human face so as to form robust human face image feature vectors. A score of the similarity between two human face feature vectors is calculated in combination with a bilinear similarity function and a Mahalanobis distance. The score of the similarity is subjected to binary classification by adopting a KELM classifier; for human face pictures with relatively high scores, two human face pictures are both judged to be from a same person; and for human face pictures with relatively low scores, two human face pictures are both judged to be from different persons. In a process of human face identification based on the human face feature vectors, the score of the similarity between the two feature vectors is calculated in combination with the bilinear similarity function and the Mahalanobis distance, so that the distinguishability of classes is enhanced.

Description

Based on the improvement SIFT face feature extraction method of key point
Technical field
The present invention relates to a kind of improvement SIFT (Scale invariant features transform) face feature extraction method based on key point, belong to field of face identification.
Background technology
Recognition of face is the biological identification technology that a kind of face feature information based on people carries out identification.Compared with other biological feature, face characteristic has natural sex, convenience and the advantage such as untouchable, makes it in security monitoring, authentication, man-machine interaction etc., have huge application prospect.Therefore, face recognition technology very has researching value.Generally speaking, face recognition process is divided into two processes: face characteristic extracts and human face similarity degree score value calculates.Face characteristic leaching process is that some key features extracting face picture form face feature vector, human face similarity degree score value computation process is the similarity between calculating two face feature vector, similarity is higher, shows that two face picture more likely come from same person, otherwise, then more show that two face picture come from different people.In some cases, more it is of concern that face characteristic Extraction parts.
Existing face feature extraction method comprises LBP (local binary patterns) method and its mutation method etc., these Local textural feature extracting method form histogram vectors by carrying out block statistics to whole face picture, and the histogram vectors cascade of each block is finally formed face feature vector.Because this method carries out Local textural feature extraction to whole face, therefore, its feature vector dimension formed is larger, and wherein contains some redundant informations.In addition, this mode is for the change not robust of expression or attitude under complex environment.
SIFT feature extracting method has been widely used in the identification of general object, and its main thought finds the key point of image under different scale, and describe the proper vector of key point as image with direction histogram.But, when SIFT method is used for facial image, accurately can not navigates to the key point in face, because it is mainly applicable to the identification of the general object with higher contrast, and there is between facial image higher similarity.
Summary of the invention
Fundamental purpose of the present invention is to provide a kind of improvement SIFT face feature extraction method based on key point.
Be different from traditional feature extracting method based on whole face, concrete innovative point of the present invention is to have employed the method that the improvement SIFT face characteristic based on key point extracts.By the crucial pixel of five in locating human face, and utilize the direction histogram in SIFT method to describe this five key points, thus form the facial image proper vector of robust.Described five crucial pixels are respectively the pixel in left eye centre position, the pixel in right eye centre position, supratip pixel, the pixel of the left corners of the mouth and the pixel of the right corners of the mouth.
Technical scheme of the present invention specifically, mainly comprises following technology contents:
1, five the crucial pixels (pixel in left eye centre position, the pixel in right eye centre position, supratip pixel, the pixel of the left corners of the mouth and the pixel of the right corners of the mouth) in three layer depth convolutional network cascade locating human face are adopted.
2, SIFT feature extracting method is improved: replace with the crucial pixel of five in face the key point that SIFT feature extracting method self detects, reduce characteristic dimension, the redundant information in rejecting characteristic.
3, by the maps feature vectors of face in intra-personal subspace, ensure same person different face picture between there is unchangeability in class.
4, the similarity score value between two face feature vector is calculated in conjunction with bilinearity similarity function and mahalanobis distance.This score value is higher, show that two face picture more may come from same person, otherwise this score value is lower, shows that two face picture more may come from different people.
5, KELM sorter (extreme learning machine based on core) is adopted to carry out two-value classification to similarity score value, the class face picture that score value is higher, two face picture are all judged to and come from same person, and the class face picture that score value is lower, two face picture are all judged to and come from different people.
As shown in Figure 1, implementing procedure is as follows for process flow diagram of the present invention:
Step 1, reading face picture, and utilize five crucial pixel (pixels in left eye centre position on three layer depth convolutional network cascade locating human face pictures, the pixel in right eye centre position, supratip pixel, the pixel of the left corners of the mouth and the pixel of the right corners of the mouth).
Degree of depth convolutional network cascade used in this step comprises three layers.Ground floor utilizes degree of depth convolutional network accurately to locate five crucial pixels, and other two-layer convolutional network positioning results to ground floor that utilize are reaffirmed.In order to ensure the accuracy of locating, in every layer, the positioning result of each degree of depth convolutional network is merged the positioning result be averaged as final.Deep layer convolutional network comprises four convolutional layers, pond layer and two full articulamentums, initial layers obtains the global context information of face picture, because convolutional network predicts five crucial pixels simultaneously, so relative position between each crucial pixel is also encoded while convolutional network training, and then weaken the impact that expression shape change, illumination variation and other environmental factors cause.
Step 2, for five the crucial pixels extracted in step 1, utilize SIFT method to carry out feature interpretation, extract the feature of face picture.
SIFT feature extracting method is a kind of method detecting local feature, not only has scale invariability, has rotational invariance simultaneously.The method generally comprises Four processes: (1) builds metric space, detects key point; (2) unstable key point is rejected; (3) be key point assignment direction parameter; (4) descriptor of key point is generated.Generally speaking, SIFT feature extracting method is applicable to the identification of the general object with higher contrast, and face picture has lower contrast and skirt response, because SIFT feature extracting method can not key point accurately in locating human face's picture, therefore replace step (1) in SIFT feature extracting method and step (2) by the method for locator key pixel in step 1.
Then, for five the crucial pixels obtained in step 1, get some pixels in each crucial pixel field, calculate gradient modulus value and the direction of each pixel.The coordinate defining certain pixel is P (x, y):
m ( x , y ) = ( P ( x + 1 , y ) - P ( x - 1 , y ) ) 2 + ( P ( x , y + 1 ) - P ( x , y - 1 ) ) 2
θ(x,y)=tan -1((P(x,y+1)-P(x,y-1))/(P(x+1,y)-P(x-1,y)))
Wherein, the gradient modulus value that m (x, y) is this pixel, the gradient direction that θ (x, y) is this pixel.
According to the result of calculation of above formula, utilize the gradient direction of pixel in statistics with histogram field.For reducing sudden change impact, need with Gaussian function smoothing to histogram.So, histogrammic peak value represents the gradient principal direction of crucial pixel field pixel, is also the direction of crucial pixel.
In order to keep rotational invariance, by the direction that X-axis rotate is crucial pixel, then centered by key point, get the field window of 16 × 16 sizes, in the grid of every 4 × 4 sizes, calculate the histogram of 8 gradient directions, finally form the SIFT face feature vector of 4 × 4 × 8=128 dimension.
Step 3, by the maps feature vectors that obtains in step 2 in intra-personal subspace.
In this step, in order to the impact of attenuating noise, first need the proper vector obtained in step 2 to be utilized PCA method (principal component analysis (PCA)) to carry out dimensionality reduction, morphogenesis characters face.Its covariance matrix expression formula is as follows:
C = Σ i = 1 n ( x i - m ) ( x i - m ) T
Wherein, n is face sample size, x irepresent face vector, m is the average of n face vector.Because covariance matrix describes the correlativity between vector, therefore the proper vector of above-mentioned covariance matrix forms mapping matrix, maps face image data according to mapping matrix, can morphogenesis characters face.Then, in order to ensure unchangeability in the class between the face picture that same person is different, eigenface be mapped in intra-personal subspace, its covariance matrix expression formula is as follows:
C S = Σ ( i , j ) ∈ S ( x i - x j ) ( x i - x j ) T
Wherein, S represents the face picture set of same person, x iand x jrepresent the face vector of two face picture different in the face picture set of same person.∧={ λ 1..., λ kand V={v 1..., v krepresent front k eigenwert and the proper vector of above formula covariance matrix respectively.Similarly, front k proper vector forms mapping matrix, above-mentioned eigenface data is mapped by this mapping matrix, thus unchangeability in class between the face picture ensureing same person.If C sbe reversible, so, eigenface is mapped to intra-personal subspace following formula and expresses:
L S = V d i a g ( λ 1 1 / 2 , ... , λ k 1 / 2 )
x ~ = L S - 1 x
Wherein, V is the mapping matrix formed by an above-mentioned k proper vector, the diagonal matrix be made up of an above-mentioned k eigenwert, the i.e. final eigenmatrix formed.
Step 4, bilinearity similarity function and mahalanobis distance is utilized to calculate similarity score value between two face feature vector.
Mahalanobis distance is widely used in field of face identification, but its recognition effect is not good especially, and research in recent years shows, bilinearity similarity function has obtained good effect at picture analogies degree search field.Therefore, in this method, calculate the similarity score value between two face feature vector in conjunction with bilinearity similarity function and mahalanobis distance, its expression formula is as follows:
f ( M , G ) ( x ~ i , x ~ j ) = S G ( x ~ i , x ~ j ) - d M ( x ~ i , x ~ j )
S G ( x ~ i , x ~ j ) = x i T Gx j
d M ( x ~ i , x ~ j ) = ( x i - x j ) T M ( x i - x j )
Wherein, represent with bilinearity similarity function between eigenmatrix, represent with mahalanobis distance between eigenmatrix.G and M is the matrix of k × k size, needs to train suitable M and G to ensure the maximum distinguishability ensured while unchangeability in class between class as far as possible.Therefore, the expression formula that intra-personal subspace similarity measure learns is defined as following form:
m i n M , G ∈ S d Σ t ∈ P ξ t + γ 2 ( | | M - I | | F 2 + | | G - I | | F 2 )
s . t . y i j · [ f ( M , G ) ( x ~ i , x ~ j ) ] ≥ 1 - ξ i j
ξ t ≥ 0 , ∀ t = ( i , j ) ∈ P = S ∪ D
Wherein, S with D represents similar face to (i.e. two face picture of same person) and dissimilar face respectively to the label of (i.e. two face picture of different people).|| .|| fthe F norm of matrix, i.e. the quadratic sum of matrix element absolute value evolution again.Its effect is similar to 2 norms of vector, therefore, and expression formula effectively Expired Drugs is prevented while unchangeability in guarantee class.ξ tbe the loss function that experience differentiates, minimize this parameter and can strengthen distinguishability between class.As can be seen here, ξ tensure that the maximum distinguishability between class, and ensure that unchangeability in class, and positive number γ is used for coordinating the impact that this two expression formulas are brought.For the inequality in above-mentioned expression formula, when a pair face picture comes from same person, y ij=1, and ξ ijvalue less, so value can be large as far as possible.And when a pair face picture comes from different people, y ij=-1, and ξ ijvalue less, so value can be little as far as possible.Therefore, when when being worth larger, then show that a pair face picture comes from same person, otherwise, then show that a pair face comes from different people.
Step 5, whether two pictures come from same person to utilize KELM sorter to judge.
ELM is the neural network only comprising a hidden layer and an output layer.Its most outstanding feature is that its hidden layer parameter does not need to debug, but random setting, there is stronger generalization ability.Suppose that the hidden layer of ELM comprises L node, then its output function is as follows:
f L ( x ) = Σ i = 1 L β i h ( ω i , b i , x ) = h ( x ) β = y
Wherein, x ∈ R d, y ∈ R c, β represents L of the hidden layer weight between node and output layer, and h (x) represents the relation between L node and input x, it is a nonlinear excitation function (as sigmoid function), in fact, its effect d dimension data is mapped in L dimension data space, w irepresent the connection weight of hidden layer i-th node and input layer, b irepresent the deviation of hidden layer i-th node.
On the basis of ELM, also been proposed a kind of ELM method based on kernel function, i.e. KELM method.The method carrys out the original excitation function H of hiding ELM by kernel function, thus improve the generalization ability of algorithm better.For a certain sample x i, it is as follows that its output function expresses formula:
f L ( x i ) = K ( x i , x 1 ) . . . K ( x i , x n ) T ( I C + K ) - 1 Y
Y=[y 1;...;y n]∈R n×c
Wherein, C is a regression coefficient.
Using the similarity score value that obtains in step 4 as the input of KELM sorter, if the output obtained 1, then show that two face picture come from same person, if 0, then show that two face picture come from different people.
Compared with prior art, tool of the present invention has the following advantages:
Be different from traditional feature extracting method based on whole face, concrete innovative point of the present invention is to have employed the method that the improvement SIFT face characteristic based on key point extracts.By the pixel in face left eye centre position, the pixel in right eye centre position, supratip pixel, the pixel of the left corners of the mouth and the pixel of the right corners of the mouth are as five crucial pixels of whole face, and extract the proper vector of SIFT feature as face of these five key points, not only on the basis reducing intrinsic dimensionality, improve operation efficiency, and weaken due to the impact on recognition effect such as illumination, expression shape change and other environmental factors.In addition, the basis of face feature vector is carried out in the process of recognition of face, combine the similarity score value that bilinearity similarity function and mahalanobis distance calculate two proper vectors, enhance the distinguishability between class, simultaneously, in order to the similarity unchangeability between considering in class, face feature vector is mapped in intra-personal subspace.By cross-validation experiments, the retrievable face recognition accuracy rate of the present invention can reach 80.56%.
Accompanying drawing explanation
Fig. 1 is particular flow sheet of the present invention.
Embodiment
The basic procedure of the improvement SIFT face feature extraction method based on key point of the present invention as shown in Figure 1, specifically comprises the following steps:
1) data in face database are divided into 10 groups and carry out cross-validation experiments, wherein, 9 groups of data, as training data, remain 1 group of data as test data, often comprise 300 in group to the face picture and 300 from same person to the face picture from different people.For often opening face picture, three layer depth convolutional network cascades are utilized to locate the coordinate position (pixel in left eye centre position of five crucial pixels, the pixel in right eye centre position, supratip pixel, the pixel of the left corners of the mouth and the pixel of the right corners of the mouth).
In the ground floor of three layer depth convolutional network cascade structures, comprise three degree of depth convolutional network.The input of first degree of depth convolutional network is whole face, exports the position of five crucial pixels; The input of second degree of depth convolutional network is the first half of face, exports the position of the crucial pixel of eyes and nose position; The input of the 3rd degree of depth convolutional network is the latter half of face, exports the position of the crucial pixel of nose and face position.Finally, the Output rusults of these three degree of depth convolutional network is averaged obtains the final output of ground floor.The second layer of deep layer convolutional network cascade and third layer all get field in front one deck Output rusults around crucial pixel as input, reaffirm the coordinate position of crucial pixel, are supplementing ground floor Output rusults.
2) for step 1) in five crucial pixels, centered by each crucial pixel, get the neighborhood window size of 14 × 14, calculate gradient modulus value and the direction of each pixel in this window.Then the histogram in statistical gradient direction.In order to reduce the impact of sudden change, be that the Gaussian function of 1.5 × 14 is to the smoothing process of histogram with parameter σ.The last direction determining each crucial pixel with the gradient direction distribution characteristic of each crucial neighborhood of pixel points pixel.
After determining the direction of crucial pixel, be the direction of crucial pixel by X-axis rotate, to guarantee rotational invariance.Then centered by crucial pixel, get the neighborhood window size of 16 × 16, on the fritter of every 4 × 4 sizes, then calculate the histogram of gradients in 8 directions, finally, each proper vector forms the descriptor of 4 × 4 × 8=128 dimension.Because the crucial pixel number often opening face picture in the present invention is 5, therefore, the proper vector of often opening face picture is 5 × 128=640 dimension, this not only greatly reduces intrinsic dimensionality, and weaken the impact that expression shape change, illumination variation and other environmental factors etc. cause, enhance the robustness of face characteristic.
3) by step 2) in the proper vector that obtains utilize PCA (principal component analysis (PCA)) method to carry out dimensionality reduction, obtain front 400 main variables, form the eigenface of 400 dimensions.Then, in order to ensure unchangeability in the class between the face picture that same person is different, eigenface be mapped in intra-personal subspace, wherein, the intrinsic dimensionality in intra-personal space gets 300.
4) the M parameter and G parameter that obtain in bilinearity similarity function and mahalanobis distance is trained by 9 groups of training datas.Similarity score value in the M parameter utilizing training process to obtain and G parameter calculation training data and test data between each pair of face picture.
5) whether two pictures come from same person to utilize KELM sorter to judge.In KELM sorter, the present invention selects Radial basis kernel function (RBF) as kernel function, regression coefficient C=1024.Using step 4) in similarity score value between two face picture obtaining as the input of KELM sorter, if the sorter obtained exports 1, then show that two face picture come from same person, if 0, then show that two face picture come from different people.

Claims (2)

1., based on the improvement SIFT face feature extraction method of key point, it is characterized in that: the implementing procedure of this method is as follows,
Step 1, reading face picture, and utilize five crucial pixels on three layer depth convolutional network cascade locating human face pictures; These five pixels are respectively the pixel in left eye centre position, the pixel in right eye centre position, supratip pixel, the pixel of the left corners of the mouth and the pixel of the right corners of the mouth;
Degree of depth convolutional network cascade used in this step comprises three layers; Ground floor utilizes degree of depth convolutional network accurately to locate five crucial pixels, and other two-layer convolutional network positioning results to ground floor that utilize are reaffirmed; In order to ensure the accuracy of locating, in every layer, the positioning result of each degree of depth convolutional network is merged the positioning result be averaged as final; Deep layer convolutional network comprises four convolutional layers, pond layer and two full articulamentums, initial layers obtains the global context information of face picture, because convolutional network predicts five crucial pixels simultaneously, so relative position between each crucial pixel is also encoded while convolutional network training, and then weaken the impact that expression shape change, illumination variation and other environmental factors cause;
Step 2, for five the crucial pixels extracted in step 1, utilize SIFT method to carry out feature interpretation, extract the feature of face picture;
SIFT feature extracting method is a kind of method detecting local feature, not only has scale invariability, has rotational invariance simultaneously; The method generally comprises Four processes: (1) builds metric space, detects key point; (2) unstable key point is rejected; (3) be key point assignment direction parameter; (4) descriptor of key point is generated; Generally speaking, SIFT feature extracting method is applicable to the identification of the general object with higher contrast, and face picture has lower contrast and skirt response, because SIFT feature extracting method can not key point accurately in locating human face's picture, therefore replace step (1) in SIFT feature extracting method and step (2) by the method for locator key pixel in step 1;
Then, for five the crucial pixels obtained in step 1, get some pixels in each crucial pixel field, calculate gradient modulus value and the direction of each pixel; The coordinate defining certain pixel is P (x, y):
m ( x , y ) = ( P ( x + 1 , y ) - P ( x - 1 , y ) ) 2 + ( P ( x , y + 1 ) - P ( x , y - 1 ) ) 2
θ(x,y)=tan -1((P(x,y+1)-P(x,y-1))/(P(x+1,y)-P(x-1,y)))
Wherein, the gradient modulus value that m (x, y) is this pixel, the gradient direction that θ (x, y) is this pixel;
According to the result of calculation of above formula, utilize the gradient direction of pixel in statistics with histogram field; For reducing sudden change impact, need with Gaussian function smoothing to histogram; So, histogrammic peak value represents the gradient principal direction of crucial pixel field pixel, is also the direction of crucial pixel;
In order to keep rotational invariance, by the direction that X-axis rotate is crucial pixel, then centered by key point, get the field window of 16 × 16 sizes, in the grid of every 4 × 4 sizes, calculate the histogram of 8 gradient directions, finally form the SIFT face feature vector of 4 × 4 × 8=128 dimension;
Step 3, by the maps feature vectors that obtains in step 2 in intra-personal subspace;
In this step, in order to the impact of attenuating noise, first need the proper vector obtained in step 2 to be utilized PCA method (principal component analysis (PCA)) to carry out dimensionality reduction, morphogenesis characters face; Its covariance matrix expression formula is as follows:
C = Σ i = 1 n ( x i - m ) ( x i - m ) T
Wherein, n is face sample size, x irepresent face vector, m is the average of n face vector; Because covariance matrix describes the correlativity between vector, therefore the proper vector of above-mentioned covariance matrix forms mapping matrix, maps face image data according to mapping matrix, gets final product morphogenesis characters face; Then, in order to ensure unchangeability in the class between the face picture that same person is different, eigenface be mapped in intra-personal subspace, its covariance matrix expression formula is as follows:
C S = Σ ( i , j ) ∈ S ( x i - x j ) ( x i - x j ) T
Wherein, S represents the face picture set of same person, x iand x jrepresent the face vector of two face picture different in the face picture set of same person; ∧={ λ 1..., λ kand V=(v 1..., v k) represent front k eigenwert and the proper vector of above formula covariance matrix respectively; Similarly, front k proper vector forms mapping matrix, above-mentioned eigenface data is mapped by this mapping matrix, thus unchangeability in class between the face picture ensureing same person; If C sbe reversible, so, eigenface is mapped to intra-personal subspace following formula and expresses:
L S = V d i a g ( λ 1 1 / 2 , ... , λ k 1 / 2 )
x ~ = L S - 1 x
Wherein, V is the mapping matrix formed by an above-mentioned k proper vector, the diagonal matrix be made up of an above-mentioned k eigenwert, the i.e. final eigenmatrix formed;
Step 4, bilinearity similarity function and mahalanobis distance is utilized to calculate similarity score value between two face feature vector;
Mahalanobis distance is widely used in field of face identification, but its recognition effect is not good especially, and research in recent years shows, bilinearity similarity function has obtained good effect at picture analogies degree search field; Therefore, in this method, calculate the similarity score value between two face feature vector in conjunction with bilinearity similarity function and mahalanobis distance, its expression formula is as follows:
f ( M , G ) ( x ~ i , x ~ j ) = S G ( x ~ i , x ~ j ) - d M ( x ~ i , x ~ j )
S G ( x ~ i , x ~ j ) = x i T Gx j
d M ( x ~ i , x ~ j ) = ( x i - x j ) T M ( x i - x j )
Wherein, represent with bilinearity similarity function between eigenmatrix, represent with mahalanobis distance between eigenmatrix; G and M is the matrix of k × k size, needs to train suitable M and G to ensure the maximum distinguishability ensured while unchangeability in class between class as far as possible; Therefore, the expression formula that intra-personal subspace similarity measure learns is defined as following form:
m i n M , G ∈ S d Σ t ∈ P ξ t + γ 2 ( | | M - I | | F 2 + | | G - I | | F 2 )
s . t . y i j [ f ( M , G ) ( x ~ i , x ~ j ) ] ≥ 1 - ξ i j
ξ t ≥ 0 , ∀ t = ( i , j ) ∈ P = S ∪ D
Wherein, S with D represents similar face to (i.e. two face picture of same person) and dissimilar face respectively to the label of (i.e. two face picture of different people); || || fthe F norm of matrix, i.e. the quadratic sum of matrix element absolute value evolution again; Its effect is similar to 2 norms of vector, therefore, and expression formula effectively Expired Drugs is prevented while unchangeability in guarantee class; ξ tbe the loss function that experience differentiates, minimize this parameter and can strengthen distinguishability between class; As can be seen here, ξ tensure that the maximum distinguishability between class, and ensure that unchangeability in class, and positive number γ is used for coordinating the impact that this two expression formulas are brought; For the inequality in above-mentioned expression formula, when a pair face picture comes from same person, y ij=1, and ξ ijvalue less, so value can be large as far as possible; And when a pair face picture comes from different people, y ij=-1, and ξ ijvalue less, so value can be little as far as possible; Therefore, when when being worth larger, then show that a pair face picture comes from same person, otherwise, then show that a pair face comes from different people;
Step 5, whether two pictures come from same person to utilize KELM sorter to judge;
ELM is the neural network only comprising a hidden layer and an output layer; Its most outstanding feature is that its hidden layer parameter does not need to debug, but random setting, there is stronger generalization ability; Suppose that the hidden layer of ELM comprises L node, then its output function is as follows:
f L ( x ) = Σ i = 1 L β i h ( ω i , b i , x ) = h ( x ) β = y
Wherein, x ∈ R d, y ∈ R c, β represents L of the hidden layer weight between node and output layer, and h (x) represents the relation between L node and input x, it is a nonlinear excitation function (as sigmoid function), in fact, its effect d dimension data is mapped in L dimension data space, w irepresent the connection weight of hidden layer i-th node and input layer, b irepresent the deviation of hidden layer i-th node;
On the basis of ELM, also been proposed a kind of ELM method based on kernel function, i.e. KELM method; The method carrys out the original excitation function H of hiding ELM by kernel function, thus improve the generalization ability of algorithm better; For a certain sample x i, it is as follows that its output function expresses formula:
f L ( x i ) = K ( x i , x 1 ) . . . K ( x i , x n ) T ( I C + K ) - 1 Y
Y=[y 1;...;y n]∈R n×c
Wherein, C is a regression coefficient;
Using the similarity score value that obtains in step 4 as the input of KELM sorter, if the output obtained 1, then show that two face picture come from same person, if 0, then show that two face picture come from different people.
2. the improvement SIFT face feature extraction method based on key point according to claim 1, is characterized in that: this method specifically comprises the following steps,
1) data in face database are divided into 10 groups and carry out cross-validation experiments, wherein, 9 groups of data, as training data, remain 1 group of data as test data, often comprise 300 in group to the face picture and 300 from same person to the face picture from different people; For often opening face picture, three layer depth convolutional network cascades are utilized to locate the coordinate position of five crucial pixels, the pixel in left eye centre position, the pixel in right eye centre position, supratip pixel, the pixel of the left corners of the mouth and the pixel of the right corners of the mouth;
In the ground floor of three layer depth convolutional network cascade structures, comprise three degree of depth convolutional network; The input of first degree of depth convolutional network is whole face, exports the position of five crucial pixels; The input of second degree of depth convolutional network is the first half of face, exports the position of the crucial pixel of eyes and nose position; The input of the 3rd degree of depth convolutional network is the latter half of face, exports the position of the crucial pixel of nose and face position; Finally, the Output rusults of these three degree of depth convolutional network is averaged obtains the final output of ground floor; The second layer of deep layer convolutional network cascade and third layer all get field in front one deck Output rusults around crucial pixel as input, reaffirm the coordinate position of crucial pixel, are supplementing ground floor Output rusults;
2) for step 1) in five crucial pixels, centered by each crucial pixel, get the neighborhood window size of 14 × 14, calculate gradient modulus value and the direction of each pixel in this window; Then the histogram in statistical gradient direction; In order to reduce the impact of sudden change, be that the Gaussian function of 1.5 × 14 is to the smoothing process of histogram with parameter σ; The last direction determining each crucial pixel with the gradient direction distribution characteristic of each crucial neighborhood of pixel points pixel;
After determining the direction of crucial pixel, be the direction of crucial pixel by X-axis rotate, to guarantee rotational invariance; Then centered by crucial pixel, get the neighborhood window size of 16 × 16, on the fritter of every 4 × 4 sizes, then calculate the histogram of gradients in 8 directions, finally, each proper vector forms the descriptor of 4 × 4 × 8=128 dimension; Because the crucial pixel number often opening face picture in the present invention is 5, therefore, the proper vector of often opening face picture is 5 × 128=640 dimension, this not only greatly reduces intrinsic dimensionality, and weaken the impact that expression shape change, illumination variation and other environmental factors etc. cause, enhance the robustness of face characteristic;
3) by step 2) in the proper vector that obtains utilize PCA (principal component analysis (PCA)) method to carry out dimensionality reduction, obtain front 400 main variables, form the eigenface of 400 dimensions; Then, in order to ensure unchangeability in the class between the face picture that same person is different, eigenface be mapped in intra-personal subspace, wherein, the intrinsic dimensionality in intra-personal space gets 300;
4) the M parameter and G parameter that obtain in bilinearity similarity function and mahalanobis distance is trained by 9 groups of training datas; Similarity score value in the M parameter utilizing training process to obtain and G parameter calculation training data and test data between each pair of face picture;
5) whether two pictures come from same person to utilize KELM sorter to judge; In KELM sorter, the present invention selects Radial basis kernel function (RBF) as kernel function, regression coefficient C=1024; Using step 4) in similarity score value between two face picture obtaining as the input of KELM sorter, if the sorter obtained exports 1, then show that two face picture come from same person, if 0, then show that two face picture come from different people.
CN201510977092.3A 2015-12-23 2015-12-23 Improvement SIFT face feature extraction method based on key point Active CN105550657B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510977092.3A CN105550657B (en) 2015-12-23 2015-12-23 Improvement SIFT face feature extraction method based on key point

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510977092.3A CN105550657B (en) 2015-12-23 2015-12-23 Improvement SIFT face feature extraction method based on key point

Publications (2)

Publication Number Publication Date
CN105550657A true CN105550657A (en) 2016-05-04
CN105550657B CN105550657B (en) 2019-01-29

Family

ID=55829840

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510977092.3A Active CN105550657B (en) 2015-12-23 2015-12-23 Improvement SIFT face feature extraction method based on key point

Country Status (1)

Country Link
CN (1) CN105550657B (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106534576A (en) * 2016-12-07 2017-03-22 深圳市传奇数码有限公司 Side face unlocking method applied to mobile phone
CN106528662A (en) * 2016-10-20 2017-03-22 中山大学 Quick retrieval method and system of vehicle image on the basis of feature geometric constraint
CN106778531A (en) * 2016-11-25 2017-05-31 北京小米移动软件有限公司 Face detection method and device
CN106778775A (en) * 2016-11-29 2017-05-31 南京邮电大学 A kind of image classification method based on the soft matching of SIFT feature
CN106845331A (en) * 2016-11-18 2017-06-13 深圳云天励飞技术有限公司 A kind of image processing method and terminal
CN107169420A (en) * 2017-04-19 2017-09-15 南京邮电大学 A kind of eyeball center positioning method based on image gradient
WO2017215540A1 (en) * 2016-06-12 2017-12-21 广州广电运通金融电子股份有限公司 Offline identity authentication method and apparatus
CN107945219A (en) * 2017-11-23 2018-04-20 翔创科技(北京)有限公司 Face image alignment schemes, computer program, storage medium and electronic equipment
CN108921140A (en) * 2018-08-07 2018-11-30 安徽云森物联网科技有限公司 Pedestrian's recognition methods again
CN109344760A (en) * 2018-09-26 2019-02-15 江西师范大学 A kind of construction method of natural scene human face expression data collection
WO2019134327A1 (en) * 2018-01-03 2019-07-11 东北大学 Facial expression recognition feature extraction method employing edge detection and sift
CN110751005A (en) * 2018-07-23 2020-02-04 合肥工业大学 Pedestrian detection method integrating depth perception features and kernel extreme learning machine
CN111008589A (en) * 2019-12-02 2020-04-14 杭州网易云音乐科技有限公司 Face key point detection method, medium, device and computing equipment
CN114936597A (en) * 2022-05-20 2022-08-23 电子科技大学 Method for extracting space true and false target characteristics of local information enhancer

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20100073749A (en) * 2008-12-23 2010-07-01 고려대학교 산학협력단 Apparatus and method for extracting feature point based on sift, and face recognition system using thereof
CN103729625A (en) * 2013-12-31 2014-04-16 青岛高校信息产业有限公司 Face identification method
CN104036524A (en) * 2014-06-18 2014-09-10 哈尔滨工程大学 Fast target tracking method with improved SIFT algorithm
CN104794441A (en) * 2015-04-15 2015-07-22 重庆邮电大学 Human face feature extracting method based on active shape model and POEM (patterns of oriented edge magnituedes) texture model in complicated background
CN105095857A (en) * 2015-06-26 2015-11-25 上海交通大学 Face data enhancement method based on key point disturbance technology

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20100073749A (en) * 2008-12-23 2010-07-01 고려대학교 산학협력단 Apparatus and method for extracting feature point based on sift, and face recognition system using thereof
CN103729625A (en) * 2013-12-31 2014-04-16 青岛高校信息产业有限公司 Face identification method
CN104036524A (en) * 2014-06-18 2014-09-10 哈尔滨工程大学 Fast target tracking method with improved SIFT algorithm
CN104794441A (en) * 2015-04-15 2015-07-22 重庆邮电大学 Human face feature extracting method based on active shape model and POEM (patterns of oriented edge magnituedes) texture model in complicated background
CN105095857A (en) * 2015-06-26 2015-11-25 上海交通大学 Face data enhancement method based on key point disturbance technology

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10417532B2 (en) 2016-06-12 2019-09-17 Grg Banking Equipment Co., Ltd. Offline identity authentication method and apparatus
WO2017215540A1 (en) * 2016-06-12 2017-12-21 广州广电运通金融电子股份有限公司 Offline identity authentication method and apparatus
CN106528662A (en) * 2016-10-20 2017-03-22 中山大学 Quick retrieval method and system of vehicle image on the basis of feature geometric constraint
CN106845331A (en) * 2016-11-18 2017-06-13 深圳云天励飞技术有限公司 A kind of image processing method and terminal
CN106845331B (en) * 2016-11-18 2018-03-13 深圳云天励飞技术有限公司 A kind of image processing method and terminal
WO2018090937A1 (en) * 2016-11-18 2018-05-24 深圳云天励飞技术有限公司 Image processing method, terminal and storage medium
US10467743B1 (en) 2016-11-18 2019-11-05 Shenzhen Intellifusion Technologies Co., Ltd. Image processing method, terminal and storage medium
CN106778531A (en) * 2016-11-25 2017-05-31 北京小米移动软件有限公司 Face detection method and device
CN106778775A (en) * 2016-11-29 2017-05-31 南京邮电大学 A kind of image classification method based on the soft matching of SIFT feature
CN106778775B (en) * 2016-11-29 2020-08-18 南京邮电大学 Image classification method based on SIFT feature soft matching
CN106534576A (en) * 2016-12-07 2017-03-22 深圳市传奇数码有限公司 Side face unlocking method applied to mobile phone
CN107169420B (en) * 2017-04-19 2020-04-21 南京邮电大学 Eyeball center positioning method based on image gradient
CN107169420A (en) * 2017-04-19 2017-09-15 南京邮电大学 A kind of eyeball center positioning method based on image gradient
CN107945219B (en) * 2017-11-23 2019-12-03 翔创科技(北京)有限公司 Face image alignment schemes, computer program, storage medium and electronic equipment
CN107945219A (en) * 2017-11-23 2018-04-20 翔创科技(北京)有限公司 Face image alignment schemes, computer program, storage medium and electronic equipment
WO2019134327A1 (en) * 2018-01-03 2019-07-11 东北大学 Facial expression recognition feature extraction method employing edge detection and sift
CN110751005A (en) * 2018-07-23 2020-02-04 合肥工业大学 Pedestrian detection method integrating depth perception features and kernel extreme learning machine
CN110751005B (en) * 2018-07-23 2022-09-30 合肥工业大学 Pedestrian detection method integrating depth perception features and kernel extreme learning machine
CN108921140A (en) * 2018-08-07 2018-11-30 安徽云森物联网科技有限公司 Pedestrian's recognition methods again
CN109344760A (en) * 2018-09-26 2019-02-15 江西师范大学 A kind of construction method of natural scene human face expression data collection
CN111008589A (en) * 2019-12-02 2020-04-14 杭州网易云音乐科技有限公司 Face key point detection method, medium, device and computing equipment
CN111008589B (en) * 2019-12-02 2024-04-09 杭州网易云音乐科技有限公司 Face key point detection method, medium, device and computing equipment
CN114936597A (en) * 2022-05-20 2022-08-23 电子科技大学 Method for extracting space true and false target characteristics of local information enhancer
CN114936597B (en) * 2022-05-20 2023-04-07 电子科技大学 Method for extracting space true and false target characteristics of local information enhancer

Also Published As

Publication number Publication date
CN105550657B (en) 2019-01-29

Similar Documents

Publication Publication Date Title
CN105550657A (en) Key point based improved SIFT human face feature extraction method
CN111488804B (en) Labor insurance product wearing condition detection and identity identification method based on deep learning
Mahmood et al. WHITE STAG model: Wise human interaction tracking and estimation (WHITE) using spatio-temporal and angular-geometric (STAG) descriptors
CN101510257B (en) Human face similarity degree matching method and device
US9811718B2 (en) Method and a system for face verification
CN105373777B (en) A kind of method and device for recognition of face
CN106355138A (en) Face recognition method based on deep learning and key features extraction
Setiowati et al. A review of optimization method in face recognition: Comparison deep learning and non-deep learning methods
Wrobel et al. Using a probabilistic neural network for lip-based biometric verification
Wang et al. Feature representation for facial expression recognition based on FACS and LBP
Rai et al. Gender classification using radon and wavelet transforms
CN108446672A (en) A kind of face alignment method based on the estimation of facial contours from thick to thin
CN106611158A (en) Method and equipment for obtaining human body 3D characteristic information
Ngxande et al. Detecting inter-sectional accuracy differences in driver drowsiness detection algorithms
CN104732247A (en) Human face feature positioning method
Amin A face recognition system based on deep learning (FRDLS) to support the entry and supervision procedures on electronic exams
CN105718858A (en) Pedestrian recognition method based on positive-negative generalized max-pooling
Feng et al. The application of scale invariant feature transform fused with shape model in the human face recognition
CN106709442A (en) Human face recognition method
Li et al. Pedestrian detection based on clustered poselet models and hierarchical and–or grammar
Tani et al. Robust human detection to pose and occlusion using bag-of-words
Shen et al. Double complete d-lbp with extreme learning machine auto-encoder and cascade forest for facial expression analysis
Ding et al. Object registration for remote sensing images using robust kernel pattern vectors
Wang et al. Driver Fatigue Detection Using Improved Deep Learning and Personalized Framework
Hafez et al. 3D face recognition based on normal map features using selected Gabor filters and linear discriminant analysis

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant