CN109815889B - Cross-resolution face recognition method based on feature representation set - Google Patents

Cross-resolution face recognition method based on feature representation set Download PDF

Info

Publication number
CN109815889B
CN109815889B CN201910055693.7A CN201910055693A CN109815889B CN 109815889 B CN109815889 B CN 109815889B CN 201910055693 A CN201910055693 A CN 201910055693A CN 109815889 B CN109815889 B CN 109815889B
Authority
CN
China
Prior art keywords
resolution
image
low
image block
vec
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910055693.7A
Other languages
Chinese (zh)
Other versions
CN109815889A (en
Inventor
高广谓
汪焰南
朱冬
吴松松
荆晓远
岳东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Posts and Telecommunications
Original Assignee
Nanjing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Posts and Telecommunications filed Critical Nanjing University of Posts and Telecommunications
Priority to CN201910055693.7A priority Critical patent/CN109815889B/en
Publication of CN109815889A publication Critical patent/CN109815889A/en
Application granted granted Critical
Publication of CN109815889B publication Critical patent/CN109815889B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a cross-resolution face recognition method based on a feature representation set, which comprises the following steps: firstly, acquiring image blocks of each pixel position of a high-resolution training sample image, a low-resolution testing sample image and a high-resolution and low-resolution training dictionary image; then for each image block in the low-quality test image, obtaining linear representation of an image block set of the image block on a corresponding position in the low-quality training dictionary image by using a constrained P-norm regular regression method of the image block, and obtaining linear representation of the image block set of the high-resolution training sample image on the corresponding position of the high-resolution dictionary by using the same method; then carrying out similarity measurement on the low-resolution test image block feature representation set and the high-resolution training image feature representation set; and finally testing the image category. The invention has the advantages that: the face image with inconsistent resolution can be accurately identified, and the problem that the face image is difficult to identify due to inconsistent resolution is effectively solved.

Description

Cross-resolution face recognition method based on feature representation set
Technical Field
The invention relates to the technical field of image processing, in particular to a cross-resolution face recognition method based on a feature representation set.
Background
Human face recognition is a popular research topic in the field of computer vision, and it combines computer image processing technology and statistical technology, and is widely applied in various fields, such as: financial field, public security system, social security field, airport frontier inspection face recognition, and the like. The current face recognition methods can be divided into two categories: a global-based face recognition method and a local-based face recognition method. The global face recognition method retains a global face structure, but neglects face feature details except a principal component; the local face recognition method mostly adopts a face recognition method based on image blocks, and under the constraint of sparse conditions, the face image blocks to be recognized and the training sample image blocks are regarded as linear combinations of the training image blocks to obtain expression coefficient matrixes with the same dimension, so that the recognition work is completed. However, the image block-based face recognition method has a disadvantage: the modules are independent from each other, the associated information between the modules is lost, and because the quality of the face images obtained in most applications is poor and the resolution of the face images is inconsistent in the actual application process, the identity recognition is difficult to complete sometimes, so that the problem that the resolution of the face images is inconsistent and difficult to recognize exists.
Disclosure of Invention
The invention aims to provide a cross-resolution face recognition method based on a feature representation set, which can accurately identify the face images with inconsistent resolution.
In order to realize the purpose, the invention adopts the following technical scheme: a cross-resolution face recognition method based on a feature representation set comprises the following steps:
the method comprises the following steps: acquiring image blocks of each pixel position in a high-resolution training sample image, a low-resolution testing sample image and a high-resolution and low-resolution training dictionary image;
step two: for each image block in the low-resolution test sample image, a regression representation method based on P norm regularization is applied to obtain linear representation of an image block set at a corresponding position on the low-resolution training dictionary image; meanwhile, for each image block in the high-resolution training sample image, a regression representation method based on P norm regularization is applied to obtain linear representation of an image block set at a corresponding position on the high-resolution training dictionary image;
step three: performing similarity measurement on the low-resolution test sample image feature representation set and the high-resolution training sample image feature representation set obtained in the second step;
step four: and D, finishing classification of the low-resolution test sample images according to the data obtained in the step three, and outputting the classification of the low-resolution test sample images.
Further, the foregoing cross-resolution face recognition method based on a feature representation set, wherein: in the second step, a regression representation method based on P-norm regularization is applied to obtain a linear representation of the image block set at the corresponding position on the low-resolution training dictionary image, and the specific method is as follows:
step (1): for a low-resolution test image block, dividing the low-resolution test image block into S blocks, and for a low-resolution test image block y, performing linear representation by using image blocks at corresponding positions on a low-resolution training dictionary sample image:
y=x 1 A 1 +x 2 A 2 +...+x N A N +E
wherein A is i An image block which represents a corresponding position of the ith low-resolution training dictionary image, i ═ 1,2,. and N, N represents the number of low-resolution training sample dictionary sample images, and x represents the number of the low-resolution training sample dictionary sample images i Representing a coefficient corresponding to the ith element of the coefficient vector x, and E represents a residual error item; the solution to represent the vector coefficients x is as follows:
defining slave space
Figure BDA0001951165050000021
To
Figure BDA0001951165050000022
A (x) x 1 A 1 +x 2 A 2 +...+x N A N
For the low-resolution test image block, a linear representation of the low-resolution test image block is obtained on a low-resolution training sample dictionary image by using a P-norm regular regression method, and an objective function is given as follows:
Figure BDA0001951165050000023
s.t.y-A(x)=E
wherein the content of the first and second substances,
Figure BDA0001951165050000024
represents the Schatten-P norm of the matrix,
Figure BDA0001951165050000025
σ i is the ith singular value of E; λ is a regularization parameter, D ═ D 1 ,D 2 ,...,D N ) Representing a Euclidean distance matrix between the low-resolution test image block and the low-resolution training sample dictionary image block; p is between (0,1), P is chosen 1/2, Schatten-1/2 norm closer to the rank function; the above model is represented as:
Figure BDA0001951165050000026
s.t.y-A(x)=E
its lagrange function is expressed as:
Figure BDA0001951165050000027
where μ > 0 is a penalty parameter, Z is the lagrange multiplier, tr (-) is the trace operation, the above equation can be written as:
Figure BDA0001951165050000028
the model is solved by adopting an alternative direction multiplier method ADMM, and the specific process is as follows:
< a > fix E, update x:
Figure BDA0001951165050000031
the solution of the above formula is:
x k+1 =(G+τD 2 )\ones(N,1)
wherein ones (M,1) is an elementAn M × 1-dimensional vector with a 1 element, "\" denotes a left matrix division operation, τ ═ 2 λ μ, and G is a covariance matrix G ═ C T C:
Figure BDA0001951165050000032
Wherein, H ═ vec (A) 1 ,Vec(A 2 ),...,Vec(A N )]Vec (·) denotes the vectorization operator.
< b > fix x, update E:
Figure BDA0001951165050000033
to solve the above equation, a single-valued function threshold theorem based on the Schatten-1/2 norm is introduced:
theorem: given a constant of eta > 0 and a matrix of rank r
Figure BDA0001951165050000034
The objective function becomes:
Figure BDA0001951165050000035
Figure BDA0001951165050000036
wherein sigma 12 ,...,σ r Is the positive singular value of G, U l×r And V m×r Is a corresponding matrix with orthogonal columns;
Figure BDA0001951165050000037
Figure BDA0001951165050000038
and is provided with
Figure BDA0001951165050000039
According to the above theorem, the objective function becomes:
Figure BDA00019511650500000310
<c>selecting a suitable termination parameter epsilon 1 And ε 2 The following termination conditions are satisfied:
||y-A(x k+1 )-E||/||y||<ε 1
max(||x k+1 -x k ||,||E k+1 -E k ||)/||y||<ε 2
wherein, | | | | is a given matrix norm, and if the maximum iteration number is reached or the termination condition is met, x is output k+1 As x, otherwise, returning to step<a>;
Step (2): obtaining the expression coefficient vector of each low-resolution image block, and integrating the expression coefficient vectors into a coefficient matrix X with dimension of NxSxM; and M is the number of samples in the test set, S is the number of low-resolution test image blocks, and N represents the number of low-resolution training sample dictionary sample images.
Further, the foregoing cross-resolution face recognition method based on a feature representation set, wherein: in the second step, a regression representation method based on P-norm regularization is used to obtain a linear representation of the image block set at the corresponding position on the high-resolution training dictionary image, and the specific method is as follows:
step (1): for a high resolution training image block, divide it into S blocks, for a high resolution training image block y 1 And performing linear representation by using image blocks at corresponding positions on the high-resolution training dictionary sample image:
y 1 =c 1 G 1 +c 2 G 2 +...+c N G N +E 2
wherein G is i Representing the image block at the corresponding position of the ith low-resolution training dictionary image, i ═ 1, 2.. multidot.N }, N represents the number of high-resolution training sample dictionary sample images, c i Representing coefficient vector c ithCoefficient of element correspondence, E 2 Representing a residual term; the solution to represent the vector coefficients c is as follows:
defining slave space
Figure BDA0001951165050000044
To
Figure BDA0001951165050000045
A linear mapping of g (c) ═ c 1 G 1 +c 2 G 2 +...+c N G N
For the low-resolution test image block, a linear representation of the low-resolution test image block is obtained on a low-resolution training sample dictionary image by using a P-norm regular regression method, and an objective function is given as follows:
Figure BDA0001951165050000041
s.t.y 1 -G(c)=E 2
wherein the content of the first and second substances,
Figure BDA0001951165050000042
representing the Schatten-P norm of the matrix,
Figure BDA0001951165050000043
σ i is the ith singular value of E; λ is the regularization parameter, D ═ D 1 ,D 2 ,...,D N ) Representing a Euclidean distance matrix between the low-resolution test image block and the low-resolution training sample dictionary image block; p is between (0,1), P is chosen 1/2, Schatten-1/2 norm closer to the rank function; the above model is represented as:
Figure BDA0001951165050000051
s.t.y 1 -G(c)=E 2
its lagrange function is expressed as:
Figure BDA0001951165050000052
where μ > 0 is a penalty parameter, Z is the lagrange multiplier, tr (-) is the trace operation, and the above equation can be written as:
Figure BDA0001951165050000053
the model is solved by adopting an alternating direction multiplier method ADMM, and the specific process is as follows:
<a>fixing E 2 And updating c:
Figure BDA0001951165050000054
the solution of the above formula is:
c k+1 =(O+τD 2 )\ones(N,1)
where ons (M,1) is an M × 1-dimensional vector with an element of 1, "\" denotes a left matrix division operation, τ ═ 2 λ μ, and O is a covariance matrix O ═ C T C:
Figure BDA0001951165050000055
Wherein H ═ vec (G) 1 ,Vec(G 2 ),...,Vec(G N )]Vec (·) denotes the vectorization operator;
<b>fix c, update E 2
Figure BDA0001951165050000056
To solve the above equation, a single-valued function threshold theorem based on the Schatten-1/2 norm is introduced:
theorem: given a constant of eta > 0 and a matrix of rank r
Figure BDA0001951165050000057
The objective function becomes:
Figure BDA0001951165050000061
Figure BDA0001951165050000062
wherein sigma 12 ,...,σ r Is G 1 Positive singular value of (U), U l×r And V m×r Is a corresponding matrix with orthogonal columns;
Figure BDA0001951165050000063
Figure BDA0001951165050000064
and is
Figure BDA0001951165050000065
According to the above theorem, the objective function becomes:
Figure BDA0001951165050000066
<c>selecting a suitable termination parameter ε 1 And ε 2 The following termination conditions are satisfied:
||y 1 -G(c k+1 )-E 2 ||/||y 1 ||<ε 1
max(||c k+1 -c k ||,||E 2 k+1 -E 2 k |)/||y 1 ||<ε 2
wherein, | | | | is a given matrix norm, and if the maximum iteration number is reached or the termination condition is met, c is output k+1 As c, otherwise, returning to step<a>;
Step (2): obtaining the representative coefficient vector of each high-resolution training image block, and applying themIntegrating to obtain a matrix of expression coefficients C ═ C 1 ,C 2 ,...,C K ) Dimension is nxsxl; wherein L represents the total number of high-resolution face sample images, C i And representing a coefficient matrix of the ith high-resolution training sample face image set on the high-resolution training dictionary, wherein S is the number of high-resolution training image blocks, and N is the number of high-resolution training sample dictionary sample images.
Further, the foregoing cross-resolution face recognition method based on a feature representation set, wherein: in step three, the specific method for performing similarity measurement on the low-resolution test sample image feature representation set and the high-resolution training sample image feature representation set is as follows:
step (31): establishing a similarity measurement model as follows:
Figure BDA0001951165050000067
s.t.∑α i =1
wherein α and β represent coefficient vectors, and X (α) ═ α 1 X 12 X 2 +...+α M X M ,C(β)=β 1 C 12 C 2 +...+β L C L ,∑α i Avoiding an explicit solution (α ═ 0), the above equation can be written as:
Figure BDA0001951165050000071
s.t.E 1 =X(α)-C(β)
the lagrangian form of the above equation is:
Figure BDA0001951165050000072
step (32): fix α, β, update E 1
Figure BDA0001951165050000073
The optimal solution of the above formula can be obtained through a singular value threshold operator, and a matrix with the rank r is given
Figure BDA0001951165050000074
Its singular value decomposition is:
Figure BDA0001951165050000075
wherein σ 1 ,...σ r Representing singular values, U and V being orthogonal matrices; for a given one τ>0, singular value operator T τ (. cndot.) is defined as:
Figure BDA0001951165050000076
the solution to the above problem is:
Figure BDA0001951165050000077
step (33): fixing E 1 Updating α, β:
the lagrangian equation above becomes:
Figure BDA0001951165050000078
where e is a row vector with elements all being 1, Vec (X) ═ Vec (X) 1 )Vec(X 2 )...Vec(X M )], Vec(-C)=[Vec(-C 1 )Vec(-C 2 )...Vec(-C M )](ii) a Is provided with
Figure BDA0001951165050000081
The above equation becomes:
Figure BDA0001951165050000082
Figure BDA0001951165050000083
Figure BDA0001951165050000084
from the above equation can be calculated:
Figure BDA0001951165050000085
wherein h is 0 =(J T J+K) -1
Step (34): selecting a suitable termination parameter ε 3 The following termination conditions are satisfied:
||X(α k )-C(β k )-E 1 ||/||X(α k )||<ε 3
wherein, | | | | is a given matrix norm, and if the maximum iteration number is reached or the termination condition is met, alpha is output k+1 As α, β k+1 As beta, otherwise, returning to step (32).
Further, the foregoing cross-resolution face recognition method based on a feature representation set, wherein: in step four, each group C was used i The residual error of X alpha is expressed to determine the class label of X, and the proposed classifier is as follows:
identity(X)=argmin i (r i )
test sample X and a training sample C i The distance between is expressed as
Figure BDA0001951165050000086
Wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0001951165050000087
and
Figure BDA0001951165050000088
represents the optimal coefficient vector, | · | calculation * Represents the nuclear norm, and identity (X) represents the class label of the test sample X.
Through the implementation of the technical scheme, the invention has the beneficial effects that: the method can accurately identify the face images with inconsistent resolution, and effectively solves the problem that the face images are difficult to identify due to inconsistent resolution.
Drawings
Fig. 1 is a schematic flow chart of a cross-resolution face recognition method based on a feature representation set according to the present invention.
Detailed Description
The invention is described in further detail below with reference to the figures and specific examples.
As shown in fig. 1, the cross-resolution face recognition method based on the feature representation set includes the following steps:
the method comprises the following steps: acquiring image blocks of each pixel position in a high-resolution training sample image, a low-resolution testing sample image and a high-resolution and low-resolution training dictionary image;
step two: for each image block in the low-resolution test sample image, linear representation of an image block set at a corresponding position on the low-resolution training dictionary image is obtained by applying a regression representation method based on P-norm regularization; meanwhile, for each image block in the high-resolution training sample image, a regression representation method based on P norm regularization is applied to obtain linear representation of an image block set at a corresponding position on the high-resolution training dictionary image;
the specific method for obtaining the linear representation of the image block set at the corresponding position on the low-resolution training dictionary image by applying the regression representation method based on the P norm regulation is as follows:
step (1): for a low-resolution test image block, dividing the low-resolution test image block into S blocks, and for a low-resolution test image block y, performing linear representation by using image blocks at corresponding positions on a low-resolution training dictionary sample image:
y=x 1 A 1 +x 2 A 2 +...+x N A N +E
wherein A is i An image block which represents a corresponding position of the ith low-resolution training dictionary image, i ═ 1,2,. and N, N represents the number of low-resolution training sample dictionary sample images, and x represents the number of the low-resolution training sample dictionary sample images i Representing a coefficient corresponding to the ith element of the coefficient vector x, and E represents a residual error item; the solution to represent the vector coefficients x is as follows:
defining slave space
Figure BDA0001951165050000095
To
Figure BDA0001951165050000096
A (x) x 1 A 1 +x 2 A 2 +...+x N A N
For the low-resolution test image block, a P-norm regular regression method is used for obtaining linear representation of the low-resolution test image block on a low-resolution training sample dictionary image, and an objective function is given as follows:
Figure BDA0001951165050000091
s.t.y-A(x)=E
wherein the content of the first and second substances,
Figure BDA0001951165050000092
represents the Schatten-P norm of the matrix,
Figure BDA0001951165050000093
σ i is the ith singular value of E; λ is the regularization parameter, D ═ D 1 ,D 2 ,...,D N ) Representing a Euclidean distance matrix between a low-resolution test image block and a low-resolution training sample dictionary image block; p is between (0,1), P is chosen 1/2, Schatten-1/2 norm is closer to rankA function; the above model is represented as:
Figure BDA0001951165050000094
s.t.y-A(x)=E
its lagrange function is expressed as:
Figure BDA0001951165050000101
where μ > 0 is a penalty parameter, Z is the lagrange multiplier, tr (-) is the trace operation, the above equation can be written as:
Figure BDA0001951165050000102
the model is solved by adopting an alternative direction multiplier method ADMM, and the specific process is as follows:
< a > fix E, update x:
Figure BDA0001951165050000103
the solution of the above formula is:
x k+1 =(G+τD 2 )\ones(N,1)
where ons (M,1) is an M × 1-dimensional vector with an element of 1, "\" denotes a left matrix division operation, τ ═ 2 λ μ, and G is a covariance matrix G ═ C T C:
Figure BDA0001951165050000104
Wherein, H ═ vec (A) 1 ,Vec(A 2 ),...,Vec(A N )]Vec (·) denotes the vectorization operator.
< b > fix x, update E:
Figure BDA0001951165050000105
to solve the above equation, a single-valued function threshold theorem based on the Schatten-1/2 norm is introduced:
the theorem is as follows: given a constant of eta > 0 and a matrix of rank r
Figure BDA0001951165050000106
The objective function becomes:
Figure BDA0001951165050000107
Figure BDA0001951165050000108
wherein σ 12 ,...,σ r Is the positive singular value of G, U l×r And V m×r Is a corresponding matrix with orthogonal columns;
Figure BDA0001951165050000109
Figure BDA0001951165050000111
and is
Figure BDA0001951165050000112
According to the above theorem, the objective function becomes:
Figure BDA0001951165050000113
<c>selecting a suitable termination parameter epsilon 1 And ε 2 The following termination conditions are satisfied:
||y-A(x k+1 )-E||/||y||<ε 1
max(||x k+1 -x k ||,||E k+1 -E k ||)/||y||<ε 2
wherein, | | | | is a given matrix norm, and if the maximum iteration number is reached or the termination condition is met, x is output k+1 As x, otherwise, returning to step<a>;
Step (2): obtaining the expression coefficient vector of each low-resolution image block, and integrating the expression coefficient vectors into a coefficient matrix X with dimension of NxSxM; wherein M is the number of samples in the test set, S is the number of low-resolution test image blocks, and N represents the number of low-resolution training sample dictionary sample images;
the specific method for obtaining the linear representation of the image block set at the corresponding position on the high-resolution training dictionary image by applying the regression representation method based on the P norm regulation is as follows:
step (1): for a high resolution training image block, divide it into S blocks, for a high resolution training image block y 1 And performing linear representation by using image blocks at corresponding positions on the high-resolution training dictionary sample image:
y 1 =c 1 G 1 +c 2 G 2 +...+c N G N +E 2
wherein G is i Representing the image block at the corresponding position of the ith low-resolution training dictionary image, i ═ 1, 2.. multidot.N }, N represents the number of high-resolution training sample dictionary sample images, c i Represents the coefficient corresponding to the ith element of the coefficient vector c, E 2 Representing a residual term; the solution to represent the vector coefficients c is as follows:
defining slave space
Figure BDA0001951165050000117
To
Figure BDA0001951165050000118
A linear mapping of g (c) ═ c 1 G 1 +c 2 G 2 +...+c N G N
For the low-resolution test image block, a linear representation of the low-resolution test image block is obtained on a low-resolution training sample dictionary image by using a P-norm regular regression method, and an objective function is given as follows:
Figure BDA0001951165050000114
s.t.y 1 -G(c)=E 2
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0001951165050000115
represents the Schatten-P norm of the matrix,
Figure BDA0001951165050000116
σ i is the ith singular value of E; λ is the regularization parameter, D ═ D 1 ,D 2 ,...,D N ) Representing a Euclidean distance matrix between a low-resolution test image block and a low-resolution training sample dictionary image block; p is between (0,1), P is chosen 1/2, Schatten-1/2 norm closer to the rank function; the above model is represented as:
Figure BDA0001951165050000121
s.t.y 1 -G(c)=E 2
its lagrange function is expressed as:
Figure BDA0001951165050000122
where μ > 0 is a penalty parameter, Z is the lagrange multiplier, tr (-) is the trace operation, and the above equation can be written as:
Figure BDA0001951165050000123
the model is solved by adopting an alternative direction multiplier method ADMM, and the specific process is as follows:
<a>fixing E 2 And updating c:
Figure BDA0001951165050000124
the solution of the above formula is:
c k+1 =(O+τD 2 )\ones(N,1)
where ons (M,1) is an M × 1-dimensional vector with an element of 1, "\" denotes a left matrix division operation, τ ═ 2 λ μ, and O is a covariance matrix O ═ C T C:
Figure BDA0001951165050000125
Wherein H ═ vec (G) 1 ,Vec(G 2 ),...,Vec(G N )]Vec (·) denotes the vectorization operator;
<b>fix c, update E 2
Figure BDA0001951165050000126
To solve the above equation, a single-valued function threshold theorem based on the Schatten-1/2 norm is introduced:
theorem: given a constant of eta > 0 and a matrix of rank r
Figure BDA0001951165050000131
The objective function becomes:
Figure BDA0001951165050000132
Figure BDA0001951165050000133
wherein sigma 12 ,...,σ r Is G 1 Positive singular value of (U), U l×r And V m×r Is a corresponding matrix with orthogonal columns;
Figure BDA0001951165050000134
Figure BDA0001951165050000135
and is
Figure BDA0001951165050000136
According to the above theorem, the objective function becomes:
Figure BDA0001951165050000137
<c>selecting a suitable termination parameter epsilon 1 And epsilon 2 The following termination conditions are satisfied:
||y 1 -G(c k+1 )-E 2 ||/||y 1 ||<ε 1
max(||c k+1 -c k ||,||E 2 k+1 -E 2 k ||)/||y 1 ||<ε 2
wherein, | | | | is a given matrix norm, and if the maximum iteration number is reached or the termination condition is met, c is output k+1 As c, otherwise, returning to the step<a>;
Step (2): obtaining the expression coefficient vector of each high-resolution training image block, and integrating the expression coefficient vectors to obtain an expression coefficient matrix C ═ (C) 1 ,C 2 ,...,C K ) Dimension is nxsxl; wherein L represents the total number of high-resolution face sample images, C i Representing a coefficient matrix of an ith high-resolution training sample face image set on a high-resolution training dictionary, wherein S is the number of high-resolution training image blocks, and N is the number of high-resolution training sample dictionary sample images;
step three: performing similarity measurement on the low-resolution test sample image feature representation set and the high-resolution training sample image feature representation set obtained in the second step;
the specific method for carrying out similarity measurement on the low-resolution test sample image feature representation set and the high-resolution training sample image feature representation set is as follows:
step (31): establishing a similarity measurement model as follows:
Figure BDA0001951165050000138
s.t.∑α i =1
wherein α and β represent coefficient vectors, and X (α) is α 1 X 12 X 2 +...+α M X M ,C(β)=β 1 C 12 C 2 +...+β L C L ,∑α i Avoiding an explicit solution (α ═ 0), the above equation can be written as:
Figure BDA0001951165050000141
s.t.E 1 =X(α)-C(β)
the lagrangian form of the above equation is:
Figure BDA0001951165050000142
step (32): fix α, β, update E 1
Figure BDA0001951165050000143
The optimal solution of the above formula can be obtained through a singular value threshold operator, and a matrix with the rank r is given
Figure BDA0001951165050000144
Its singular value decomposition is:
Figure BDA0001951165050000145
wherein σ 1 ,...σ r Representing singular values, U and V being orthogonal matrices; for a given one τ>0, singular value operator T τ (. cndot.) is defined as:
Figure BDA0001951165050000146
the solution to the above problem is:
Figure BDA0001951165050000147
step (33): fixing E 1 Updating α, β:
the lagrangian equation above becomes:
Figure BDA0001951165050000151
where e is a row vector with elements all being 1, Vec (X) ═ Vec (X) 1 )Vec(X 2 )...Vec(X M )], Vec(-C)=[Vec(-C 1 )Vec(-C 2 )...Vec(-C M )](ii) a Is provided with
Figure BDA0001951165050000152
The above equation becomes:
Figure BDA0001951165050000153
Figure BDA0001951165050000154
Figure BDA0001951165050000155
from the above equation:
Figure BDA0001951165050000156
wherein h is 0 =(J T J+K) -1
Step (34): selecting a suitable termination parameter ε 3 The following termination conditions are satisfied:
||X(α k )-C(β k )-E 1 ||/||X(α k )||<ε 3
wherein, | | | | is a given matrix norm, and if the maximum iteration number is reached or the termination condition is met, alpha is output k+1 As α, β k+1 As beta, otherwise, returning to the step (32);
step four: classifying the low-resolution test sample images according to the data obtained in the step three, and outputting the classes of the low-resolution test sample images;
wherein each group C is used i The residual error of X alpha is expressed to determine the class label of X, and the proposed classifier is as follows:
identity(X)=argmin i (r i )
test sample X and a training sample C i The distance between is expressed as
Figure BDA0001951165050000161
Wherein the content of the first and second substances,
Figure BDA0001951165050000162
and
Figure BDA0001951165050000163
represents the optimal coefficient vector, | · | calculation * Represents the nuclear norm, and identity (X) represents the class label of the test sample X.
The invention has the advantages that: the face image with inconsistent resolution can be accurately identified, and the problem that the face image is difficult to identify due to inconsistent resolution is effectively solved.

Claims (3)

1. A cross-resolution face recognition method based on a feature representation set is characterized in that: the method comprises the following steps:
the method comprises the following steps: acquiring image blocks of each pixel position in a high-resolution training sample image, a low-resolution testing sample image and a high-resolution and low-resolution training dictionary image;
step two: for each image block in the low-resolution test sample image, a regression representation method based on P norm regularization is used for obtaining linear representation of an image block set at a corresponding position on the low-resolution training dictionary image; meanwhile, for each image block in the high-resolution training sample image, a regression representation method based on P norm regularization is used for obtaining linear representation of an image block set at a corresponding position on the high-resolution training dictionary image;
the specific method for obtaining the linear representation of the image block set at the corresponding position on the low-resolution training dictionary image by applying the regression representation method based on the P norm regulation is as follows:
step (1): for a low-resolution test image block, dividing the low-resolution test image block into S blocks, and for a low-resolution test image block y, performing linear representation by using image blocks at corresponding positions on a low-resolution training dictionary sample image:
y=x 1 A 1 +x 2 A 2 +...+x N A N +E
wherein A is i An image block which represents a corresponding position of the ith low-resolution training dictionary image, i ═ 1,2,. and N, N represents the number of low-resolution training sample dictionary sample images, and x represents the number of the low-resolution training sample dictionary sample images i Representing a coefficient corresponding to the ith element of the coefficient vector x, and E represents a residual error item; the solution to represent the vector coefficients x is as follows:
defining slave space
Figure FDA0003717627290000011
To
Figure FDA0003717627290000012
A linear mapping ofRay A (x) x 1 A 1 +x 2 A 2 +...+x N A N
For the low-resolution test image block, a linear representation of the low-resolution test image block is obtained on a low-resolution training sample dictionary image by using a P-norm regular regression method, and an objective function is given as follows:
Figure FDA0003717627290000013
s.t.y-A(x)=E
wherein the content of the first and second substances,
Figure FDA0003717627290000014
representing the Schatten-P norm of the matrix,
Figure FDA0003717627290000015
σ i is the ith singular value of E; λ is a regularization parameter, D ═ D 1 ,D 2 ,...,D N ) Representing a Euclidean distance matrix between the low-resolution test image block and the low-resolution training sample dictionary image block; p is between (0,1), P is chosen 1/2, Schatten-1/2 norm closer to the rank function; the model is represented as:
Figure FDA0003717627290000016
s.t.y-A(x)=E
its lagrange function is expressed as:
Figure FDA0003717627290000021
where μ > 0 is a penalty parameter, Z is the lagrange multiplier, tr (-) is the trace operation, and the above equation can be written as:
Figure FDA0003717627290000022
the model is solved by adopting an alternative direction multiplier method ADMM, and the specific process is as follows:
< a > fix E, update x:
Figure FDA0003717627290000023
the solution of the above formula is:
x k+1 =(G+τD 2 )\ones(N,1)
where ones (M,1) is an M × 1-dimensional vector with an element of 1, "\" indicates a left matrix division operation, τ ═ 2 λ/μ, and G is a covariance matrix G ═ C T C:
Figure FDA0003717627290000024
Wherein, H ═ vec (A) 1 ) ,Vec(A 2 ),...,Vec(A N )]Vec (·) denotes the vectorization operator;
< b > fix x, update E:
Figure FDA0003717627290000025
to solve the above equation, a single-valued function threshold theorem based on the Schatten-1/2 norm is introduced:
theorem: given a constant of eta > 0 and a matrix of rank r
Figure FDA0003717627290000026
The objective function becomes:
Figure FDA0003717627290000027
Figure FDA0003717627290000028
wherein σ 12 ,...,σ r Is the positive singular value of G, U l×r And V m×r Is a corresponding matrix with orthogonal columns;
Figure FDA0003717627290000029
Figure FDA0003717627290000031
and is provided with
Figure FDA0003717627290000032
According to the above theorem, the objective function becomes:
Figure FDA0003717627290000033
<c>selecting a termination parameter ε 1 And ε 2 The following termination conditions are satisfied:
||y-A(x k+1 )-E||/||y||<ε 1
max(||x k+1 -x k ||,||E k+1 -E k ||)/||y||<ε 2
wherein, | | | | is a given matrix norm, if the maximum iteration times is reached or the termination condition is met, x is output k+1 As x, otherwise, returning to step<a>;
Step (2): obtaining the expression coefficient vector of each low-resolution image block, and integrating the expression coefficient vectors into a coefficient matrix X with the dimension of NxSxM; wherein M is the number of samples in the test set, S is the number of low-resolution test image blocks, and N represents the number of low-resolution training sample dictionary sample images;
the specific method for obtaining the linear representation of the image block set at the corresponding position on the high-resolution training dictionary image by applying the regression representation method based on the P norm regulation is as follows:
step (1): for a high resolution training image block, divide it into S blocks, for a high resolution training image block y 1 And performing linear representation by using image blocks at corresponding positions on the high-resolution training dictionary sample image:
y 1 =c 1 G 1 +c 2 G 2 +...+c N G N +E 2
wherein G is i Representing the image block at the corresponding position of the ith low-resolution training dictionary image, i ═ 1, 2.. multidot.N }, N represents the number of high-resolution training sample dictionary sample images, c i Represents the coefficient corresponding to the ith element of the coefficient vector c, E 2 Representing a residual term; the solution to represent the vector coefficients c is as follows:
defining slave space
Figure FDA0003717627290000034
To
Figure FDA0003717627290000035
A linear mapping of g (c) ═ c 1 G 1 +c 2 G 2 +...+c N G N
For the low-resolution test image block, a linear representation of the low-resolution test image block is obtained on a low-resolution training sample dictionary image by using a P-norm regular regression method, and an objective function is given as follows:
Figure FDA0003717627290000036
s.t.y 1 -G(c)=E 2
wherein the content of the first and second substances,
Figure FDA0003717627290000037
represents the Schatten-P norm of the matrix,
Figure FDA0003717627290000038
σ i is the ith singularity of EA value; λ is a regularization parameter, D ═ D 1 ,D 2 ,...,D N ) Representing a Euclidean distance matrix between the low-resolution test image block and the low-resolution training sample dictionary image block; p is between (0,1), P is chosen to be 1/2, Schatten-1/2 norm is closer to the rank function; the model is represented as:
Figure FDA0003717627290000041
s.t.y 1 -G(c)=E 2
its lagrange function is expressed as:
Figure FDA0003717627290000042
where μ > 0 is a penalty parameter, Z is the lagrange multiplier, tr (-) is the trace operation, and the above equation can be written as:
Figure FDA0003717627290000043
solving the model by adopting an alternative direction multiplier method ADMM, which comprises the following specific steps:
<a>fixing E 2 And updating c:
Figure FDA0003717627290000044
the solution of the above formula is:
c k+1 =(O+τD 2 )\ones(N,1)
where ons (M,1) is an M × 1-dimensional vector with an element of 1, "\" denotes a left matrix division operation, τ ═ 2 λ/μ, and O is a covariance matrix O ═ C T C:
Figure FDA0003717627290000045
Wherein H ═ vec (G) 1 ) ,Vec(G 2 ),...,Vec(G N )]Vec (·) represents the vectorization operator;
<b>fix c, update E 2
Figure FDA0003717627290000046
To solve the above equation, a single-valued function threshold theorem based on the Schatten-1/2 norm is introduced:
theorem: given a constant of eta > 0 and a matrix of rank r
Figure FDA0003717627290000051
The objective function becomes:
Figure FDA0003717627290000052
Figure FDA0003717627290000053
wherein sigma 12 ,...,σ r Is G 1 Positive singular value of (U), U l×r And V m×r Is a corresponding matrix with orthogonal columns;
Figure FDA0003717627290000054
Figure FDA0003717627290000055
and is
Figure FDA0003717627290000056
According to the above theorem, the objective function becomes:
Figure FDA0003717627290000057
<c>selecting a termination parameter ε 1 And ε 2 The following termination conditions are satisfied:
||y 1 -G(c k+1 )-E 2 ||/||y 1 ||<ε 1
max(||c k+1 -c k ||,||E 2 k+1 -E 2 k ||)/||y 1 ||<ε 2
wherein, | | | | is a given matrix norm, and if the maximum iteration number is reached or a termination condition is met, c is output k+1 As c, otherwise, returning to step<a>;
Step (2): obtaining a vector of the representation coefficients of each high-resolution training image block, and integrating the vectors to obtain a matrix of the representation coefficients C ═ C 1 ,C 2 ,...,C K ) Dimension is nxsxl; wherein L represents the total number of high-resolution face sample images, C i Representing a coefficient matrix of an ith high-resolution training sample face image set on a high-resolution training dictionary, wherein S is the number of high-resolution training image blocks, and N is the number of high-resolution training sample dictionary sample images;
step three: performing similarity measurement on the low-resolution test sample image feature representation set and the high-resolution training sample image feature representation set obtained in the second step;
step four: and D, finishing classification of the low-resolution test sample images according to the data obtained in the step three, and outputting the classification of the low-resolution test sample images.
2. The cross-resolution face recognition method based on the feature representation set according to claim 1, characterized in that: in step three, the specific method for performing similarity measurement on the low-resolution test sample image feature representation set and the high-resolution training sample image feature representation set is as follows:
step (31): establishing a similarity measurement model as follows:
Figure FDA0003717627290000061
s.t.∑α i =1
wherein α and β represent coefficient vectors, and X (α) ═ α 1 X 12 X 2 +...+α M X M ,C(β)=β 1 C 12 C 2 +...+β L C L ,∑α i Avoiding an explicit solution (α ═ 0), the above equation can be written as:
Figure FDA0003717627290000062
s.t.E 1 =X(α)-C(β)
the lagrange form of the above equation is:
Figure FDA0003717627290000063
step (32): fix α, β, update E 1
Figure FDA0003717627290000064
The optimal solution of the above formula can be obtained through a singular value threshold operator, and a matrix with the rank r is given
Figure FDA0003717627290000065
Its singular value decomposition is:
Figure FDA0003717627290000066
wherein sigma 1 ,...σ r Representing singular values, U and V being orthogonal matrices; for theGiven a τ>0, singular value operator T τ (. cndot.) is defined as:
Figure FDA0003717627290000067
the solution to the problem is:
Figure FDA0003717627290000068
step (33): fixing E 1 Updating α, β:
the lagrangian equation becomes:
Figure FDA0003717627290000071
where e is a row vector with elements all being 1, Vec (X) ═ Vec (X) 1 )Vec(X 2 )...Vec(X M )],Vec(-C)=[Vec(-C 1 )Vec(-C 2 )...Vec(-C M )](ii) a Is provided with
J=[Vec(X)Vec(-C)],
Figure FDA0003717627290000072
d=[e 0] T ,
Figure FDA0003717627290000073
The above equation becomes:
Figure FDA0003717627290000074
Figure FDA0003717627290000075
Figure FDA0003717627290000076
it can be calculated from the equation:
Figure FDA0003717627290000077
wherein h is 0 =(J T J+K) -1
Step (34): selecting a termination parameter epsilon 3 The following termination conditions are satisfied:
||X(α k )-C(β k )-E 1 ||/||X(α k )||<ε 3
wherein, | | | | is a given matrix norm, and if the maximum iteration number is reached or a termination condition is met, alpha is output k+1 As α, β k+1 As beta, otherwise, returning to step (32).
3. The method of claim 2, wherein the method comprises the following steps: in step four, each group C was used i The residual error of X alpha is expressed to determine the class label of X, and the proposed classifier is as follows:
identity(X)=argmin i (r i )
test sample X and a training sample C i The distance between is expressed as
Figure FDA0003717627290000081
Wherein the content of the first and second substances,
Figure FDA0003717627290000082
and
Figure FDA0003717627290000083
represents the optimal coefficient vector, | · | calculation * Represents the nuclear norm, and identity (X) represents the class label of the test sample X.
CN201910055693.7A 2019-01-21 2019-01-21 Cross-resolution face recognition method based on feature representation set Active CN109815889B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910055693.7A CN109815889B (en) 2019-01-21 2019-01-21 Cross-resolution face recognition method based on feature representation set

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910055693.7A CN109815889B (en) 2019-01-21 2019-01-21 Cross-resolution face recognition method based on feature representation set

Publications (2)

Publication Number Publication Date
CN109815889A CN109815889A (en) 2019-05-28
CN109815889B true CN109815889B (en) 2022-08-30

Family

ID=66604643

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910055693.7A Active CN109815889B (en) 2019-01-21 2019-01-21 Cross-resolution face recognition method based on feature representation set

Country Status (1)

Country Link
CN (1) CN109815889B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110443255B (en) * 2019-06-14 2022-04-15 深圳大学 Image recognition method for image feature extraction
CN111444802B (en) * 2020-03-18 2022-05-20 重庆邮电大学 Face recognition method and device and intelligent terminal

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106157251A (en) * 2015-04-01 2016-11-23 武汉大学 A kind of face super-resolution method based on Cauchy's regularization
CN107194873A (en) * 2017-05-11 2017-09-22 南京邮电大学 Low-rank nuclear norm canonical facial image ultra-resolution method based on coupling dictionary learning
KR101829287B1 (en) * 2016-11-29 2018-02-14 인천대학교 산학협력단 Nonsubsampled Contourlet Transform Based Infrared Image Super-Resolution

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106157251A (en) * 2015-04-01 2016-11-23 武汉大学 A kind of face super-resolution method based on Cauchy's regularization
KR101829287B1 (en) * 2016-11-29 2018-02-14 인천대학교 산학협력단 Nonsubsampled Contourlet Transform Based Infrared Image Super-Resolution
CN107194873A (en) * 2017-05-11 2017-09-22 南京邮电大学 Low-rank nuclear norm canonical facial image ultra-resolution method based on coupling dictionary learning

Also Published As

Publication number Publication date
CN109815889A (en) 2019-05-28

Similar Documents

Publication Publication Date Title
CN108133232B (en) Radar high-resolution range profile target identification method based on statistical dictionary learning
CN109060001B (en) Multi-working-condition process soft measurement modeling method based on feature transfer learning
CN107273927B (en) Unsupervised field adaptive classification method based on inter-class matching
CN110135459B (en) Zero sample classification method based on double-triple depth measurement learning network
CN108734199B (en) Hyperspectral image robust classification method based on segmented depth features and low-rank representation
CN105574548A (en) Hyperspectral data dimensionality-reduction method based on sparse and low-rank representation graph
CN105787516A (en) High-spectral image classification method base on space spectral locality low-rank hypergraph learning
CN108460400B (en) Hyperspectral image classification method combining various characteristic information
CN107392107A (en) A kind of face feature extraction method based on isomery tensor resolution
CN111598167B (en) Small sample image identification method and system based on graph learning
CN110490894B (en) Video foreground and background separation method based on improved low-rank sparse decomposition
CN110598636B (en) Ship target identification method based on feature migration
CN109815889B (en) Cross-resolution face recognition method based on feature representation set
CN106971189B (en) A kind of noisy method for recognising star map of low resolution
CN108985161B (en) Low-rank sparse representation image feature learning method based on Laplace regularization
CN110705384B (en) Vehicle re-identification method based on cross-domain migration enhanced representation
CN110659663A (en) Unsupervised bidirectional reconstruction field self-adaption method
CN106908774B (en) One-dimensional range profile identification method based on multi-scale nuclear sparse preserving projection
CN112132257A (en) Neural network model training method based on pyramid pooling and long-term memory structure
CN104866905A (en) Nonparametric sparse tensor dictionary learning method based on beta process
CN108388918B (en) Data feature selection method with structure retention characteristics
CN106960225B (en) sparse image classification method based on low-rank supervision
CN111695570B (en) Small sample lower semantic segmentation method based on variational prototype reasoning
CN110363713B (en) Hyperspectral image noise reduction method based on recursive sample scaling and bilinear factorization
Vilasini et al. Deep learning techniques to detect learning disabilities among children using handwriting

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant