CN102629374B - Image super resolution (SR) reconstruction method based on subspace projection and neighborhood embedding - Google Patents

Image super resolution (SR) reconstruction method based on subspace projection and neighborhood embedding Download PDF

Info

Publication number
CN102629374B
CN102629374B CN201210049804.1A CN201210049804A CN102629374B CN 102629374 B CN102629374 B CN 102629374B CN 201210049804 A CN201210049804 A CN 201210049804A CN 102629374 B CN102629374 B CN 102629374B
Authority
CN
China
Prior art keywords
prime
resolution
training
image
proper vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201210049804.1A
Other languages
Chinese (zh)
Other versions
CN102629374A (en
Inventor
李小燕
和红杰
尹忠科
陈帆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southwest Jiaotong University
Original Assignee
Southwest Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southwest Jiaotong University filed Critical Southwest Jiaotong University
Priority to CN201210049804.1A priority Critical patent/CN102629374B/en
Publication of CN102629374A publication Critical patent/CN102629374A/en
Application granted granted Critical
Publication of CN102629374B publication Critical patent/CN102629374B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses an image super resolution (SR) reconstruction method based on subspace projection and neighborhood embedding. The method is characterized by: using first and secondary subspace projection methods to project original high-dimensional data to a low-dimensional space, using dimension reduction feature vectors to show a feature of a low-resolution image block so that global structure information and local structure information of original data can be maintained; comparing a Euclidean distance between the dimension reduction feature vectors in the low-dimensional space, finding a neighborhood block which is most matched with the low-resolution image block to be reconstructed, using a similarity and a scale factor between the feature vectors to construct an accurate embedded weight coefficient so that a searching speed and matching precision can be increased; then constructing the similarity and the scale factor between the feature vectors, calculating the accurate weight coefficient and acquiring more high frequency information from a training database; finally, according to the weight coefficient and the neighborhood block, estimating the high-resolution image block with high precision, reconstructing the image which has the high similarity with a real object, which is good for later-stage real object identification processing.

Description

Based on the image super-resolution rebuilding method of subspace projection and neighborhood embedding
Technical field
The present invention relates to the method for image processing, relate in particular to a kind of image super-resolution rebuilding method based on subspace projection and neighborhood embedding.
Background technology
Image super-resolution (Super Resolution, SR) reconstruction technique refers to and utilize one or more low resolution (the Low Resolution capturing in the different observation angles of Same Scene, different observation time or different sensors situation, LR) mutual information in image, adopt the method for Digital Image Processing to reconstruct a panel height resolution (High Resolution, HR) image.Under the condition that this reconstruction technique can limit at hardware devices such as charge-coupled image sensors (Charge Coupled Device, CCD), estimate the high-frequency information of its loss by one or more low-resolution image.In fields such as remote sensing satellite, military surveillance, medical imaging, security monitoring and traffic administrations, image super-resolution rebuilding technology not only has important researching value, is also with a wide range of applications.
The essence of image super-resolution rebuilding is how to solve the accurate estimation problem of losing high-frequency information, this is a classical one-to-many ill-conditioning problem, theory to this problem and applied research are still in the exploratory stage, and the aspects such as the quality of the time efficiency of image super-resolution rebuilding method and super-resolution rebuilding image all need further to be improved.Existing image super-resolution rebuilding method is mainly divided into interpolation, reconstruct and study three major types.Method of interpolation is a kind of method of utilizing known neighbour's point value to estimate interpolation point value, and these class methods are difficult to estimate the high-frequency information of loss.Reconstruction Method is to utilize the mutual information between several low-resolution images comprehensively to estimate high-frequency information, and its reconstruction effect is better than method of interpolation.But, Reconstruction Method needs the low-resolution image after several accuracy registrations, and the precision of current most image registration algorithms is not very high, can affect the quality of super-resolution rebuilding image.Learning method is the corresponding relation between low-resolution image piece and high-definition picture piece in first learning training storehouse, and estimates the high-definition picture piece corresponding with low-resolution image piece to be reconstructed according to this relation.Learning method can be realized single image super-resolution rebuilding, does not need image to carry out registration, and can from training storehouse, get more high-frequency information, thereby obtain better rebuilding effect.Therefore, learning method is the study hotspot of current image super-resolution rebuilding technology.
The image super-resolution rebuilding embedding based on neighborhood is an important branch in learning method, obtains in recent years good achievement in research.(the document 1:H.Chang such as Chang, D.-Y.Yeung, and Y.Xiong, " Super-resolution through neighbor embedding ", IEEE Conference on Computer Vision Pattern Recognition, pp.275-282, 2004) first propose to utilize the thought of neighborhood embedding to realize image super-resolution rebuilding, this method is to have under the prerequisite of similar local geometric features hypothesis at low-resolution image piece and corresponding high-definition picture piece, for each low-resolution image piece to be reconstructed, utilize local linear to embed (Locally Linear Embedding, LLE) algorithm calculates the embedding weight coefficient of several training low-resolution image pieces similar to low-resolution image piece to be reconstructed, each similar training low-resolution image piece is as a neighborhood piece, and then embed weight coefficients and corresponding neighborhood piece comprehensively estimates the high-definition picture piece after reconstruction with these.But, the method is not considered image block type and the impact of neighborhood piece number on super-resolution rebuilding effect.Subsequently, (the document 2:T.-M.Chan such as Chan, J.Zhang, J.Pu, and H.Huang, " Neighbor embedding based super-resolution algorithm through edge detection and feature selection ", Pattern Recognition Letters, vol.30, pp.494-502, 2009) propose a kind of neighborhood based on rim detection and Feature Selection and embed algorithm, this method can be determined the number that will choose neighborhood piece according to the type of image block (He Fei edge, edge), but, in the time there is serious blurring effect in input low-resolution image to be reconstructed, this algorithm can be difficult to judge the type of image block, make super-resolution rebuilding poor effect.(the document 3:K.Zhang such as Zhang, X.Gao, X.Li, and D.Tao, " Partially supervised neighbor embedding for example-based image super-resolution ", IEEE Journal of Selected Topics in Signal Processing, vol.5, no.2, pp.230-239, 2011) utilize class label information to construct a semi-supervised distance function, can from training storehouse, select the neighborhood piece that low-resolution image piece more and to be reconstructed mates according to this semi-supervised distance function, the super-resolution rebuilding effect of the method depends on the accuracy of identification of sorter to a certain extent.Algorithm described in document 1~3 is all on original high-dimensional feature space, to find several neighborhood pieces, because the dimension of original feature vector is very large, only rely on certain distance function to find neighborhood piece, the similarity between the neighborhood piece searching out and low-resolution image piece to be reconstructed is not high.Feature Dimension Reduction is a kind of feasible method addressing the above problem, based on this thinking, (the document 4:X.Gao such as Gao, K.Zhang, D.Tao, and X.Li, " Joint learning for single image super-resolution via coupled constraint ", IEEE Transactions on Image Processing, 2011) proposition is low by training by combination learning method, high-resolution features vector projects on a low-dimensional proper subspace simultaneously, and then on this low-dimensional proper subspace, find out the neighborhood piece that several mate with low-resolution image piece to be reconstructed.Although document 4 has been used the method for Feature Dimension Reduction, but when original low, high-resolution features is vectorial while projecting on unified low-dimensional proper subspace simultaneously, due to large than original low resolution proper vector of the dimension of original high resolution proper vector, inevitably can damage the useful information in original high-resolution features vector, also can produce harmful effect to the neighborhood piece of finding coupling.In addition, the algorithm of document 1~4 is all to adopt local linear to embed algorithm in the time calculating embedding weight coefficient, the embedding weight coefficient calculating is likely negative value, this can affect neighbour and train high-definition picture piece that the degree of high-frequency information can be provided, and the computation complexity of this algorithm is high, and arithmetic speed is slow.
Summary of the invention
The object of the present invention is to provide a kind of image super-resolution rebuilding method based on subspace projection and neighborhood embedding, the method realizes more effective Feature Dimension Reduction, estimate more accurately and rebuild rear high-definition picture piece, similarity between reconstruction image and real-world object is higher, super-resolution rebuilding better effects if.
The present invention solves its technical matters, and the technical scheme adopting is:
A, training:
Using high-definition picture identical L width resolution, that size is identical as training high-definition picture
Figure BDA0000139605960000031
l=1,2 ..., L, L=3~80; To every width training high-definition picture
Figure BDA0000139605960000032
carry out obtaining N after overlap partition 1individual size is the training high-definition picture piece of z*z, N 1=1000~7000, z=6,9,12,15, obtain altogether N=L*N 1individual training high-definition picture piece, extract after the standardization brightness of each training high-definition picture piece as a training high resolving power standardization brightness image block, i training high resolving power standardization brightness image block is converted to i by an order being listed as and trains high-resolution features vector
Figure BDA0000139605960000033
i=1,2 ..., N, each training high-resolution features vector
Figure BDA0000139605960000034
dimension be d 1=z 2, all training high-resolution features vectors
Figure BDA0000139605960000035
training high-resolution features matrix of (1≤i≤N) composition
Figure BDA0000139605960000036
To l (1≤l≤L) width training high-definition picture
Figure BDA0000139605960000041
do a times of down-sampling processing, a=2,3,4,5, obtain corresponding l width training low-resolution image
Figure BDA0000139605960000042
again to every width training low-resolution image
Figure BDA0000139605960000043
do after a times of up-sampling processed and obtain corresponding training interpolation image
Figure BDA0000139605960000044
extract every width training interpolation image
Figure BDA0000139605960000045
the vertical and second order level of single order level, single order, second order VG (vertical gradient) feature, this four width Gradient Features image is carried out to overlap partition, every width training interpolation image simultaneously
Figure BDA0000139605960000046
after overlap partition, obtain 4*N 1individual size is the training low resolution Gradient Features image block of z*z, obtain altogether 4*N training low resolution Gradient Features image block, every four training low resolution Gradient Features image blocks, as a training low resolution Gradient Features image block group, are converted to i training low resolution proper vector by i training low resolution Gradient Features image block group by an order being listed as
Figure BDA0000139605960000047
each training low resolution proper vector dimension be d 2=4*z 2, all training low resolution proper vectors
Figure BDA0000139605960000049
training low resolution eigenmatrix of (1≤i≤N) composition X s = [ x 1 s , x 2 s , · · · , x N s ] ;
B, pre-service:
Input low-resolution image R to be reconstructed d, its resolution and all training low-resolution images
Figure BDA00001396059600000411
the resolution of (1≤l≤L) is identical, by low-resolution image R to be reconstructed dcarry out after a times of up-sampling processed obtaining interpolation image R to be reconstructed c, extract interpolation image R to be reconstructed cthe vertical and second order level of single order level, single order, second order VG (vertical gradient) feature, this four width Gradient Features image is carried out obtaining altogether 4*N after overlap partition simultaneously 2individual size is the low resolution Gradient Features image block to be reconstructed of z*z, N 2=1000~7000, every four low resolution Gradient Features image blocks to be reconstructed, as a low resolution Gradient Features image block group to be reconstructed, are converted to j low resolution proper vector to be reconstructed by j low resolution Gradient Features image block group to be reconstructed by an order being listed as j=1,2 ..., N 2, each low resolution proper vector to be reconstructed
Figure BDA00001396059600000413
dimension and all training low resolution proper vectors
Figure BDA00001396059600000414
the dimension of (1≤i≤N) is identical, is d 2=4*z 2, all low resolution proper vectors to be reconstructed
Figure BDA00001396059600000415
(1≤j≤N 2) a low resolution eigenmatrix to be reconstructed of composition X t = [ x 1 t , x 2 t , · · · , x N 2 t ] ;
C, super-resolution rebuilding:
C1, subcharacter matrix generate: to the low resolution proper vector each to be reconstructed obtaining in step B
Figure BDA0000139605960000051
(1≤j≤N 2), the training low resolution eigenmatrix obtaining in steps A successively in find out and j low resolution proper vector to be reconstructed
Figure BDA0000139605960000053
between n-1=100~300 the training low resolution proper vector of Euclidean distance minimum, by j low resolution proper vector to be reconstructed
Figure BDA0000139605960000054
as subcharacter vector x 1j, and by the n-1 searching out a training low resolution proper vector by with j low resolution proper vector to be reconstructed
Figure BDA0000139605960000055
between Euclidean distance from small to large successively as subcharacter vector x 2j, x 3j..., x nj, form thus j sub-eigenmatrix
C2, dimensionality reduction eigenmatrix generate: adopt I and II subspace projection method, by j sub-eigenmatrix carry out Feature Dimension Reduction, each subcharacter vector x i ' j(1≤i '≤n) project on a low n-dimensional subspace n, obtain altogether n dimensionality reduction proper vector
Figure BDA0000139605960000058
each dimensionality reduction proper vector
Figure BDA0000139605960000059
dimension be m < d 2; Each subcharacter vector x i ' jwith corresponding dimensionality reduction proper vector
Figure BDA00001396059600000510
form mapping relations one by one, after I and II subspace projection, dimension is d 2subcharacter vector x i ' jconverting dimension to is the dimensionality reduction proper vector of m
Figure BDA00001396059600000511
realization character dimensionality reduction, all dimensionality reduction proper vectors (1≤i '≤n) j dimensionality reduction eigenmatrix of composition X j m = { x i &prime; j m } i &prime; = 1 n ;
C3, weight coefficient calculate: the dimensionality reduction proper vector set obtaining at step C2 in find out and dimensionality reduction proper vector
Figure BDA00001396059600000515
between k dimensionality reduction proper vector of Euclidean distance minimum, k=5~10, according to step C1 neutron proper vector x i ' jwith dimensionality reduction proper vector in step C2
Figure BDA00001396059600000516
between mapping relations one by one, and the call number of the k searching out a dimensionality reduction proper vector, the training that obtains in steps A is successively low, high-resolution features matrix X sand Y smiddlely find out that corresponding k is low to training, high-resolution features is vectorial, and, high-resolution features low to training by this k is vectorial forms respectively that j neighbour is low, high-resolution features matrix
Figure BDA00001396059600000517
with
Figure BDA00001396059600000518
according to j neighbour's low resolution eigenmatrix
Figure BDA00001396059600000519
with j low resolution proper vector to be reconstructed
Figure BDA00001396059600000520
calculate j similarity group
Figure BDA00001396059600000521
with j scale factor group
Figure BDA00001396059600000522
and associating
Figure BDA00001396059600000523
with
Figure BDA00001396059600000524
calculate j normalization and embed weight coefficient group
Figure BDA0000139605960000061
C4, reconstruction high-definition picture piece: embed weight coefficient group according to j normalization
Figure BDA0000139605960000062
with j neighbour's high-resolution features matrix
Figure BDA0000139605960000063
linear weighted function form, estimate j rebuild after high-resolution features vector
Figure BDA0000139605960000064
again with the brightness average of j low-resolution image piece to be reconstructed be added, and be converted to image block form, can show that j is rebuild rear high-definition picture piece;
Repeat above step C1~step C4, obtain altogether N 2individual size is high-definition picture piece after the reconstruction of z*z, obtains the preliminary image H of super-resolution rebuilding after lap splice 0;
D, aftertreatment:
Adopt the preliminary image H of iteration back-projection algorithms to super-resolution rebuilding 0carry out Q iterative computation processing, Q=10~30, obtain the final image H of super-resolution rebuilding *.
Compared with prior art, the invention has the beneficial effects as follows:
One, the present invention adopts I and II subspace projection method by subcharacter matrix projection to lower dimensional space, obtain corresponding dimensionality reduction eigenmatrix, reduce to greatest extent the dimension of original low resolution proper vector, and more effectively represent the feature of low-resolution image piece.Therefore, the Time & Space Complexity of the inventive method reduces, and arithmetic speed improves.
Two, the directly Euclidean distance between dimensionality reduction proper vector relatively at lower dimensional space but not on original higher dimensional space of the present invention, due to the overall situation that has comprised raw data in dimensionality reduction eigenmatrix and local structural information, can find out the neighborhood piece mating the most with low-resolution image piece to be reconstructed more efficiently according to dimensionality reduction proper vector.Visible, the present invention has higher matching precision to finding neighborhood piece.
Three, the present invention utilizes the similarity between weight coefficient and two proper vectors to have approximate proportional relation, construct similarity between low resolution proper vector to be reconstructed and neighbour's low resolution proper vector and corresponding scale factor, calculate more accurate weight coefficient, be conducive to estimate high-definition picture piece.Visible, the present invention has higher estimated accuracy and stronger adaptability.
In a word, the inventive method adopts I and II subspace projection method that original high dimensional data is projected on a lower dimensional space, make the character representation of low-resolution image piece become more succinct, efficient, can search out the more neighborhood piece of coupling, and construct more accurate weight coefficient by the similarity between proper vector and scale factor, rebuild rear high-definition picture piece thereby estimate more accurately, super-resolution rebuilding better effects if, the similarity of rebuilding image and real-world object is higher, is conducive to later stage real-world object identifying processing.
In above-mentioned step C2, adopt I and II subspace projection method, by j sub-eigenmatrix
Figure BDA0000139605960000071
carry out obtaining j dimensionality reduction eigenmatrix after Feature Dimension Reduction specific practice be:
One-level subspace projection:
Calculate j sub-eigenmatrix j corresponding radial basis kernel matrix K j = { K i &prime; j &prime; j } i &prime; , j &prime; = 1 n = { exp ( - | | x i &prime; j - x j &prime; j | | 2 / &beta; j ) } i &prime; , j &prime; = 1 n , Wherein the parameter of j radial basis kernel function is
Figure BDA0000139605960000075
then to j radial basis kernel matrix K jcarry out obtaining j centralization kernel matrix after centralization processing
Figure BDA0000139605960000076
g=I-1/n in formula, I is a unit matrix that size is n*n, 1 is that a size is n*n, and all values is 1 matrix; Solve j centralization kernel matrix descending Eigenvalues Decomposition equation
Figure BDA0000139605960000078
i '=1,2,3 ..., n, j=1,2,3 ..., N 2, λ in formula i ' jfor the individual eigenwert of i ' by after descending sort, α i ' jit is the individual eigenvalue λ of i ' i ' jthe corresponding individual proper vector of i '; Again to the individual proper vector α of i ' i ' jcarry out obtaining orthogonalization proper vector after orthogonalization process
Figure BDA0000139605960000079
choose front r eigenvalue of maximum λ i ' jcorresponding orthogonalization proper vector r=50~n, the r selecting an orthogonalization proper vector
Figure BDA00001396059600000711
form j one-level subspace projection matrix
Figure BDA00001396059600000712
front r described eigenvalue of maximum
Figure BDA00001396059600000713
sum accounts for all eigenwerts
Figure BDA00001396059600000714
the more than 99% of sum; By formula draw j one-level mapping matrix
Figure BDA00001396059600000716
t is matrix transpose computing, all one-level mapping vectors
Figure BDA00001396059600000717
(1≤i '≤dimension n) is r;
Secondary subspace projection:
Calculate j one-level mapping matrix j corresponding core distance function matrix D j r = { D i &prime; j &prime; j r } i &prime; , j &prime; = 1 n = { 2 - 2 * exp ( - | | x i &prime; j r - x j &prime; j r | | 2 / &gamma; j ) } i &prime; , j &prime; = 1 n , Wherein the parameter of j core distance function is then construct j adjacency matrix
Figure BDA0000139605960000083
as i ' ∈ Λ j ' jor j ' ∈ Λ i ' jtime, otherwise, S i ' j ' j=0, wherein Λ j ' jand Λ i ' jbe expressed as
Figure BDA0000139605960000085
neighborhood tally set, the element number of neighborhood tally set is b=5~10; Calculate j adjacency matrix S jj corresponding Laplacian Matrix F j=diag (S j* 1 n)-S j, wherein 1 nbeing expressed as a size is n*1, and all values is 1 column vector, and diag () is expressed as diagonalization of matrix computing; Then calculate respectively j symmetric matrix U j = X j r * F j * F j * ( X j r ) T With j diagonal matrix V j = X j r * { diag ( S j * 1 n ) } * ( X j r ) T , J symmetric matrix U jwith j diagonal matrix V jare all size matrixes for r*r, combine and solve U jand V jbroad sense ascending order Eigenvalues Decomposition equation
Figure BDA0000139605960000088
in formula for the individual eigenwert of i ' after arranging by ascending order, p i ' jit is the individual eigenwert of i '
Figure BDA00001396059600000810
the corresponding individual proper vector of i '; Choose a front m minimal eigenvalue
Figure BDA00001396059600000811
characteristic of correspondence vector p i ' j, m=10~r, the m a selecting proper vector
Figure BDA00001396059600000812
form j secondary subspace projection matrix P j=[p 1j, p 2j..., p mj], by formula
Figure BDA00001396059600000813
can draw j dimensionality reduction eigenmatrix
Figure BDA00001396059600000814
all dimensionality reduction proper vectors
Figure BDA00001396059600000815
(1≤i '≤dimension n) is m.
Adopt after above one-level subspace projection method, the global structure information that has comprised original high dimensional data in one-level mapping matrix, has retained more partial structurtes information after above secondary subspace projection in the dimensionality reduction eigenmatrix obtaining.Adopt I and II subspace projection method, dimensionality reduction eigenmatrix has not only retained the global structure information of raw data but also has retained partial structurtes information, realization character dimensionality reduction, can more effectively express the feature of low-resolution image piece, thereby makes super-resolution rebuilding become succinct and efficient.
In above-mentioned step C3, combine j similarity group
Figure BDA00001396059600000816
with j scale factor group calculate j normalization and embed weight coefficient group
Figure BDA00001396059600000818
specific practice be:
Calculate respectively j neighbour's low resolution eigenmatrix that step C3 obtains
Figure BDA00001396059600000819
in k training low resolution proper vector (1≤i "≤k) with j low resolution proper vector to be reconstructed
Figure BDA0000139605960000092
between similarity
Figure BDA0000139605960000093
with corresponding scale factor
Figure BDA0000139605960000094
wherein
Figure BDA0000139605960000095
with
Figure BDA0000139605960000096
be expressed as training low resolution proper vector
Figure BDA0000139605960000097
with j low resolution proper vector to be reconstructed in u brightness value, ε is 1*10 -4~1*10 -2positive number; The k calculating is to similarity
Figure BDA0000139605960000099
with scale factor c i " jform respectively j similarity group
Figure BDA00001396059600000910
with j scale factor group combine j similarity group
Figure BDA00001396059600000912
with j scale factor group
Figure BDA00001396059600000913
calculate j and embed weight coefficient group
Figure BDA00001396059600000914
and embed weight coefficient group to j
Figure BDA00001396059600000915
after doing normalized, obtain j normalization embedding weight coefficient group { w ^ i &prime; &prime; j } i &prime; &prime; = 1 k = { w i &prime; &prime; j / &Sigma; i &prime; &prime; j k w i &prime; &prime; j } i &prime; &prime; = 1 k .
Construct similarity and the scale factor between low resolution proper vector to be reconstructed and the training low resolution proper vector of mating the most by above method, with they calculate more accurate weight coefficient, meet the requirement that weight coefficient and similarity become to be similar to direct ratio, can from training storehouse, get more high-frequency information, be conducive to improve the estimated accuracy of high-definition picture piece after rebuilding, thereby obtain better super-resolution rebuilding effect.
Below in conjunction with the drawings and specific embodiments, the present invention is described in further detail.
Accompanying drawing explanation
Training image and the test pattern of Fig. 1 (a)~Fig. 1 (f) for using in the embodiment of the present invention.
Fig. 2 (A)~Fig. 2 (E) is for adopting algorithms of different to realize the simulation experiment result of image super-resolution rebuilding.
In Fig. 2 (A)~Fig. 2 (E), intermediate rectangular block diagram is the regional area of rebuilding on image, corner block diagram is that this regional area amplifies the design sketch of 2 times, and Fig. 2 (A) is for adopting the result figure of NESR algorithm (document 1); Fig. 2 (B) is for adopting the result figure of NeedFS algorithm (document 2); Fig. 2 (C) is for adopting the result figure of CSNE algorithm (document 3); Fig. 2 (D) is for adopting the result figure of JLNE algorithm (document 4); The result figure that Fig. 2 (E) is the inventive method.
Fig. 3 is that existing four kinds of methods and the inventive method are carried out image block root-mean-square error value E after super-resolution rebuilding to (a)~(f) the six width image in Fig. 1 respectively phistogram.
Fig. 4 is that existing four kinds of methods and the inventive method are carried out image block structural similarity value S after super-resolution rebuilding to (a)~(f) the six width image in Fig. 1 respectively phistogram.
Embodiment
Embodiment
Based on an image super-resolution rebuilding method for subspace projection and neighborhood embedding, comprise the following steps:
A, training:
Using high-definition picture identical L width resolution, that size is identical as training high-definition picture l=1,2 ..., L, L=3~80; To every width training high-definition picture
Figure BDA0000139605960000102
carry out obtaining N after overlap partition 1individual size is the training high-definition picture piece of z*z, N 1=1000~7000, z=6,9,12,15, obtain altogether N=L*N 1individual training high-definition picture piece, extract after the standardization brightness of each training high-definition picture piece as a training high resolving power standardization brightness image block, i training high resolving power standardization brightness image block is converted to i by an order being listed as and trains high-resolution features vector
Figure BDA0000139605960000103
i=1,2 ..., N, each training high-resolution features vector
Figure BDA0000139605960000104
dimension be d 1=z 2, all training high-resolution features vectors
Figure BDA0000139605960000105
(1≤i≤n) training high-resolution features matrix of composition
Figure BDA0000139605960000106
To l (1≤l≤L) width training high-definition picture
Figure BDA0000139605960000107
do a times of down-sampling processing, a=2,3,4,5, obtain corresponding l width training low-resolution image again to every width training low-resolution image do after a times of up-sampling processed and obtain corresponding training interpolation image
Figure BDA00001396059600001010
extract every width training interpolation image
Figure BDA00001396059600001011
the vertical and second order level of single order level, single order, second order VG (vertical gradient) feature, this four width Gradient Features image is carried out to overlap partition, every width training interpolation image simultaneously
Figure BDA00001396059600001012
after overlap partition, obtain 4*N 1individual size is the training low resolution Gradient Features image block of z*z, obtain altogether 4*N training low resolution Gradient Features image block, every four training low resolution Gradient Features image blocks, as a training low resolution Gradient Features image block group, are converted to i training low resolution proper vector by i training low resolution Gradient Features image block group by an order being listed as
Figure BDA00001396059600001013
each training low resolution proper vector dimension be d 2=4*z 2, all training low resolution proper vectors
Figure BDA0000139605960000111
training low resolution eigenmatrix of (1≤i≤N) composition X s = [ x 1 s , x 2 s , &CenterDot; &CenterDot; &CenterDot; , x N s ] ;
B, pre-service:
Input low-resolution image R to be reconstructed d, its resolution and all training low-resolution images
Figure BDA0000139605960000113
the resolution of (1≤l≤L) is identical, by low-resolution image R to be reconstructed dcarry out after a times of up-sampling processed obtaining interpolation image R to be reconstructed c, extract interpolation image R to be reconstructed cthe vertical and second order level of single order level, single order, second order VG (vertical gradient) feature, this four width Gradient Features image is carried out obtaining altogether 4*N after overlap partition simultaneously 2individual size is the low resolution Gradient Features image block to be reconstructed of z*z, N 2=1000~7000, every four low resolution Gradient Features image blocks to be reconstructed, as a low resolution Gradient Features image block group to be reconstructed, are converted to j low resolution proper vector to be reconstructed by j low resolution Gradient Features image block group to be reconstructed by an order being listed as
Figure BDA0000139605960000114
j=1,2 ..., N 2, each low resolution proper vector to be reconstructed
Figure BDA0000139605960000115
dimension and all training low resolution proper vectors
Figure BDA0000139605960000116
the dimension of (1≤i≤N) is identical, is d 2=4*z 2, all low resolution proper vectors to be reconstructed
Figure BDA0000139605960000117
(1≤j≤N 2) a low resolution eigenmatrix to be reconstructed of composition X t = [ x 1 t , x 2 t , &CenterDot; &CenterDot; &CenterDot; , x N 2 t ] ;
C, super-resolution rebuilding:
C1, subcharacter matrix generate: to the low resolution proper vector each to be reconstructed obtaining in step B (1≤j≤N 2), the training low resolution eigenmatrix obtaining in steps A successively
Figure BDA00001396059600001110
in find out and j low resolution proper vector to be reconstructed
Figure BDA00001396059600001111
between n-1=100~300 the training low resolution proper vector of Euclidean distance minimum, by j low resolution proper vector to be reconstructed
Figure BDA00001396059600001112
as subcharacter vector x 1j, and by the n-1 searching out a training low resolution proper vector by with j low resolution proper vector to be reconstructed
Figure BDA00001396059600001113
between Euclidean distance from small to large successively as subcharacter vector x 2j, x 3j..., x nj, form thus j sub-eigenmatrix
C2, dimensionality reduction eigenmatrix generate: adopt I and II subspace projection method, by j sub-eigenmatrix
Figure BDA0000139605960000121
carry out Feature Dimension Reduction, each subcharacter vector x i ' j(1≤i '≤n) project on a low n-dimensional subspace n, obtain altogether n dimensionality reduction proper vector
Figure BDA0000139605960000122
each dimensionality reduction proper vector
Figure BDA0000139605960000123
dimension be m < d 2; Each subcharacter vector x i ' jwith corresponding dimensionality reduction proper vector
Figure BDA0000139605960000124
form mapping relations one by one, after I and II subspace projection, dimension is d 2subcharacter vector x i ' jconverting dimension to is the dimensionality reduction proper vector of m
Figure BDA0000139605960000125
realization character dimensionality reduction, all dimensionality reduction proper vectors
Figure BDA0000139605960000126
(1≤i '≤n) j dimensionality reduction eigenmatrix of composition X j m = { x i &prime; j m } i &prime; = 1 n ;
More than adopt I and II subspace projection method, by j sub-eigenmatrix
Figure BDA0000139605960000128
carry out obtaining j dimensionality reduction eigenmatrix after Feature Dimension Reduction specific practice be:
One-level subspace projection:
Calculate j sub-eigenmatrix
Figure BDA00001396059600001210
j corresponding radial basis kernel matrix K j = { K i &prime; j &prime; j } i &prime; , j &prime; = 1 n = { exp ( - | | x i &prime; j - x j &prime; j | | 2 / &beta; j ) } i &prime; , j &prime; = 1 n , Wherein the parameter of j radial basis kernel function is then to j radial basis kernel matrix K jcarry out obtaining j centralization kernel matrix after centralization processing
Figure BDA00001396059600001213
g=I-1/n in formula, I is a unit matrix that size is n*n, 1 is that a size is n*n, and all values is 1 matrix; Solve j centralization kernel matrix
Figure BDA00001396059600001214
descending Eigenvalues Decomposition equation
Figure BDA00001396059600001215
i '=1,2,3 ..., n, j=1,2,3 ..., N 2, λ in formula i ' jfor the individual eigenwert of i ' by after descending sort, α i ' jit is the individual eigenvalue λ of i ' i ' jthe corresponding individual proper vector of i '; Again to the individual proper vector α of i ' i ' jcarry out obtaining orthogonalization proper vector after orthogonalization process
Figure BDA00001396059600001216
choose front r eigenvalue of maximum λ i ' jcorresponding orthogonalization proper vector
Figure BDA00001396059600001217
r=50~n, the r selecting an orthogonalization proper vector
Figure BDA00001396059600001218
form j one-level subspace projection matrix front r described eigenvalue of maximum sum accounts for all eigenwerts
Figure BDA00001396059600001221
the more than 99% of sum; By formula draw j one-level mapping matrix
Figure BDA00001396059600001223
t is matrix transpose computing, all one-level mapping vectors
Figure BDA00001396059600001224
(1≤i '≤dimension n) is r;
Secondary subspace projection:
Calculate j one-level mapping matrix j corresponding core distance function matrix D j r = { D i &prime; j &prime; j r } i &prime; , j &prime; = 1 n = { 2 - 2 * exp ( - | | x i &prime; j r - x j &prime; j r | | 2 / &gamma; j ) } i &prime; , j &prime; = 1 n , Wherein the parameter of j core distance function is
Figure BDA0000139605960000133
then construct j adjacency matrix
Figure BDA0000139605960000134
as i ' ∈ Λ j ' jor j ' ∈ Λ i ' jtime,
Figure BDA0000139605960000135
otherwise, S i ' j ' j=0, wherein Λ j ' jand Λ i ' jbe expressed as
Figure BDA0000139605960000136
neighborhood tally set, the element number of neighborhood tally set is b=5~10; Calculate j adjacency matrix S jj corresponding Laplacian Matrix F j=diag (S j* 1 n)-S j, wherein 1 nbeing expressed as a size is n*1, and all values is 1 column vector, and diag () is expressed as diagonalization of matrix computing; Then calculate respectively j symmetric matrix U j = X j r * F j * ( X j r ) T With j diagonal matrix V j = X j r * { diag ( S j * 1 n ) } * ( X j r ) T , J symmetric matrix U jwith j diagonal matrix V jare all size matrixes for r*r, combine and solve U jand V jbroad sense ascending order Eigenvalues Decomposition equation
Figure BDA0000139605960000139
in formula
Figure BDA00001396059600001310
for the individual eigenwert of i ' after arranging by ascending order, p i ' jit is the individual eigenwert of i '
Figure BDA00001396059600001311
the corresponding individual proper vector of i '; Choose a front m minimal eigenvalue
Figure BDA00001396059600001312
characteristic of correspondence vector p i ' j, m=10~r, the m a selecting proper vector
Figure BDA00001396059600001313
form j secondary subspace projection matrix P j=[p 1j, p 2j..., p mj], by formula
Figure BDA00001396059600001314
can draw j dimensionality reduction eigenmatrix
Figure BDA00001396059600001315
all dimensionality reduction proper vectors
Figure BDA00001396059600001316
(1≤i '≤dimension n) is m;
C3, weight coefficient calculate: the dimensionality reduction proper vector set obtaining at step C2 in find out and dimensionality reduction proper vector
Figure BDA00001396059600001318
between k dimensionality reduction proper vector of Euclidean distance minimum, k=5~10, according to step C1 neutron proper vector x i ' jwith dimensionality reduction proper vector in step C2 between mapping relations one by one, and the call number of the k searching out a dimensionality reduction proper vector, the training that obtains in steps A is successively low, high-resolution features matrix X sand Y smiddlely find out that corresponding k is low to training, high-resolution features is vectorial, and, high-resolution features low to training by this k is vectorial forms respectively that j neighbour is low, high-resolution features matrix
Figure BDA0000139605960000141
with
Figure BDA0000139605960000142
according to j neighbour's low resolution eigenmatrix
Figure BDA0000139605960000143
with j low resolution proper vector to be reconstructed calculate j similarity group
Figure BDA0000139605960000145
with j scale factor group
Figure BDA0000139605960000146
and associating
Figure BDA0000139605960000147
with
Figure BDA0000139605960000148
calculate j normalization and embed weight coefficient group
Figure BDA0000139605960000149
More than combine j similarity group
Figure BDA00001396059600001410
with j scale factor group
Figure BDA00001396059600001411
calculate j normalization and embed weight coefficient group specific practice be:
Calculate respectively j neighbour's low resolution eigenmatrix that step C3 obtains
Figure BDA00001396059600001413
in k training low resolution proper vector
Figure BDA00001396059600001414
(1≤i "≤k and j low resolution proper vector to be reconstructed
Figure BDA00001396059600001415
between similarity
Figure BDA00001396059600001416
with corresponding scale factor wherein
Figure BDA00001396059600001418
with be expressed as training low resolution proper vector
Figure BDA00001396059600001420
with j low resolution proper vector to be reconstructed
Figure BDA00001396059600001421
in u brightness value, ε is 1*10 -4~1*10 -2positive number; The k calculating is to similarity with scale factor c i " jform respectively j similarity group
Figure BDA00001396059600001423
with j scale factor group
Figure BDA00001396059600001424
combine j similarity group
Figure BDA00001396059600001425
with j scale factor group calculate j and embed weight coefficient group
Figure BDA00001396059600001427
and embed weight coefficient group to j
Figure BDA00001396059600001428
after doing normalized, obtain j normalization embedding weight coefficient group { w ^ i &prime; &prime; j } i &prime; &prime; = 1 k = { w i &prime; &prime; j / &Sigma; i &prime; &prime; = 1 k w i &prime; &prime; j } i &prime; &prime; = 1 k ;
C4, reconstruction high-definition picture piece: embed weight coefficient group according to j normalization
Figure BDA00001396059600001430
with j neighbour's high-resolution features matrix
Figure BDA00001396059600001431
linear weighted function form, estimate j rebuild after high-resolution features vector
Figure BDA00001396059600001432
again with the brightness average of j low-resolution image piece to be reconstructed be added, and be converted to image block form, can show that j is rebuild rear high-definition picture piece;
Repeat above step C1~step C4, obtain altogether N 2individual size is high-definition picture piece after the reconstruction of z*z, obtains the preliminary image H of super-resolution rebuilding after lap splice 0;
D, aftertreatment:
Adopt the preliminary image H of iteration back-projection algorithms to super-resolution rebuilding 0carry out Q iterative computation processing, Q=10~30, obtain the final image H* of super-resolution rebuilding.
Emulation experiment:
The condition of emulation experiment and design parameter are:
In Fig. 1, choose the big or small natural image for 384*510 of L=5 width as training high-definition picture, a remaining width is as test reference image, and emulation experiment has been done in rotation altogether 6 times successively.
The sampling multiple a=3 of training high-definition picture, tile size is z*z=9*9, obtains altogether N=L*N after doubling of the image piecemeal 1=5*5440=27200 image block, the dimension of training high-resolution features vector is d 1=z 2=9 2=81, the dimension of training low resolution proper vector is d 2=4*z 2=4*9 2=324.
Image using test reference image after 3 times of down-samplings is as low-resolution image to be reconstructed, and size is 128 × 170.The interpolation image that low-resolution image to be reconstructed is carried out to obtain after 3 times of up-samplings carries out obtaining N after overlap partition 2=5440 low-resolution image pieces to be reconstructed, the dimension of low resolution proper vector to be reconstructed is d 2=324.All subcharacter matrix X jvectorial number in (1≤j≤5440) is n=101, it is r=100 that one-level is shone upon vectorial dimension, in secondary subspace projection process, the element number of neighborhood tally set is b=5, the dimension of dimensionality reduction proper vector is m=22, the number of the dimensionality reduction proper vector of finding is k=9, ε=1*10 -3, the iterations that aftertreatment is used is Q=20.
Adopt existing four kinds of methods simultaneously, be respectively NESR algorithm (document 1), NeedFS algorithm (document 2), CSNE algorithm (document 3) and JLNE algorithm (document 4), and the inventive method is similarly being carried out super-resolution emulation reconstruction under emulation experiment condition to (a)~(f) the six width image in Fig. 1.
Subjective vision Contrast on effect result as shown in Figure 2, Fig. 2 only provide (b) in Fig. 1, (c) and (d) three width images carry out the result of emulation experiment.Fig. 2 (A)~Fig. 2 (D) is respectively the simulation experiment result of document 1~4, the simulation experiment result that Fig. 2 (E) is the inventive method.In Fig. 2 (A)~Fig. 2 (E), intermediate rectangular block diagram is the regional area of rebuilding on image, and corner block diagram is that this regional area amplifies the design sketch of 2 times.Relatively partial enlarged drawing can be found out: the reconstruction image of NESR, NeedFS and CSNE algorithm there will be fuzzy and texture aliasing, in the reconstruction image of JLNE algorithm, texture edge is level and smooth not, detailed information in the reconstruction image of the inventive method is more clear, and edge is more level and smooth.Visible, the inventive method is better than existing four kinds of methods in subjective vision effect.
In order more accurately the whole bag of tricks to be carried out to objective evaluation, below with image block root-mean-square error value E pwith image block structural similarity value S pas the objective indicator of evaluating the whole bag of tricks quality, wherein E pand S pcomputing formula as follows:
E p = 1 N 2 &Sigma; j = 1 N 2 1 d 1 &Sigma; v = 1 d 1 ( y jv - y jv t ) 2
S p = 1 N 2 &Sigma; j = 1 N 2 ( 2 &mu; 1 j &mu; 2 j + &epsiv; 1 ) ( 2 &sigma; 12 j + &epsiv; 2 ) ( &mu; 1 j 2 + &mu; 2 j 2 + &epsiv; 1 ) ( &sigma; 1 j 2 + &sigma; 2 j 2 + &epsiv; 2 )
In the same form, y jvfor v brightness value of j image block on original high resolution image,
Figure BDA0000139605960000163
for v brightness value of j image block on super-resolution rebuilding image, d 1for the number of the contained brightness value of each image block, identical with the dimension of training high-resolution features vector, be d 1=81, N 2for rebuilding rear high-definition picture piece number.
In two formulas, μ 1jand μ 2jbe respectively the brightness average of j image block on original high resolution image and super-resolution rebuilding image, σ 1jand σ 2jbe respectively the standard deviation of j image block on original high resolution image and super-resolution rebuilding image, σ 12jfor the standard covariance of j image block on j image block on original high resolution image and super-resolution rebuilding image, ε 1=6.5, ε 2=58.5.
Fig. 3 is that existing four kinds of methods and the inventive method are carried out image block root-mean-square error value E after super-resolution rebuilding to (a)~(f) the six width image in Fig. 1 respectively phistogram, as can be seen from this figure, 6 width of the inventive method are rebuild the image block root-mean-square error value E of images pall, lower than existing four kinds of methods, the difference minimum of reconstruction image and the original high resolution image of the inventive method is described, reconstruction effect is best.
Fig. 4 is that existing four kinds of methods and the inventive method are carried out image block structural similarity value S after super-resolution rebuilding to (a)~(f) the six width image in Fig. 1 respectively phistogram, as can be seen from this figure, 6 width of the inventive method are rebuild the image block structural similarity value S of images pall, higher than existing four kinds of methods, also illustrate that the inventive method can reconstruct more high-frequency information, closer to original high resolution image.
Above the simulation experiment result shows, subjective vision effect and objective evaluation index this aspect two on, the inventive method is all better than existing four kinds of methods, has feasibility and applicability in the application of image super-resolution rebuilding.

Claims (3)

1. the image super-resolution rebuilding method based on subspace projection and neighborhood embedding, comprises the following steps:
A, training:
Using high-definition picture identical L width resolution, that size is identical as training high-definition picture
Figure FDA0000446053710000011
l=1,2 ..., L, L=3~80; To every width training high-definition picture
Figure FDA0000446053710000012
carry out obtaining N after overlap partition 1individual size is the training high-definition picture piece of z*z, N 1=1000~7000, z=6,9,12,15, obtain altogether N=L*N 1individual training high-definition picture piece, extract the standardization brightness of each training high-definition picture piece as a training high resolving power standardization brightness image block, i training high resolving power standardization brightness image block is converted to i training high-resolution features vector by an order being listed as
Figure FDA0000446053710000013
i=1,2 ..., N, each training high-resolution features vector
Figure FDA0000446053710000014
dimension be d 1=z 2, all training high-resolution features vectors
Figure FDA0000446053710000015
form a training high-resolution features matrix
To l (1≤l≤L) width training high-definition picture
Figure FDA0000446053710000017
do a times of down-sampling processing, a=2,3,4,5, obtain corresponding l width training low-resolution image again to every width training low-resolution image
Figure FDA0000446053710000019
do after a times of up-sampling processed and obtain corresponding training interpolation image
Figure FDA00004460537100000110
extract every width training interpolation image the vertical and second order level of single order level, single order, second order VG (vertical gradient) feature, this four width Gradient Features image is carried out to overlap partition, every width training interpolation image simultaneously
Figure FDA00004460537100000112
after overlap partition, obtain 4*N 1individual size is the training low resolution Gradient Features image block of z*z, obtain altogether 4*N training low resolution Gradient Features image block, four dissimilar Gradient Features image blocks of each training low-resolution image piece, as a training low resolution Gradient Features image block group, are converted to i training low resolution proper vector by i training low resolution Gradient Features image block group by an order being listed as
Figure FDA00004460537100000113
each training low resolution proper vector
Figure FDA00004460537100000114
dimension be d 2=4*z 2, all training low resolution proper vectors
Figure FDA00004460537100000115
form a training low resolution eigenmatrix X s = [ x 1 s , x 2 s , . . . , x N s ] ;
B, pre-service:
Input low-resolution image R to be reconstructed d, its resolution and all training low-resolution images
Figure FDA0000446053710000021
resolution identical, by low-resolution image R to be reconstructed dcarry out after a times of up-sampling processed obtaining interpolation image R to be reconstructed c, extract interpolation image R to be reconstructed cthe vertical and second order level of single order level, single order, second order VG (vertical gradient) feature, this four width Gradient Features image is carried out obtaining altogether 4*N after overlap partition simultaneously 2individual size is the low resolution Gradient Features image block to be reconstructed of z*z, N 2=1000~7000, four dissimilar Gradient Features image blocks of each low-resolution image piece to be reconstructed, as a low resolution Gradient Features image block group to be reconstructed, are converted to j low resolution proper vector to be reconstructed by j low resolution Gradient Features image block group to be reconstructed by an order being listed as j=1,2 ..., N 2, each low resolution proper vector to be reconstructed
Figure FDA0000446053710000023
dimension and all training low resolution proper vectors
Figure FDA0000446053710000024
dimension identical, be d 2=4*z 2, all low resolution proper vectors to be reconstructed
Figure FDA0000446053710000025
form a low resolution eigenmatrix to be reconstructed X t = [ x 1 t , x 2 t , . . . , x N 2 t ] ;
C, super-resolution rebuilding:
C1, subcharacter matrix generate: to the low resolution proper vector each to be reconstructed obtaining in step B
Figure FDA0000446053710000027
the training low resolution eigenmatrix obtaining in steps A successively
Figure FDA0000446053710000028
in find out and j low resolution proper vector to be reconstructed
Figure FDA0000446053710000029
between n-1=100~300 the training low resolution proper vector of Euclidean distance minimum, by j low resolution proper vector to be reconstructed
Figure FDA00004460537100000210
as subcharacter vector x 1j, and by the n-1 searching out a training low resolution proper vector by with j low resolution proper vector to be reconstructed
Figure FDA00004460537100000211
between Euclidean distance from small to large successively as subcharacter vector x 2j, x 3j..., x nj, form thus j sub-eigenmatrix X j = { x i &prime; j } i &prime; = 1 n ;
C2, dimensionality reduction eigenmatrix generate: adopt I and II subspace projection method, by j sub-eigenmatrix
Figure FDA00004460537100000213
carry out Feature Dimension Reduction, each subcharacter vector x i ' j(1≤i '≤n) project on a low n-dimensional subspace n, obtain altogether n dimensionality reduction proper vector
Figure FDA00004460537100000214
each dimensionality reduction proper vector dimension be m<d 2; Each subcharacter vector x i ' jwith corresponding dimensionality reduction proper vector
Figure FDA00004460537100000216
form mapping relations one by one, after I and II subspace projection, dimension is d 2subcharacter vector x i ' jconverting dimension to is the dimensionality reduction proper vector of m
Figure FDA0000446053710000031
realization character dimensionality reduction, all dimensionality reduction proper vectors
Figure FDA0000446053710000032
form j dimensionality reduction eigenmatrix X j m = { x i &prime; j m } i &prime; = 1 n ;
C3, weight coefficient calculate: the dimensionality reduction proper vector set obtaining at step C2
Figure FDA0000446053710000034
in find out and dimensionality reduction proper vector
Figure FDA0000446053710000035
between k dimensionality reduction proper vector of Euclidean distance minimum, k=5~10, according to step C1 neutron proper vector x i ' jwith dimensionality reduction proper vector in step C2
Figure FDA0000446053710000036
between mapping relations one by one, and the call number of the k searching out a dimensionality reduction proper vector, the training that obtains in steps A is successively low, high-resolution features matrix X sand Y smiddlely find out that corresponding k is low to training, high-resolution features is vectorial, and, high-resolution features low to training by this k is vectorial forms respectively that j neighbour is low, high-resolution features matrix X j f = { x i &Prime; j f } i &Prime; = 1 k With Y j f = { y i &Prime; j f } i &Prime; = 1 k ; According to j neighbour's low resolution eigenmatrix X j f = { x i &Prime; j f } i &Prime; = 1 k With j low resolution proper vector to be reconstructed
Figure FDA00004460537100000310
calculate j similarity group
Figure FDA00004460537100000311
with j scale factor group
Figure FDA00004460537100000312
and associating with
Figure FDA00004460537100000314
calculate j normalization and embed weight coefficient group
Figure FDA00004460537100000321
C4, reconstruction high-definition picture piece: embed weight coefficient group according to j normalization
Figure FDA00004460537100000316
with j neighbour's high-resolution features matrix
Figure FDA00004460537100000317
linear weighted function form, estimate j rebuild after high-resolution features vector
Figure FDA00004460537100000318
again with the brightness average of j low-resolution image piece to be reconstructed
Figure FDA00004460537100000319
be added, and be converted to image block form, can show that j is rebuild rear high-definition picture piece;
Repeat above step C1~step C4, obtain altogether N 2individual size is high-definition picture piece after the reconstruction of z*z, obtains the preliminary image H of super-resolution rebuilding after lap splice 0;
D, aftertreatment:
Adopt the preliminary image H of iteration back-projection algorithms to super-resolution rebuilding 0carry out Q iterative computation processing, Q=10~30, obtain the final image H of super-resolution rebuilding *.
2. image super-resolution rebuilding method according to claim 1, is characterized in that, adopts I and II subspace projection method in described step C2, by j sub-eigenmatrix carry out obtaining j dimensionality reduction eigenmatrix after Feature Dimension Reduction
Figure FDA0000446053710000041
specific practice be:
One-level subspace projection:
Calculate j sub-eigenmatrix
Figure FDA0000446053710000042
j corresponding radial basis kernel matrix K j = { K i &prime; j &prime; j } i &prime; , j &prime; = 1 n = { exp ( - | | x i &prime; j - x j &prime; j | | 2 / &beta; j ) } i &prime; , j &prime; = 1 n , Wherein the parameter of j radial basis kernel function is &beta; j = ( &Sigma; i &prime; , j &prime; = 1 n | | x i &prime; j - x j &prime; j | | / n 2 ) 2 ; Then to j radial basis kernel matrix K jcarry out obtaining j centralization kernel matrix after centralization processing
Figure FDA0000446053710000045
g=I-1/n in formula, I is a unit matrix that size is n*n, 1 is that a size is n*n, and all values is 1 matrix; Solve j centralization kernel matrix descending Eigenvalues Decomposition equation
Figure FDA0000446053710000047
i '=1,2,3 ..., n, j=1,2,3 ..., N 2, λ in formula i ' jfor the individual eigenwert of i ' by after descending sort, α i ' jit is the individual eigenvalue λ of i ' i ' jthe corresponding individual proper vector of i '; Again to the individual proper vector α of i ' i ' jcarry out obtaining orthogonalization proper vector after orthogonalization process
Figure FDA0000446053710000048
choose front r eigenvalue of maximum λ i ' jcorresponding orthogonalization proper vector
Figure FDA0000446053710000049
r=50~n, the r selecting an orthogonalization proper vector
Figure FDA00004460537100000410
form j one-level subspace projection matrix A j = [ &alpha; ^ 1 j , &alpha; ^ 2 j , . . . , &alpha; ^ rj ] , Front r described eigenvalue of maximum
Figure FDA00004460537100000412
sum accounts for all eigenwerts the more than 99% of sum; By formula
Figure FDA00004460537100000414
draw j one-level mapping matrix
Figure FDA00004460537100000415
t is matrix transpose computing, all one-level mapping vectors
Figure FDA00004460537100000416
dimension be r;
Secondary subspace projection:
Calculate j one-level mapping matrix
Figure FDA00004460537100000417
j corresponding core distance function matrix D j r = { D i &prime; j &prime; j r } i &prime; , j &prime; = 1 n = { 2 - 2 * exp ( - | | x i &prime; j r - x j &prime; j r | | 2 / &gamma; j ) } i &prime; , j &prime; = 1 n , Wherein the parameter of j core distance function is &gamma; j = ( &Sigma; i &prime; , j &prime; = 1 n | | x i &prime; j r - x j &prime; j r | | / n 2 ) 2 ; Then construct j adjacency matrix S j = { S i &prime; j &prime; j } i &prime; , j &prime; = 1 n , As i ' ∈ Λ j ' jor j ' ∈ Λ i ' jtime, S i &prime; j &prime; j = 1 - D i &prime; j &prime; j r ; Otherwise, S i ' j ' j=0, wherein Λ j ' jand Λ i ' jbe expressed as
Figure FDA00004460537100000422
neighborhood tally set, the element number of neighborhood tally set is b=5~10; Calculate j adjacency matrix S jj corresponding Laplacian Matrix F j=diag (S j* 1 n)-S j, wherein 1 nbeing expressed as a size is n*1, and all values is 1 column vector, and diag () is expressed as diagonalization of matrix computing; Then calculate respectively j symmetric matrix U j = X j r * F j * ( X j r ) T With j diagonal matrix V j = X j r * { diag ( S j * 1 n ) } * ( X j r ) T , J symmetric matrix U jwith j diagonal matrix V jare all size matrixes for r*r, combine and solve U jand V jbroad sense ascending order Eigenvalues Decomposition equation
Figure FDA0000446053710000053
in formula
Figure FDA0000446053710000054
for the individual eigenwert of i ' after arranging by ascending order, p i ' jit is the individual eigenwert of i '
Figure FDA0000446053710000055
the corresponding individual proper vector of i '; Choose a front m minimal eigenvalue
Figure FDA0000446053710000056
characteristic of correspondence vector p i ' j, m=10~r, the m a selecting proper vector
Figure FDA0000446053710000057
form j secondary subspace projection matrix P j=[p 1j, p 2j..., p mj], by formula can draw j dimensionality reduction eigenmatrix
Figure FDA0000446053710000059
all dimensionality reduction proper vectors
Figure FDA00004460537100000510
dimension be m.
3. image super-resolution rebuilding method according to claim 1, is characterized in that, combines j similarity group in described step C3
Figure FDA00004460537100000511
with j scale factor group
Figure FDA00004460537100000512
calculate j normalization and embed weight coefficient group
Figure FDA00004460537100000513
specific practice be:
Calculate respectively j neighbour's low resolution eigenmatrix that step C3 obtains
Figure FDA00004460537100000514
in k training low resolution proper vector
Figure FDA00004460537100000515
with j low resolution proper vector to be reconstructed
Figure FDA00004460537100000516
between similarity K i &Prime; j f = exp ( - | | x i &Prime; j f - x j t | | 2 / &beta; j ) With corresponding scale factor c i &Prime; j = ( &Sigma; u = 1 d 2 | x i &Prime; ju f - x ju t | + &epsiv; ) - 1 , Wherein
Figure FDA00004460537100000519
with
Figure FDA00004460537100000520
be expressed as training low resolution proper vector
Figure FDA00004460537100000521
with j low resolution proper vector to be reconstructed
Figure FDA00004460537100000522
in u brightness value, ε is 1*10 -4~1*10 -2positive number; The k calculating is to similarity
Figure FDA00004460537100000523
with scale factor c i ' ' jform respectively j similarity group with j scale factor group combine j similarity group
Figure FDA00004460537100000526
with j scale factor group calculate j and embed weight coefficient group
Figure FDA00004460537100000528
and embed weight coefficient group to j
Figure FDA0000446053710000061
after doing normalized, obtain j normalization embedding weight coefficient group { w ^ i &Prime; j } i &Prime; = 1 k = { w i &Prime; j / &Sigma; i &Prime; = 1 k w i &Prime; j } i &Prime; = 1 k .
CN201210049804.1A 2012-02-29 2012-02-29 Image super resolution (SR) reconstruction method based on subspace projection and neighborhood embedding Expired - Fee Related CN102629374B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210049804.1A CN102629374B (en) 2012-02-29 2012-02-29 Image super resolution (SR) reconstruction method based on subspace projection and neighborhood embedding

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210049804.1A CN102629374B (en) 2012-02-29 2012-02-29 Image super resolution (SR) reconstruction method based on subspace projection and neighborhood embedding

Publications (2)

Publication Number Publication Date
CN102629374A CN102629374A (en) 2012-08-08
CN102629374B true CN102629374B (en) 2014-05-21

Family

ID=46587632

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210049804.1A Expired - Fee Related CN102629374B (en) 2012-02-29 2012-02-29 Image super resolution (SR) reconstruction method based on subspace projection and neighborhood embedding

Country Status (1)

Country Link
CN (1) CN102629374B (en)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103236049B (en) * 2013-05-05 2015-10-28 西安电子科技大学 Based on the partial K space image reconstruction method of sequence similarity interpolation
CN104182931B (en) * 2013-05-21 2017-04-26 北京大学 Super resolution method and device
CN103780863B (en) * 2014-01-17 2017-08-25 Tcl集团股份有限公司 A kind of high-definition picture generation method and device
CN104484418B (en) * 2014-12-17 2017-10-31 中国科学技术大学 A kind of characteristic quantification method and system based on dual resolution design
CN105046664B (en) * 2015-07-13 2018-05-25 广东工业大学 A kind of image de-noising method based on adaptive EPLL algorithms
CN105069825B (en) * 2015-08-14 2018-06-12 厦门大学 Image super-resolution rebuilding method based on depth confidence network
CN106920214B (en) * 2016-07-01 2020-04-14 北京航空航天大学 Super-resolution reconstruction method for space target image
CN106228583A (en) * 2016-07-14 2016-12-14 北京大学 The device that a kind of abdomen images is rebuild
CN107066984A (en) * 2017-04-20 2017-08-18 四川大学 Algorithm for gait recognition based on subspace integrated study
CN108537264B (en) * 2018-03-30 2021-09-07 西安电子科技大学 Heterogeneous image matching method based on deep learning
CN108616590B (en) * 2018-04-26 2020-07-31 清华大学 Billion-scale network embedded iterative random projection algorithm and device
CN109636727B (en) * 2018-12-17 2022-11-15 辽宁工程技术大学 Super-resolution reconstruction image spatial resolution evaluation method
CN110363235B (en) * 2019-06-29 2021-08-06 苏州浪潮智能科技有限公司 High-resolution image matching method and system

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070019887A1 (en) * 2004-06-30 2007-01-25 Oscar Nestares Computing a higher resolution image from multiple lower resolution images using model-base, robust bayesian estimation
CN101976435A (en) * 2010-10-07 2011-02-16 西安电子科技大学 Combination learning super-resolution method based on dual constraint

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070019887A1 (en) * 2004-06-30 2007-01-25 Oscar Nestares Computing a higher resolution image from multiple lower resolution images using model-base, robust bayesian estimation
CN101976435A (en) * 2010-10-07 2011-02-16 西安电子科技大学 Combination learning super-resolution method based on dual constraint

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Hong Chang et al.Super-Resolution Through Neighbor Embedding.《Proceedings of the 2004 IEEE Computer Society Conference》.2004,第1卷第1-8页.
Super-Resolution Through Neighbor Embedding;Hong Chang et al;《Proceedings of the 2004 IEEE Computer Society Conference》;20040627;第1卷;第1-8页 *
李民等.非局部联合稀疏近似的超分辨率重建算法.《电子与信息学报》.2011,第33卷(第6期),第1407-1412页.
非局部联合稀疏近似的超分辨率重建算法;李民等;《电子与信息学报》;20110630;第33卷(第6期);第1407-1412页 *

Also Published As

Publication number Publication date
CN102629374A (en) 2012-08-08

Similar Documents

Publication Publication Date Title
CN102629374B (en) Image super resolution (SR) reconstruction method based on subspace projection and neighborhood embedding
CN108573276B (en) Change detection method based on high-resolution remote sensing image
CN101976435B (en) Combination learning super-resolution method based on dual constraint
Chen et al. Convolutional neural network based dem super resolution
CN103034982B (en) Image super-resolution rebuilding method based on variable focal length video sequence
CN110458192B (en) Hyperspectral remote sensing image classification method and system based on visual saliency
CN104036296B (en) A kind of expression of image and processing method and processing device
CN110400276B (en) Hyperspectral image denoising method and device
CN104268833A (en) New image fusion method based on shift invariance shearlet transformation
CN104077742B (en) Human face sketch synthetic method and system based on Gabor characteristic
He et al. Learning group-based sparse and low-rank representation for hyperspectral image classification
CN115496928A (en) Multi-modal image feature matching method based on multi-feature matching
CN109034213B (en) Hyperspectral image classification method and system based on correlation entropy principle
Yuan et al. ROBUST PCANet for hyperspectral image change detection
Li et al. An aerial image segmentation approach based on enhanced multi-scale convolutional neural network
Khurshid et al. A residual-dyad encoder discriminator network for remote sensing image matching
CN107392211A (en) The well-marked target detection method of the sparse cognition of view-based access control model
CN112668662B (en) Outdoor mountain forest environment target detection method based on improved YOLOv3 network
Zhao et al. Multiple endmembers based unmixing using archetypal analysis
CN105046286A (en) Supervision multi-view feature selection method based on automatic generation of view and unit with l1 and l2 norm minimization
CN103345739B (en) A kind of high-resolution remote sensing image building area index calculation method based on texture
Yue et al. Remote-sensing image super-resolution using classifier-based generative adversarial networks
Liu et al. Fuzzy clustering algorithm for integrating multiscale spatial context in image segmentation by hidden Markov random field models
Kasem et al. DRCS-SR: Deep robust compressed sensing for single image super-resolution
Li et al. Focus on local: transmission line defect detection via feature refinement

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20140521

Termination date: 20190228

CF01 Termination of patent right due to non-payment of annual fee