CN111833252A - Image super-resolution method based on SAE dictionary learning and neighborhood regression - Google Patents
Image super-resolution method based on SAE dictionary learning and neighborhood regression Download PDFInfo
- Publication number
- CN111833252A CN111833252A CN202010670836.8A CN202010670836A CN111833252A CN 111833252 A CN111833252 A CN 111833252A CN 202010670836 A CN202010670836 A CN 202010670836A CN 111833252 A CN111833252 A CN 111833252A
- Authority
- CN
- China
- Prior art keywords
- dictionary
- image
- sae
- dictionary learning
- input data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 58
- 239000011159 matrix material Substances 0.000 claims abstract description 27
- 238000012549 training Methods 0.000 claims abstract description 23
- 238000013507 mapping Methods 0.000 claims description 22
- 238000012360 testing method Methods 0.000 claims description 14
- 230000006870 function Effects 0.000 claims description 9
- 238000002360 preparation method Methods 0.000 claims description 8
- 230000004913 activation Effects 0.000 claims description 6
- 238000001914 filtration Methods 0.000 claims description 6
- 238000000513 principal component analysis Methods 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 6
- 238000005070 sampling Methods 0.000 claims description 5
- 238000000638 solvent extraction Methods 0.000 claims description 5
- 238000007781 pre-processing Methods 0.000 claims description 4
- 238000000354 decomposition reaction Methods 0.000 claims description 3
- 238000011478 gradient descent method Methods 0.000 claims description 3
- 210000002569 neuron Anatomy 0.000 claims description 3
- 230000001105 regulatory effect Effects 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 8
- 238000004590 computer program Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 4
- 238000002474 experimental method Methods 0.000 description 4
- 238000004088 simulation Methods 0.000 description 4
- 238000003860 storage Methods 0.000 description 3
- 101100365548 Caenorhabditis elegans set-14 gene Proteins 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- RZVAJINKPMORJF-UHFFFAOYSA-N Acetaminophen Chemical compound CC(=O)NC1=CC=C(O)C=C1 RZVAJINKPMORJF-UHFFFAOYSA-N 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4053—Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2135—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/28—Determining representative reference patterns, e.g. by averaging or distorting; Generating dictionaries
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
The invention relates to an image super-resolution method based on SAE dictionary learning and neighborhood regression, which comprises the steps of firstly preparing input data aiming at a dictionary learning model SAE, and constructing and training a dictionary; then solving a projection matrix by combining a neighborhood regression theory and a dictionary; and finally, reconstructing the image based on the projection matrix to obtain a high-resolution image. On one hand, the method improves the feature expression capability of the dictionary and reduces the dependency of the reconstruction result on the dictionary; on the other hand, a neighborhood regression theory is integrated, and the reconstruction speed is improved.
Description
Technical Field
The invention relates to the field of image super-resolution method design, in particular to an image super-resolution method based on SAE dictionary learning and neighborhood regression.
Background
In reality, due to the limitations of image acquisition equipment, scene changes, light sources and other factors, high-quality images cannot be obtained, and the requirements of practical application cannot be met when the resolution of the images is low. The Super-Resolution (SR) method utilizes an image signal processing technique to reconstruct a single or multiple low-Resolution (LR) images into High-Resolution (HR) images, and is characterized in that certain additional information is added during the reconstruction process to compensate for the detail information lost during the image degradation process. The SR reconstruction can break through the limitation of the intrinsic resolution of the imaging device to realize the improvement of the image resolution, so that the SR reconstruction has important application value in the fields of remote sensing, medical treatment, video monitoring and the like.
Currently, SR methods are mainly divided into 3 types: interpolation-based, reconstruction-based, and learning-based SR methods. Among them, the SR method based on learning is a hot spot direction in recent years, and the SR method based on dictionary learning is the most popular one among the SR methods based on learning, and was proposed by Yang and the like for the earliest time. The method is provided based on a compressed sensing theory, a dictionary joint learning method is adopted to learn HR and LR dictionary pairs, LR sparse coefficients are obtained by utilizing LR image blocks and an LR dictionary, and then the LR sparse coefficients and the HR dictionary are used for reconstructing HR image blocks on the basis of the assumption that the HR and LR image blocks have the same sparse representation coefficients. The algorithm can acquire sufficient priori knowledge, has a good subjective visual effect, but has a remarkable blocking effect, and the reconstructed effect has great dependence on a learned dictionary and the reconstruction consumes a long time.
Disclosure of Invention
In view of the above, the invention aims to provide an image super-resolution method based on SAE dictionary learning and neighborhood regression, which improves the feature expression capability of a dictionary and reduces the dependency of a reconstruction result on the dictionary; on the other hand, a neighborhood regression theory is integrated, and the reconstruction speed is improved.
The invention is realized by adopting the following scheme: an image super-resolution method based on SAE dictionary learning and neighborhood regression specifically comprises the following steps:
preparing input data aiming at a dictionary learning model SAE, and constructing and training a dictionary;
solving a projection matrix by combining a neighborhood regression theory and a dictionary;
and reconstructing the image based on the projection matrix to obtain a high-resolution image.
Further, the preparing input data for the dictionary learning model SAE specifically includes:
HR image sample IhDownsampling to obtain LR image IlAnd up-sampling the LR image to obtain an intermediate image Im;
HR input data preparation: HR image IhAnd an intermediate image ImSubtracting to obtain a difference image IdFor difference image IdPartitioning and normalizing to obtain HR input datam is the number of samples of HR input data;
LR input data preparation: for intermediate image ImFiltering to obtain a filtered image, normalizing and partitioning the filtered image, and reducing dimensions of the filtered image block, which is recorded asn is the number of samples of LR input data;
input data of the dictionary learning model SAE is represented as S ═ Sh,Sl]。
Further, when LR input data is prepared, the Principal Component Analysis (PCA) is adopted to reduce the dimension of the filtering image block.
Further, the constructing and training of the dictionary specifically includes:
combining with the requirement of dictionary learning, in the cost function of SAE, adopting average absolute value error to replace mean square error to obtain an improved SAE dictionary learning model;
input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,Dl}。
Further, in combination with the requirement of dictionary learning, in the cost function of SAE, the mean absolute value error is used instead of the mean square error, and the improved SAE dictionary learning model is specifically:
let siE.g. S as input data, oiE O is taken as output data, and the improved SAE dictionary model is as follows:
wherein the first item JMAE(θ) is a reconstruction error term, here expressed in mean absolute error, m and n represent the number of samples of HR and LR input data, respectively; second item Jweight(theta) is a weight decay term used to reduce the magnitude of the weights, preventing overfitting, lambda is the tuning parameter of this term,represents the connection weight of the l-1 layer node i and the l layer node j, wherein l represents the layer number of the network, and NlIndicates the number of nodes of the l-th layer, Nl+1Number of nodes of layer l + 1; third item Jsparse(theta) is a hidden layer sparsity regularization term,for the mean activation of the neurons in the hidden layer, ρ is the set expected activation, γ is the regulatory parameter, N2Represents the number of nodes of layer 2; wherein,expressed by formula (2):
further, the input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,DlThe concrete steps are as follows:
in the training process of the SAE dictionary learning model, updating parameters by combining a gradient descent method, and finally obtaining the connection weight W from the input layer to the hidden layer1Wherein W is1={wi1,2, i, m + n; according to the relation between the network weight and the dictionary, the dictionary D is equivalent to the link weight W of the input layer and the hidden layer1Denoted HR dictionary Dh={w1,w2,…,wmR, LR dictionary Dl={wm+1,wm+2,…,wm+nDenoted D ═ D for dictionary pairh,Dl) Wherein w isi∈W1And w isi={w1,i,w2,i,...,wk,iK is the dimension of the dictionary, wk,iRepresenting the weight of the ith dictionary atom in the kth dimension.
Further, the solving of the projection matrix by combining the neighborhood regression theory and the dictionary specifically includes: first, the HR and LR dictionary pair { D is calculated by adopting a nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,Nl}; then, based on the mapping { Nh,NlAnd solving a projection matrix P by using a ridge regression method.
Further, the HR and LR dictionary pair { D is calculated by adopting a nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,NlThe concrete steps are as follows:
let S be ═ Sh,Sl]For training samples, { Dh,DlThe atom is calculated by Euclidean distanceIn LR dictionary DlK nearest neighbor domain block sets N in (1)l,q:
In the formula,represents the LR training sample SlThe p-th training sample of (1),the qth atom representing the LR dictionary;
from equation (3), K LR neighboring image blocks N corresponding to each LR dictionary atom can be calculatedl,qAccording to the positions of the K adjacent image blocks, corresponding K HR adjacent image blocks N can be obtained from the HR training samplesh,qWhen traversing the entire LR dictionary DlAll atoms of (2) can obtain a mapping relation N formed by combining nearest neighbor LR image blocks of all atomslAnd a nearest neighbor mapping relation N formed by combining the nearest neighbor HR image blocks corresponding to the nearest neighbor HR image blockshFinally obtain { Nh,Nl}。
Further, the mapping relation is based on { Nh,NlSolving the projection matrix P by using a ridge regression method specifically comprises the following steps:
using LR nearest neighbor mapping relation NlReplacing dictionary DlThen the expression for the reconstruction of beta is
In the formula, beta represents a coefficient matrix, Y represents a low-resolution image to be reconstructed, and eta is a weight coefficient and is used for relieving the singularity problem and ensuring the stability of coefficient decomposition;
solving the formula (4) by using a ridge regression method, and calculating the coefficientIs shown as
Reconstructed HR image X-passOver-mapping relation NhSum coefficientThe method comprises the steps of (1) obtaining,
Further, the image reconstruction based on the projection matrix to obtain the high-resolution image specifically includes:
firstly, preprocessing an image Y to be reconstructed to obtain an LR test characteristic image Yt={y1,y2,…,yi,…,yn}; then, the Euclidean distance is adopted to be in LR dictionary DlFinding LR test characteristic image block y iniCorresponding nearest neighbor dictionary atom dkThe expression isThen through atom dkFind its corresponding projection matrix PtThen using the expression xi=PtyiObtain a correspondence yiHR image block xiAll LR test characteristic image blocks are reconstructed in this way, and an HR image X is formed.
Compared with the prior art, the invention has the following beneficial effects: firstly, an improved sparse automatic encoder is adopted for dictionary learning, the outstanding feature learning capability of the improved sparse automatic encoder is fully utilized, the feature expression capability of the dictionary is enhanced, and the reconstruction quality of an image is improved; then, the invention integrates the neighborhood regression theory into the super-resolution frame based on dictionary learning, avoids the sparse coding process in the original frame, reduces the calculated amount and improves the reconstruction speed.
Drawings
FIG. 1 is a flow chart of SAE input data preprocessing according to an embodiment of the present invention.
FIG. 2 is a schematic flow chart of a method according to an embodiment of the present invention.
Detailed Description
The invention is further explained below with reference to the drawings and the embodiments.
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the disclosure. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present application. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
As shown in fig. 2, the present embodiment provides an image super-resolution method based on SAE dictionary learning and neighborhood regression, which specifically includes the following steps:
step S1: preparing input data aiming at a dictionary learning model SAE, and constructing and training a dictionary;
step S2: solving a projection matrix by combining a neighborhood regression theory and a dictionary;
step S3: and reconstructing the image based on the projection matrix to obtain a high-resolution image.
In this embodiment, as shown in fig. 1, the step S1 of preparing the input data for the dictionary learning model SAE specifically includes:
step S11: HR image sample IhDownsampling to obtain LR image IlAnd up-sampling the LR image to obtain an intermediate image Im;
Step S12: HR input data preparation: HR image IhAnd an intermediate image ImSubtracting to obtain a difference image IdFor difference image IdBlocking and normalizing asHR input data, notem is the number of samples of HR input data;
step S13: LR input data preparation: for intermediate image ImFiltering to obtain a filtered image, normalizing and partitioning the filtered image, and reducing dimensions of the filtered image block, which is recorded asn is the number of samples of LR input data;
step S14: input data of the dictionary learning model SAE is represented as S ═ Sh,Sl]。
When LR input data is prepared, the Principal Component Analysis (PCA) is adopted to reduce the dimension of the filtering image block.
In this embodiment, the constructing and training of the dictionary in step S1 specifically includes:
step S15: combining with the requirement of dictionary learning, in the cost function of SAE, adopting average absolute value error to replace mean square error to obtain an improved SAE dictionary learning model;
let siE.g. S as input data, oiE O is taken as output data, and the improved SAE dictionary model is as follows:
wherein the first item JMAE(θ) is a reconstruction error term, here expressed as Mean Squared Error (MSE), where m and n represent the number of samples of HR and LR input data, respectively; second item Jweight(theta) is a weight decay term used to reduce the magnitude of the weights, preventing overfitting, lambda is the tuning parameter of this term,represents the connection weight of the l-1 layer node i and the l layer node j, wherein l represents the layer number of the network, and NlNumber of nodes representing layer l, Nl+1Number of nodes of layer l + 1; third item Jsparse(theta) is a hidden layer sparsity regularization term,for the mean activation of hidden layer neurons, ρ is the set expected activation, with a value close to 0, γ is the regulatory parameter of the term, N2Represents the number of nodes of layer 2; for theThe case of significant deviation ρ is generally penalized with relative entropy, as shown in equation (2):
step S16: input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,Dl}。
In the training process of the SAE dictionary learning model, updating parameters by combining a gradient descent method, and finally obtaining the connection weight W from the input layer to the hidden layer1Wherein W is1={wi1,2, i, m + n; in dictionary learning, input data can be represented by a dictionary matrix and sparsely, whereas in SAE, input data can be represented by a hidden layer representation and learning weights, from the relationship between which it is known that dictionary D is equivalent to the link weights W of the input layer and hidden layer1Denoted HR dictionary Dh={w1,w2,…,wmR, LR dictionary Dl={wm+1,wm+2,…,wm+nDenoted D ═ D for dictionary pairh,Dl) Wherein w isi∈W1And w isi={w1,i,w2,i,...,wk,iK is the dimension of the dictionary, wk,iRepresenting the weight of the ith dictionary atom in the kth dimension.
In this embodiment, step S2 specifically includes:
step S21: HR and LR dictionary pair { D is calculated by adopting nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,Nl}; that is to say that the first and second electrodes,
let S be ═ Sh,Sl]For training samples, { Dh,DlThe atom is calculated by Euclidean distanceIn LR dictionary DlK nearest neighbor domain block sets N in (1)l,q:
In the formula,represents the LR training sample SlThe p-th training sample of (1),the qth atom representing the LR dictionary;
from equation (3), K LR neighboring image blocks N corresponding to each LR dictionary atom can be calculatedl,qAccording to the positions of the K adjacent image blocks, corresponding K HR adjacent image blocks N can be obtained from the HR training samplesh,qWhen traversing the entire LR dictionary DlAll atoms of (2) can obtain a mapping relation N formed by combining nearest neighbor LR image blocks of all atomslAnd a nearest neighbor mapping relation N formed by combining the nearest neighbor HR image blocks corresponding to the nearest neighbor HR image blockshFinally obtain { Nh,Nl}。
Step S22: based on the mapping relation { Nh,NlSolving a projection matrix P by a ridge regression method, namely,
using LR nearest neighbor mapping relation NlReplacing dictionary DlThen the expression for the reconstruction of beta is
In the formula, beta represents a coefficient matrix, Y represents a low-resolution image to be reconstructed, and eta is a weight coefficient and is used for relieving the singularity problem and ensuring the stability of coefficient decomposition;
solving the formula (4) by using a ridge regression method, and calculating the coefficientIs shown as
The reconstructed HR image X passes through the mapping relation NhSum coefficientThe method comprises the steps of (1) obtaining,
In this embodiment, step S3 specifically includes:
firstly, preprocessing an image Y to be reconstructed to obtain an LR test characteristic image Yt={y1,y2,…,yi,…,yn}; then, the Euclidean distance is adopted to be in LR dictionary DlFinding LR test characteristic image block y iniCorresponding nearest neighbor dictionary atom dkThe expression isThen through atom dkFind its corresponding projection matrix PtThen use the tableExpression xi=PtyiObtain a correspondence yiHR image block xiAll LR test characteristic image blocks are reconstructed in this way, and an HR image X is formed.
Next, the present embodiment will be further described using the following simulation experiment.
The simulation tool adopted in the embodiment is MATLAB, and the evaluation indexes are peak signal-to-noise ratio PSNR and structural similarity SSIM, wherein the larger the PSNR, the closer the SSIM is to 1, and the better the super-resolution effect is.
The simulation experiment is specifically set as follows:
preparing data: in order to ensure the objectivity of the experiment, 91 universal standard HR training samples are adopted in dictionary learning, and test images are derived from standard test libraries Set5 and Set 14. In order to quantitatively evaluate the quality of the reconstructed images, these test images are taken as HR reference images, and LR images to be processed are acquired by down-sampling.
And (3) comparison algorithm: 5 SR algorithms such as Bicubic, L1SR (Super Resolution with L1 Resolution), SISR (Single Image Super Resolution), ANR (ordered neighbor Resolution), SRCNN (Super Resolution using volumetric Neural network) and the like are compared.
Setting important parameters: the sampling factor is 3 and the number of nodes of the SAE hidden layer is 1024.
The simulation experiments were mainly divided into 2 groups, as follows.
Experiment set 1: compared to different SR methods.
Table 1 lists PSNR and SSIM corresponding to the reconstructed images obtained by different SR algorithms, where the values in the last column indicate that the performance of the corresponding algorithm of this embodiment is optimal under the corresponding evaluation indexes, and 10 images are from Set5 and Set14, respectively. As can be seen from table 1, the PSNR and SSIM values obtained by the method of the present embodiment are substantially optimal, which indicates that the reconstruction effect is better.
TABLE 1 comparison of PSNR (dB) and SSIM values for different SR methods
Experiment set 2: and (4) comparing reconstruction speeds.
In the experiment, different algorithms are operated in the same equipment and environment, and the method is used for verifying the effect of reconstruction speed improved by blending the neighborhood regression thought into the SR algorithm based on dictionary learning. The average reconstruction times for the Set5 and Set14 test sets under different SR algorithms are listed in table 2. It can be seen that the present invention is significantly faster than other SR algorithms.
TABLE 2 comparison of different SR calculated mean reconstruction times(s)
Test image library | L1SR | SISR | SRCNN | Method for producing a composite material |
Set 5 | 14.28 | 0.96 | 2.98 | 0.33 |
Set 14 | 31.93 | 1.93 | 8.03 | 0.65 |
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The foregoing is directed to preferred embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow. However, any simple modification, equivalent change and modification of the above embodiments according to the technical essence of the present invention are within the protection scope of the technical solution of the present invention.
Claims (10)
1. An image super-resolution method based on SAE dictionary learning and neighborhood regression is characterized by comprising the following steps:
preparing input data aiming at a dictionary learning model SAE, and constructing and training a dictionary;
solving a projection matrix by combining a neighborhood regression theory and a dictionary;
and reconstructing the image based on the projection matrix to obtain a high-resolution image.
2. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 1, wherein the preparation of input data for the dictionary learning model SAE is specifically as follows:
HR image sample IhDownsampling to obtain LR image IlAnd up-sampling the LR image to obtain an intermediate image Im;
HR input data preparation: HR image IhAnd an intermediate image ImSubtracting to obtain a difference image IdFor difference image IdPartitioning and normalizing to obtain HR input datam is the number of samples of HR input data;
LR input data preparation: for intermediate image ImFiltering to obtain a filtered image, normalizing and partitioning the filtered image, and reducing dimensions of the filtered image block, which is recorded asn is the number of samples of LR input data;
input data of the dictionary learning model SAE is represented as S ═ Sh,Sl]。
3. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 2, wherein in LR input data preparation, a Principal Component Analysis (PCA) method is used to reduce the dimension of the filtering image block.
4. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 1, wherein said constructing and training of the dictionary specifically comprises:
combining with the requirement of dictionary learning, in the cost function of SAE, adopting average absolute value error to replace mean square error to obtain an improved SAE dictionary learning model;
input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,Dl}。
5. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 4, wherein said combination of the requirement of dictionary learning, in the cost function of SAE, the mean absolute value error is used to replace the mean square error, and the improved SAE dictionary learning model is specifically:
let siE.g. S as input data, oiE O is taken as output data, and the improved SAE dictionary model is as follows:
wherein the first item JMAE(θ) is a reconstruction error term, here expressed in mean absolute error, m and n represent HR and HR, respectivelyNumber of samples of LR input data; second item Jweight(theta) is a weight decay term used to reduce the magnitude of the weights, preventing overfitting, lambda is the tuning parameter of this term,represents the connection weight of the l-1 layer node i and the l layer node j, wherein l represents the layer number of the network, and NlIndicates the number of nodes of the l-th layer, Nl+1Number of nodes of layer l + 1; third item Jsparse(theta) is a hidden layer sparsity regularization term,for the mean activation of the neurons in the hidden layer, ρ is the set expected activation, γ is the regulatory parameter, N2Represents the number of nodes of layer 2; wherein,expressed by formula (2):
6. the image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 4, wherein said input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,DlThe concrete steps are as follows:
in the training process of the SAE dictionary learning model, updating parameters by combining a gradient descent method, and finally obtaining the connection weight W from the input layer to the hidden layer1Wherein W is1={wi1,2, i, m + n; according to the relation between the network weight and the dictionary, the dictionary D is equivalent to the link weight W of the input layer and the hidden layer1Denoted HR dictionary Dh={w1,w2,…,wmR, LR dictionary Dl={wm+1,wm+2,…,wm+nDenoted D ═ D for dictionary pairh,Dl) Wherein w isi∈W1And w isi={w1,i,w2,i,...,wk,iK is the dimension of the dictionary, wk,iRepresenting the weight of the ith dictionary atom in the kth dimension.
7. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 1, wherein said solving of projection matrix in combination with neighborhood regression theory and dictionary specifically comprises: first, the HR and LR dictionary pair { D is calculated by adopting a nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,Nl}; then, based on the mapping { Nh,NlAnd solving a projection matrix P by using a ridge regression method.
8. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 7, wherein said computing HR and LR dictionary pair { D ] by nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,NlThe concrete steps are as follows:
let S be ═ Sh,Sl]For training samples, { Dh,DlThe atom is calculated by Euclidean distanceIn LR dictionary DlK nearest neighbor domain block sets N in (1)l,q:
In the formula,represents the LR training sample SlThe p-th training sample of (1),representation of LR dictionary DlThe qth atom of (1);
from equation (3), K LR neighboring image blocks N corresponding to each LR dictionary atom can be calculatedl,qAccording to the positions of the K adjacent image blocks, obtaining corresponding K HR adjacent image blocks N from HR training samplesh,qWhen traversing the entire LR dictionary DlObtaining the mapping relation N formed by combining the nearest neighbor LR image blocks of all atomslAnd a nearest neighbor mapping relation N formed by combining the nearest neighbor HR image blocks corresponding to the nearest neighbor HR image blockshFinally obtain { Nh,Nl}。
9. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 7, wherein said mapping relation { N }h,NlSolving the projection matrix P by using a ridge regression method specifically comprises the following steps:
using LR nearest neighbor mapping relation NlReplacing dictionary DlThen the expression for the reconstruction of beta is
In the formula, beta represents a coefficient matrix, Y represents a low-resolution image to be reconstructed, and eta is a weight coefficient and is used for relieving the singularity problem and ensuring the stability of coefficient decomposition;
solving the formula (4) by using a ridge regression method, and calculating the coefficientIs shown as
The reconstructed HR image X passes through the mapping relation NhSum coefficientThe method comprises the steps of (1) obtaining,
note N in the formula (6)h(Nl TNl+ηI)-1Nl TIs a projection matrix P.
10. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 7, wherein said image reconstruction based on projection matrix to obtain high resolution image specifically comprises:
firstly, preprocessing an image Y to be reconstructed to obtain an LR test characteristic image Yt={y1,y2,…,yi,…,yn}; then, the Euclidean distance is adopted to be in LR dictionary DlFinding LR test characteristic image block y iniCorresponding nearest neighbor dictionary atom dkThe expression isThen through atom dkFind its corresponding projection matrix PtThen using the expression xi=PtyiObtain a correspondence yiHR image block xiAll LR test characteristic image blocks are reconstructed in this way, and an HR image X is formed.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010670836.8A CN111833252B (en) | 2020-07-13 | 2020-07-13 | Image super-resolution method based on SAE dictionary learning and neighborhood regression |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010670836.8A CN111833252B (en) | 2020-07-13 | 2020-07-13 | Image super-resolution method based on SAE dictionary learning and neighborhood regression |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111833252A true CN111833252A (en) | 2020-10-27 |
CN111833252B CN111833252B (en) | 2022-06-14 |
Family
ID=72922812
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010670836.8A Active CN111833252B (en) | 2020-07-13 | 2020-07-13 | Image super-resolution method based on SAE dictionary learning and neighborhood regression |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111833252B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109146785A (en) * | 2018-08-02 | 2019-01-04 | 华侨大学 | A kind of image super-resolution method based on the sparse autocoder of improvement |
CN109712074A (en) * | 2018-12-20 | 2019-05-03 | 黑龙江大学 | The remote sensing images super-resolution reconstruction method of two-parameter beta combine processes dictionary |
CN110097503A (en) * | 2019-04-12 | 2019-08-06 | 浙江师范大学 | The super-resolution method returned based on neighborhood |
EP3591608A1 (en) * | 2018-07-06 | 2020-01-08 | Tata Consultancy Services Limited | Method and system for image reconstruction using deep dictionary learning (ddl) |
-
2020
- 2020-07-13 CN CN202010670836.8A patent/CN111833252B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3591608A1 (en) * | 2018-07-06 | 2020-01-08 | Tata Consultancy Services Limited | Method and system for image reconstruction using deep dictionary learning (ddl) |
CN109146785A (en) * | 2018-08-02 | 2019-01-04 | 华侨大学 | A kind of image super-resolution method based on the sparse autocoder of improvement |
CN109712074A (en) * | 2018-12-20 | 2019-05-03 | 黑龙江大学 | The remote sensing images super-resolution reconstruction method of two-parameter beta combine processes dictionary |
CN110097503A (en) * | 2019-04-12 | 2019-08-06 | 浙江师范大学 | The super-resolution method returned based on neighborhood |
Non-Patent Citations (2)
Title |
---|
詹曙等: "基于耦合特征空间下改进字典学习的图像超分辨率重建", 《电子学报》 * |
黄德天等: "正则化技术和低秩矩阵在稀疏表示超分辨率算法中的应用", 《计算机辅助设计与图形学学报》 * |
Also Published As
Publication number | Publication date |
---|---|
CN111833252B (en) | 2022-06-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Deng et al. | Wavelet domain style transfer for an effective perception-distortion tradeoff in single image super-resolution | |
CN110443768B (en) | Single-frame image super-resolution reconstruction method based on multiple consistency constraints | |
CN103093445B (en) | Unified feature space image super-resolution reconstruction method based on joint sparse constraint | |
CN113298147B (en) | Image fusion method and device based on regional energy and intuitionistic fuzzy set | |
CN105631807A (en) | Single-frame image super resolution reconstruction method based on sparse domain selection | |
CN107590779A (en) | Image denoising and deblurring method based on image block clustering dictionary training | |
CN108830791B (en) | Image super-resolution method based on self sample and sparse representation | |
CN112150354A (en) | Single image super-resolution method combining contour enhancement and denoising statistical prior | |
CN111861886B (en) | Image super-resolution reconstruction method based on multi-scale feedback network | |
CN114820352A (en) | Hyperspectral image denoising method and device and storage medium | |
CN114648048B (en) | Electrocardiosignal noise reduction method based on variational self-coding and PixelCNN model | |
CN105590296B (en) | A kind of single-frame images Super-Resolution method based on doubledictionary study | |
CN101739670B (en) | Non-local mean space domain time varying image filtering method | |
CN115578262A (en) | Polarization image super-resolution reconstruction method based on AFAN model | |
CN115526779A (en) | Infrared image super-resolution reconstruction method based on dynamic attention mechanism | |
Roy et al. | Recent study on image denoising using deep cnn techniques | |
CN106296583B (en) | Based on image block group sparse coding and the noisy high spectrum image ultra-resolution ratio reconstructing method that in pairs maps | |
Tomosada et al. | GAN-based image deblurring using DCT discriminator | |
CN111833252B (en) | Image super-resolution method based on SAE dictionary learning and neighborhood regression | |
Hua et al. | An Efficient Multiscale Spatial Rearrangement MLP Architecture for Image Restoration | |
CN112767273A (en) | Low-dose CT image restoration method and system applying feature decoupling | |
CN116612009A (en) | Multi-scale connection generation countermeasure network medical image super-resolution reconstruction method | |
CN111951181A (en) | Hyperspectral image denoising method based on non-local similarity and weighted truncation kernel norm | |
CN108846797B (en) | Image super-resolution method based on two training sets | |
CN116071237A (en) | Video hyperspectral imaging method, system and medium based on filter sampling fusion |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |