CN111833252A - Image super-resolution method based on SAE dictionary learning and neighborhood regression - Google Patents

Image super-resolution method based on SAE dictionary learning and neighborhood regression Download PDF

Info

Publication number
CN111833252A
CN111833252A CN202010670836.8A CN202010670836A CN111833252A CN 111833252 A CN111833252 A CN 111833252A CN 202010670836 A CN202010670836 A CN 202010670836A CN 111833252 A CN111833252 A CN 111833252A
Authority
CN
China
Prior art keywords
dictionary
image
sae
dictionary learning
input data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010670836.8A
Other languages
Chinese (zh)
Other versions
CN111833252B (en
Inventor
黄炜钦
郭一晶
陈俊仁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen University Tan Kah Kee College
Original Assignee
Xiamen University Tan Kah Kee College
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen University Tan Kah Kee College filed Critical Xiamen University Tan Kah Kee College
Priority to CN202010670836.8A priority Critical patent/CN111833252B/en
Publication of CN111833252A publication Critical patent/CN111833252A/en
Application granted granted Critical
Publication of CN111833252B publication Critical patent/CN111833252B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2135Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/28Determining representative reference patterns, e.g. by averaging or distorting; Generating dictionaries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to an image super-resolution method based on SAE dictionary learning and neighborhood regression, which comprises the steps of firstly preparing input data aiming at a dictionary learning model SAE, and constructing and training a dictionary; then solving a projection matrix by combining a neighborhood regression theory and a dictionary; and finally, reconstructing the image based on the projection matrix to obtain a high-resolution image. On one hand, the method improves the feature expression capability of the dictionary and reduces the dependency of the reconstruction result on the dictionary; on the other hand, a neighborhood regression theory is integrated, and the reconstruction speed is improved.

Description

Image super-resolution method based on SAE dictionary learning and neighborhood regression
Technical Field
The invention relates to the field of image super-resolution method design, in particular to an image super-resolution method based on SAE dictionary learning and neighborhood regression.
Background
In reality, due to the limitations of image acquisition equipment, scene changes, light sources and other factors, high-quality images cannot be obtained, and the requirements of practical application cannot be met when the resolution of the images is low. The Super-Resolution (SR) method utilizes an image signal processing technique to reconstruct a single or multiple low-Resolution (LR) images into High-Resolution (HR) images, and is characterized in that certain additional information is added during the reconstruction process to compensate for the detail information lost during the image degradation process. The SR reconstruction can break through the limitation of the intrinsic resolution of the imaging device to realize the improvement of the image resolution, so that the SR reconstruction has important application value in the fields of remote sensing, medical treatment, video monitoring and the like.
Currently, SR methods are mainly divided into 3 types: interpolation-based, reconstruction-based, and learning-based SR methods. Among them, the SR method based on learning is a hot spot direction in recent years, and the SR method based on dictionary learning is the most popular one among the SR methods based on learning, and was proposed by Yang and the like for the earliest time. The method is provided based on a compressed sensing theory, a dictionary joint learning method is adopted to learn HR and LR dictionary pairs, LR sparse coefficients are obtained by utilizing LR image blocks and an LR dictionary, and then the LR sparse coefficients and the HR dictionary are used for reconstructing HR image blocks on the basis of the assumption that the HR and LR image blocks have the same sparse representation coefficients. The algorithm can acquire sufficient priori knowledge, has a good subjective visual effect, but has a remarkable blocking effect, and the reconstructed effect has great dependence on a learned dictionary and the reconstruction consumes a long time.
Disclosure of Invention
In view of the above, the invention aims to provide an image super-resolution method based on SAE dictionary learning and neighborhood regression, which improves the feature expression capability of a dictionary and reduces the dependency of a reconstruction result on the dictionary; on the other hand, a neighborhood regression theory is integrated, and the reconstruction speed is improved.
The invention is realized by adopting the following scheme: an image super-resolution method based on SAE dictionary learning and neighborhood regression specifically comprises the following steps:
preparing input data aiming at a dictionary learning model SAE, and constructing and training a dictionary;
solving a projection matrix by combining a neighborhood regression theory and a dictionary;
and reconstructing the image based on the projection matrix to obtain a high-resolution image.
Further, the preparing input data for the dictionary learning model SAE specifically includes:
HR image sample IhDownsampling to obtain LR image IlAnd up-sampling the LR image to obtain an intermediate image Im
HR input data preparation: HR image IhAnd an intermediate image ImSubtracting to obtain a difference image IdFor difference image IdPartitioning and normalizing to obtain HR input data
Figure BDA0002582207870000021
m is the number of samples of HR input data;
LR input data preparation: for intermediate image ImFiltering to obtain a filtered image, normalizing and partitioning the filtered image, and reducing dimensions of the filtered image block, which is recorded as
Figure BDA0002582207870000022
n is the number of samples of LR input data;
input data of the dictionary learning model SAE is represented as S ═ Sh,Sl]。
Further, when LR input data is prepared, the Principal Component Analysis (PCA) is adopted to reduce the dimension of the filtering image block.
Further, the constructing and training of the dictionary specifically includes:
combining with the requirement of dictionary learning, in the cost function of SAE, adopting average absolute value error to replace mean square error to obtain an improved SAE dictionary learning model;
input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,Dl}。
Further, in combination with the requirement of dictionary learning, in the cost function of SAE, the mean absolute value error is used instead of the mean square error, and the improved SAE dictionary learning model is specifically:
let siE.g. S as input data, oiE O is taken as output data, and the improved SAE dictionary model is as follows:
Figure BDA0002582207870000081
wherein the first item JMAE(θ) is a reconstruction error term, here expressed in mean absolute error, m and n represent the number of samples of HR and LR input data, respectively; second item Jweight(theta) is a weight decay term used to reduce the magnitude of the weights, preventing overfitting, lambda is the tuning parameter of this term,
Figure BDA0002582207870000032
represents the connection weight of the l-1 layer node i and the l layer node j, wherein l represents the layer number of the network, and NlIndicates the number of nodes of the l-th layer, Nl+1Number of nodes of layer l + 1; third item Jsparse(theta) is a hidden layer sparsity regularization term,
Figure BDA0002582207870000033
for the mean activation of the neurons in the hidden layer, ρ is the set expected activation, γ is the regulatory parameter, N2Represents the number of nodes of layer 2; wherein,
Figure BDA0002582207870000034
expressed by formula (2):
Figure BDA0002582207870000035
further, the input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,DlThe concrete steps are as follows:
in the training process of the SAE dictionary learning model, updating parameters by combining a gradient descent method, and finally obtaining the connection weight W from the input layer to the hidden layer1Wherein W is1={wi1,2, i, m + n; according to the relation between the network weight and the dictionary, the dictionary D is equivalent to the link weight W of the input layer and the hidden layer1Denoted HR dictionary Dh={w1,w2,…,wmR, LR dictionary Dl={wm+1,wm+2,…,wm+nDenoted D ═ D for dictionary pairh,Dl) Wherein w isi∈W1And w isi={w1,i,w2,i,...,wk,iK is the dimension of the dictionary, wk,iRepresenting the weight of the ith dictionary atom in the kth dimension.
Further, the solving of the projection matrix by combining the neighborhood regression theory and the dictionary specifically includes: first, the HR and LR dictionary pair { D is calculated by adopting a nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,Nl}; then, based on the mapping { Nh,NlAnd solving a projection matrix P by using a ridge regression method.
Further, the HR and LR dictionary pair { D is calculated by adopting a nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,NlThe concrete steps are as follows:
let S be ═ Sh,Sl]For training samples, { Dh,DlThe atom is calculated by Euclidean distance
Figure BDA0002582207870000041
In LR dictionary DlK nearest neighbor domain block sets N in (1)l,q
Figure BDA0002582207870000042
In the formula,
Figure BDA0002582207870000043
represents the LR training sample SlThe p-th training sample of (1),
Figure BDA0002582207870000044
the qth atom representing the LR dictionary;
from equation (3), K LR neighboring image blocks N corresponding to each LR dictionary atom can be calculatedl,qAccording to the positions of the K adjacent image blocks, corresponding K HR adjacent image blocks N can be obtained from the HR training samplesh,qWhen traversing the entire LR dictionary DlAll atoms of (2) can obtain a mapping relation N formed by combining nearest neighbor LR image blocks of all atomslAnd a nearest neighbor mapping relation N formed by combining the nearest neighbor HR image blocks corresponding to the nearest neighbor HR image blockshFinally obtain { Nh,Nl}。
Further, the mapping relation is based on { Nh,NlSolving the projection matrix P by using a ridge regression method specifically comprises the following steps:
using LR nearest neighbor mapping relation NlReplacing dictionary DlThen the expression for the reconstruction of beta is
Figure BDA0002582207870000051
Figure BDA0002582207870000052
In the formula, beta represents a coefficient matrix, Y represents a low-resolution image to be reconstructed, and eta is a weight coefficient and is used for relieving the singularity problem and ensuring the stability of coefficient decomposition;
solving the formula (4) by using a ridge regression method, and calculating the coefficient
Figure BDA0002582207870000053
Is shown as
Figure BDA0002582207870000054
Reconstructed HR image X-passOver-mapping relation NhSum coefficient
Figure BDA0002582207870000055
The method comprises the steps of (1) obtaining,
Figure BDA0002582207870000056
in the formula (6)
Figure BDA0002582207870000057
Is a projection matrix P.
Further, the image reconstruction based on the projection matrix to obtain the high-resolution image specifically includes:
firstly, preprocessing an image Y to be reconstructed to obtain an LR test characteristic image Yt={y1,y2,…,yi,…,yn}; then, the Euclidean distance is adopted to be in LR dictionary DlFinding LR test characteristic image block y iniCorresponding nearest neighbor dictionary atom dkThe expression is
Figure BDA0002582207870000058
Then through atom dkFind its corresponding projection matrix PtThen using the expression xi=PtyiObtain a correspondence yiHR image block xiAll LR test characteristic image blocks are reconstructed in this way, and an HR image X is formed.
Compared with the prior art, the invention has the following beneficial effects: firstly, an improved sparse automatic encoder is adopted for dictionary learning, the outstanding feature learning capability of the improved sparse automatic encoder is fully utilized, the feature expression capability of the dictionary is enhanced, and the reconstruction quality of an image is improved; then, the invention integrates the neighborhood regression theory into the super-resolution frame based on dictionary learning, avoids the sparse coding process in the original frame, reduces the calculated amount and improves the reconstruction speed.
Drawings
FIG. 1 is a flow chart of SAE input data preprocessing according to an embodiment of the present invention.
FIG. 2 is a schematic flow chart of a method according to an embodiment of the present invention.
Detailed Description
The invention is further explained below with reference to the drawings and the embodiments.
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the disclosure. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present application. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
As shown in fig. 2, the present embodiment provides an image super-resolution method based on SAE dictionary learning and neighborhood regression, which specifically includes the following steps:
step S1: preparing input data aiming at a dictionary learning model SAE, and constructing and training a dictionary;
step S2: solving a projection matrix by combining a neighborhood regression theory and a dictionary;
step S3: and reconstructing the image based on the projection matrix to obtain a high-resolution image.
In this embodiment, as shown in fig. 1, the step S1 of preparing the input data for the dictionary learning model SAE specifically includes:
step S11: HR image sample IhDownsampling to obtain LR image IlAnd up-sampling the LR image to obtain an intermediate image Im
Step S12: HR input data preparation: HR image IhAnd an intermediate image ImSubtracting to obtain a difference image IdFor difference image IdBlocking and normalizing asHR input data, note
Figure BDA0002582207870000071
m is the number of samples of HR input data;
step S13: LR input data preparation: for intermediate image ImFiltering to obtain a filtered image, normalizing and partitioning the filtered image, and reducing dimensions of the filtered image block, which is recorded as
Figure BDA0002582207870000072
n is the number of samples of LR input data;
step S14: input data of the dictionary learning model SAE is represented as S ═ Sh,Sl]。
When LR input data is prepared, the Principal Component Analysis (PCA) is adopted to reduce the dimension of the filtering image block.
In this embodiment, the constructing and training of the dictionary in step S1 specifically includes:
step S15: combining with the requirement of dictionary learning, in the cost function of SAE, adopting average absolute value error to replace mean square error to obtain an improved SAE dictionary learning model;
let siE.g. S as input data, oiE O is taken as output data, and the improved SAE dictionary model is as follows:
Figure BDA0002582207870000081
wherein the first item JMAE(θ) is a reconstruction error term, here expressed as Mean Squared Error (MSE), where m and n represent the number of samples of HR and LR input data, respectively; second item Jweight(theta) is a weight decay term used to reduce the magnitude of the weights, preventing overfitting, lambda is the tuning parameter of this term,
Figure BDA0002582207870000082
represents the connection weight of the l-1 layer node i and the l layer node j, wherein l represents the layer number of the network, and NlNumber of nodes representing layer l, Nl+1Number of nodes of layer l + 1; third item Jsparse(theta) is a hidden layer sparsity regularization term,
Figure BDA0002582207870000083
for the mean activation of hidden layer neurons, ρ is the set expected activation, with a value close to 0, γ is the regulatory parameter of the term, N2Represents the number of nodes of layer 2; for the
Figure BDA0002582207870000084
The case of significant deviation ρ is generally penalized with relative entropy, as shown in equation (2):
Figure BDA0002582207870000085
step S16: input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,Dl}。
In the training process of the SAE dictionary learning model, updating parameters by combining a gradient descent method, and finally obtaining the connection weight W from the input layer to the hidden layer1Wherein W is1={wi1,2, i, m + n; in dictionary learning, input data can be represented by a dictionary matrix and sparsely, whereas in SAE, input data can be represented by a hidden layer representation and learning weights, from the relationship between which it is known that dictionary D is equivalent to the link weights W of the input layer and hidden layer1Denoted HR dictionary Dh={w1,w2,…,wmR, LR dictionary Dl={wm+1,wm+2,…,wm+nDenoted D ═ D for dictionary pairh,Dl) Wherein w isi∈W1And w isi={w1,i,w2,i,...,wk,iK is the dimension of the dictionary, wk,iRepresenting the weight of the ith dictionary atom in the kth dimension.
In this embodiment, step S2 specifically includes:
step S21: HR and LR dictionary pair { D is calculated by adopting nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,Nl}; that is to say that the first and second electrodes,
let S be ═ Sh,Sl]For training samples, { Dh,DlThe atom is calculated by Euclidean distance
Figure BDA0002582207870000091
In LR dictionary DlK nearest neighbor domain block sets N in (1)l,q
Figure BDA0002582207870000092
In the formula,
Figure BDA0002582207870000093
represents the LR training sample SlThe p-th training sample of (1),
Figure BDA0002582207870000094
the qth atom representing the LR dictionary;
from equation (3), K LR neighboring image blocks N corresponding to each LR dictionary atom can be calculatedl,qAccording to the positions of the K adjacent image blocks, corresponding K HR adjacent image blocks N can be obtained from the HR training samplesh,qWhen traversing the entire LR dictionary DlAll atoms of (2) can obtain a mapping relation N formed by combining nearest neighbor LR image blocks of all atomslAnd a nearest neighbor mapping relation N formed by combining the nearest neighbor HR image blocks corresponding to the nearest neighbor HR image blockshFinally obtain { Nh,Nl}。
Step S22: based on the mapping relation { Nh,NlSolving a projection matrix P by a ridge regression method, namely,
using LR nearest neighbor mapping relation NlReplacing dictionary DlThen the expression for the reconstruction of beta is
Figure BDA0002582207870000095
Figure BDA0002582207870000096
In the formula, beta represents a coefficient matrix, Y represents a low-resolution image to be reconstructed, and eta is a weight coefficient and is used for relieving the singularity problem and ensuring the stability of coefficient decomposition;
solving the formula (4) by using a ridge regression method, and calculating the coefficient
Figure BDA0002582207870000097
Is shown as
Figure BDA0002582207870000098
The reconstructed HR image X passes through the mapping relation NhSum coefficient
Figure BDA0002582207870000099
The method comprises the steps of (1) obtaining,
Figure BDA0002582207870000101
in the formula (6)
Figure BDA0002582207870000102
Is a projection matrix P.
In this embodiment, step S3 specifically includes:
firstly, preprocessing an image Y to be reconstructed to obtain an LR test characteristic image Yt={y1,y2,…,yi,…,yn}; then, the Euclidean distance is adopted to be in LR dictionary DlFinding LR test characteristic image block y iniCorresponding nearest neighbor dictionary atom dkThe expression is
Figure BDA0002582207870000103
Then through atom dkFind its corresponding projection matrix PtThen use the tableExpression xi=PtyiObtain a correspondence yiHR image block xiAll LR test characteristic image blocks are reconstructed in this way, and an HR image X is formed.
Next, the present embodiment will be further described using the following simulation experiment.
The simulation tool adopted in the embodiment is MATLAB, and the evaluation indexes are peak signal-to-noise ratio PSNR and structural similarity SSIM, wherein the larger the PSNR, the closer the SSIM is to 1, and the better the super-resolution effect is.
The simulation experiment is specifically set as follows:
preparing data: in order to ensure the objectivity of the experiment, 91 universal standard HR training samples are adopted in dictionary learning, and test images are derived from standard test libraries Set5 and Set 14. In order to quantitatively evaluate the quality of the reconstructed images, these test images are taken as HR reference images, and LR images to be processed are acquired by down-sampling.
And (3) comparison algorithm: 5 SR algorithms such as Bicubic, L1SR (Super Resolution with L1 Resolution), SISR (Single Image Super Resolution), ANR (ordered neighbor Resolution), SRCNN (Super Resolution using volumetric Neural network) and the like are compared.
Setting important parameters: the sampling factor is 3 and the number of nodes of the SAE hidden layer is 1024.
The simulation experiments were mainly divided into 2 groups, as follows.
Experiment set 1: compared to different SR methods.
Table 1 lists PSNR and SSIM corresponding to the reconstructed images obtained by different SR algorithms, where the values in the last column indicate that the performance of the corresponding algorithm of this embodiment is optimal under the corresponding evaluation indexes, and 10 images are from Set5 and Set14, respectively. As can be seen from table 1, the PSNR and SSIM values obtained by the method of the present embodiment are substantially optimal, which indicates that the reconstruction effect is better.
TABLE 1 comparison of PSNR (dB) and SSIM values for different SR methods
Figure BDA0002582207870000111
Experiment set 2: and (4) comparing reconstruction speeds.
In the experiment, different algorithms are operated in the same equipment and environment, and the method is used for verifying the effect of reconstruction speed improved by blending the neighborhood regression thought into the SR algorithm based on dictionary learning. The average reconstruction times for the Set5 and Set14 test sets under different SR algorithms are listed in table 2. It can be seen that the present invention is significantly faster than other SR algorithms.
TABLE 2 comparison of different SR calculated mean reconstruction times(s)
Test image library L1SR SISR SRCNN Method for producing a composite material
Set 5 14.28 0.96 2.98 0.33
Set 14 31.93 1.93 8.03 0.65
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The foregoing is directed to preferred embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow. However, any simple modification, equivalent change and modification of the above embodiments according to the technical essence of the present invention are within the protection scope of the technical solution of the present invention.

Claims (10)

1. An image super-resolution method based on SAE dictionary learning and neighborhood regression is characterized by comprising the following steps:
preparing input data aiming at a dictionary learning model SAE, and constructing and training a dictionary;
solving a projection matrix by combining a neighborhood regression theory and a dictionary;
and reconstructing the image based on the projection matrix to obtain a high-resolution image.
2. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 1, wherein the preparation of input data for the dictionary learning model SAE is specifically as follows:
HR image sample IhDownsampling to obtain LR image IlAnd up-sampling the LR image to obtain an intermediate image Im
HR input data preparation: HR image IhAnd an intermediate image ImSubtracting to obtain a difference image IdFor difference image IdPartitioning and normalizing to obtain HR input data
Figure FDA0002582207860000011
m is the number of samples of HR input data;
LR input data preparation: for intermediate image ImFiltering to obtain a filtered image, normalizing and partitioning the filtered image, and reducing dimensions of the filtered image block, which is recorded as
Figure FDA0002582207860000012
n is the number of samples of LR input data;
input data of the dictionary learning model SAE is represented as S ═ Sh,Sl]。
3. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 2, wherein in LR input data preparation, a Principal Component Analysis (PCA) method is used to reduce the dimension of the filtering image block.
4. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 1, wherein said constructing and training of the dictionary specifically comprises:
combining with the requirement of dictionary learning, in the cost function of SAE, adopting average absolute value error to replace mean square error to obtain an improved SAE dictionary learning model;
input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,Dl}。
5. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 4, wherein said combination of the requirement of dictionary learning, in the cost function of SAE, the mean absolute value error is used to replace the mean square error, and the improved SAE dictionary learning model is specifically:
let siE.g. S as input data, oiE O is taken as output data, and the improved SAE dictionary model is as follows:
Figure DEST_PATH_BDA0002582207870000081
wherein the first item JMAE(θ) is a reconstruction error term, here expressed in mean absolute error, m and n represent HR and HR, respectivelyNumber of samples of LR input data; second item Jweight(theta) is a weight decay term used to reduce the magnitude of the weights, preventing overfitting, lambda is the tuning parameter of this term,
Figure FDA0002582207860000022
represents the connection weight of the l-1 layer node i and the l layer node j, wherein l represents the layer number of the network, and NlIndicates the number of nodes of the l-th layer, Nl+1Number of nodes of layer l + 1; third item Jsparse(theta) is a hidden layer sparsity regularization term,
Figure FDA0002582207860000023
for the mean activation of the neurons in the hidden layer, ρ is the set expected activation, γ is the regulatory parameter, N2Represents the number of nodes of layer 2; wherein,
Figure FDA0002582207860000024
expressed by formula (2):
Figure FDA0002582207860000025
6. the image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 4, wherein said input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,DlThe concrete steps are as follows:
in the training process of the SAE dictionary learning model, updating parameters by combining a gradient descent method, and finally obtaining the connection weight W from the input layer to the hidden layer1Wherein W is1={wi1,2, i, m + n; according to the relation between the network weight and the dictionary, the dictionary D is equivalent to the link weight W of the input layer and the hidden layer1Denoted HR dictionary Dh={w1,w2,…,wmR, LR dictionary Dl={wm+1,wm+2,…,wm+nDenoted D ═ D for dictionary pairh,Dl) Wherein w isi∈W1And w isi={w1,i,w2,i,...,wk,iK is the dimension of the dictionary, wk,iRepresenting the weight of the ith dictionary atom in the kth dimension.
7. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 1, wherein said solving of projection matrix in combination with neighborhood regression theory and dictionary specifically comprises: first, the HR and LR dictionary pair { D is calculated by adopting a nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,Nl}; then, based on the mapping { Nh,NlAnd solving a projection matrix P by using a ridge regression method.
8. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 7, wherein said computing HR and LR dictionary pair { D ] by nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,NlThe concrete steps are as follows:
let S be ═ Sh,Sl]For training samples, { Dh,DlThe atom is calculated by Euclidean distance
Figure FDA0002582207860000031
In LR dictionary DlK nearest neighbor domain block sets N in (1)l,q
Figure FDA0002582207860000032
In the formula,
Figure FDA0002582207860000041
represents the LR training sample SlThe p-th training sample of (1),
Figure FDA0002582207860000042
representation of LR dictionary DlThe qth atom of (1);
from equation (3), K LR neighboring image blocks N corresponding to each LR dictionary atom can be calculatedl,qAccording to the positions of the K adjacent image blocks, obtaining corresponding K HR adjacent image blocks N from HR training samplesh,qWhen traversing the entire LR dictionary DlObtaining the mapping relation N formed by combining the nearest neighbor LR image blocks of all atomslAnd a nearest neighbor mapping relation N formed by combining the nearest neighbor HR image blocks corresponding to the nearest neighbor HR image blockshFinally obtain { Nh,Nl}。
9. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 7, wherein said mapping relation { N }h,NlSolving the projection matrix P by using a ridge regression method specifically comprises the following steps:
using LR nearest neighbor mapping relation NlReplacing dictionary DlThen the expression for the reconstruction of beta is
Figure FDA0002582207860000043
Figure FDA0002582207860000044
In the formula, beta represents a coefficient matrix, Y represents a low-resolution image to be reconstructed, and eta is a weight coefficient and is used for relieving the singularity problem and ensuring the stability of coefficient decomposition;
solving the formula (4) by using a ridge regression method, and calculating the coefficient
Figure FDA0002582207860000045
Is shown as
Figure FDA0002582207860000046
The reconstructed HR image X passes through the mapping relation NhSum coefficient
Figure FDA0002582207860000047
The method comprises the steps of (1) obtaining,
Figure FDA0002582207860000048
note N in the formula (6)h(Nl TNl+ηI)-1Nl TIs a projection matrix P.
10. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 7, wherein said image reconstruction based on projection matrix to obtain high resolution image specifically comprises:
firstly, preprocessing an image Y to be reconstructed to obtain an LR test characteristic image Yt={y1,y2,…,yi,…,yn}; then, the Euclidean distance is adopted to be in LR dictionary DlFinding LR test characteristic image block y iniCorresponding nearest neighbor dictionary atom dkThe expression is
Figure FDA0002582207860000051
Then through atom dkFind its corresponding projection matrix PtThen using the expression xi=PtyiObtain a correspondence yiHR image block xiAll LR test characteristic image blocks are reconstructed in this way, and an HR image X is formed.
CN202010670836.8A 2020-07-13 2020-07-13 Image super-resolution method based on SAE dictionary learning and neighborhood regression Active CN111833252B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010670836.8A CN111833252B (en) 2020-07-13 2020-07-13 Image super-resolution method based on SAE dictionary learning and neighborhood regression

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010670836.8A CN111833252B (en) 2020-07-13 2020-07-13 Image super-resolution method based on SAE dictionary learning and neighborhood regression

Publications (2)

Publication Number Publication Date
CN111833252A true CN111833252A (en) 2020-10-27
CN111833252B CN111833252B (en) 2022-06-14

Family

ID=72922812

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010670836.8A Active CN111833252B (en) 2020-07-13 2020-07-13 Image super-resolution method based on SAE dictionary learning and neighborhood regression

Country Status (1)

Country Link
CN (1) CN111833252B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109146785A (en) * 2018-08-02 2019-01-04 华侨大学 A kind of image super-resolution method based on the sparse autocoder of improvement
CN109712074A (en) * 2018-12-20 2019-05-03 黑龙江大学 The remote sensing images super-resolution reconstruction method of two-parameter beta combine processes dictionary
CN110097503A (en) * 2019-04-12 2019-08-06 浙江师范大学 The super-resolution method returned based on neighborhood
EP3591608A1 (en) * 2018-07-06 2020-01-08 Tata Consultancy Services Limited Method and system for image reconstruction using deep dictionary learning (ddl)

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3591608A1 (en) * 2018-07-06 2020-01-08 Tata Consultancy Services Limited Method and system for image reconstruction using deep dictionary learning (ddl)
CN109146785A (en) * 2018-08-02 2019-01-04 华侨大学 A kind of image super-resolution method based on the sparse autocoder of improvement
CN109712074A (en) * 2018-12-20 2019-05-03 黑龙江大学 The remote sensing images super-resolution reconstruction method of two-parameter beta combine processes dictionary
CN110097503A (en) * 2019-04-12 2019-08-06 浙江师范大学 The super-resolution method returned based on neighborhood

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
詹曙等: "基于耦合特征空间下改进字典学习的图像超分辨率重建", 《电子学报》 *
黄德天等: "正则化技术和低秩矩阵在稀疏表示超分辨率算法中的应用", 《计算机辅助设计与图形学学报》 *

Also Published As

Publication number Publication date
CN111833252B (en) 2022-06-14

Similar Documents

Publication Publication Date Title
Deng et al. Wavelet domain style transfer for an effective perception-distortion tradeoff in single image super-resolution
CN110443768B (en) Single-frame image super-resolution reconstruction method based on multiple consistency constraints
CN103093445B (en) Unified feature space image super-resolution reconstruction method based on joint sparse constraint
CN113298147B (en) Image fusion method and device based on regional energy and intuitionistic fuzzy set
CN105631807A (en) Single-frame image super resolution reconstruction method based on sparse domain selection
CN107590779A (en) Image denoising and deblurring method based on image block clustering dictionary training
CN108830791B (en) Image super-resolution method based on self sample and sparse representation
CN112150354A (en) Single image super-resolution method combining contour enhancement and denoising statistical prior
CN111861886B (en) Image super-resolution reconstruction method based on multi-scale feedback network
CN114820352A (en) Hyperspectral image denoising method and device and storage medium
CN114648048B (en) Electrocardiosignal noise reduction method based on variational self-coding and PixelCNN model
CN105590296B (en) A kind of single-frame images Super-Resolution method based on doubledictionary study
CN101739670B (en) Non-local mean space domain time varying image filtering method
CN115578262A (en) Polarization image super-resolution reconstruction method based on AFAN model
CN115526779A (en) Infrared image super-resolution reconstruction method based on dynamic attention mechanism
Roy et al. Recent study on image denoising using deep cnn techniques
CN106296583B (en) Based on image block group sparse coding and the noisy high spectrum image ultra-resolution ratio reconstructing method that in pairs maps
Tomosada et al. GAN-based image deblurring using DCT discriminator
CN111833252B (en) Image super-resolution method based on SAE dictionary learning and neighborhood regression
Hua et al. An Efficient Multiscale Spatial Rearrangement MLP Architecture for Image Restoration
CN112767273A (en) Low-dose CT image restoration method and system applying feature decoupling
CN116612009A (en) Multi-scale connection generation countermeasure network medical image super-resolution reconstruction method
CN111951181A (en) Hyperspectral image denoising method based on non-local similarity and weighted truncation kernel norm
CN108846797B (en) Image super-resolution method based on two training sets
CN116071237A (en) Video hyperspectral imaging method, system and medium based on filter sampling fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant