CN111833252A - Image Super-Resolution Method Based on SAE Dictionary Learning and Neighborhood Regression - Google Patents

Image Super-Resolution Method Based on SAE Dictionary Learning and Neighborhood Regression Download PDF

Info

Publication number
CN111833252A
CN111833252A CN202010670836.8A CN202010670836A CN111833252A CN 111833252 A CN111833252 A CN 111833252A CN 202010670836 A CN202010670836 A CN 202010670836A CN 111833252 A CN111833252 A CN 111833252A
Authority
CN
China
Prior art keywords
dictionary
image
sae
input data
dictionary learning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010670836.8A
Other languages
Chinese (zh)
Other versions
CN111833252B (en
Inventor
黄炜钦
郭一晶
陈俊仁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen University Tan Kah Kee College
Original Assignee
Xiamen University Tan Kah Kee College
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen University Tan Kah Kee College filed Critical Xiamen University Tan Kah Kee College
Priority to CN202010670836.8A priority Critical patent/CN111833252B/en
Publication of CN111833252A publication Critical patent/CN111833252A/en
Application granted granted Critical
Publication of CN111833252B publication Critical patent/CN111833252B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2135Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/28Determining representative reference patterns, e.g. by averaging or distorting; Generating dictionaries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

本发明涉及基于SAE字典学习和邻域回归的图像超分辨率方法,首先针对字典学习模型SAE准备输入数据,并进行字典的构造与训练;然后结合邻域回归理论和字典求解投影矩阵;最后基于投影矩阵进行图像重建,获得高分辨率图像。本发明一方面提高字典的特征表达能力,减小重建结果对字典的依赖性;另一方面融入邻域回归理论,提高重建速度。

Figure 202010670836

The present invention relates to an image super-resolution method based on SAE dictionary learning and neighborhood regression. First, input data is prepared for the dictionary learning model SAE, and a dictionary is constructed and trained; then, the projection matrix is solved by combining the neighborhood regression theory and the dictionary; finally, the image is reconstructed based on the projection matrix to obtain a high-resolution image. On the one hand, the present invention improves the feature expression ability of the dictionary and reduces the dependence of the reconstruction result on the dictionary; on the other hand, the neighborhood regression theory is incorporated to improve the reconstruction speed.

Figure 202010670836

Description

基于SAE字典学习和邻域回归的图像超分辨率方法Image Super-Resolution Method Based on SAE Dictionary Learning and Neighborhood Regression

技术领域technical field

本发明涉及图像超分辨率方法设计领域,特别是一种基于SAE字典学习和邻域回归的图像超分辨率方法。The invention relates to the field of image super-resolution method design, in particular to an image super-resolution method based on SAE dictionary learning and neighborhood regression.

背景技术Background technique

在现实中,由于图像采集设备的限制、场景变化以及光源等因素,往往不能得到高质量的图像,当图像的分辨率较低就无法满足实际应用的要求。图像超分辨率(Super-Resolution,SR)方法利用图像信号处理技术将已有的单幅或多幅低分辨率(LowResolution,LR)图像重建成高分辨率(High Resolution,HR)图像,其关键在于在重建过程中加入一定的附加信息来弥补图像降质过程中损失的细节信息。由于SR重建能够突破成像器件自身固有分辨率的限制实现图像分辨率的提升,因此其在遥感、医疗、视频监控等领域都具有重要的应用价值。In reality, due to factors such as the limitation of image acquisition equipment, scene changes, and light sources, high-quality images are often not obtained. When the resolution of the image is low, it cannot meet the requirements of practical applications. Image Super-Resolution (SR) method uses image signal processing technology to reconstruct existing single or multiple low-resolution (LowResolution, LR) images into high-resolution (High Resolution, HR) images. The key It is to add some additional information in the reconstruction process to make up for the detailed information lost in the process of image degradation. Since SR reconstruction can break through the limitation of the inherent resolution of the imaging device and improve the image resolution, it has important application value in remote sensing, medical treatment, video surveillance and other fields.

目前,SR方法主要分为3种类型:基于插值、基于重建和基于学习的SR方法。其中,基于学习的SR方法是近年来的热点方向,而基于字典学习的SR方法是基于学习的SR方法中最为流行的一种,最早由Yang等提出。该算法基于压缩感知理论而提出,采用字典联合学习法学习HR和LR字典对,先利用LR图像块和LR字典求得LR稀疏系数,然后基于HR和LR图像块具有相同的稀疏表示系数的假设,利用LR稀疏系数和HR字典重建HR图像块。该算法能够获取到充分的先验知识,具有较好的主观视觉效果,但是块效应较为明显,而且重建的效果对学习的字典具有很大的依赖性,而且重建耗时较长。Currently, SR methods are mainly divided into three types: interpolation-based, reconstruction-based, and learning-based SR methods. Among them, the learning-based SR method is a hot topic in recent years, and the dictionary learning-based SR method is the most popular one among the learning-based SR methods, which was first proposed by Yang et al. The algorithm is based on the compressed sensing theory. It adopts the dictionary joint learning method to learn the HR and LR dictionary pairs. First, the LR image block and the LR dictionary are used to obtain the LR sparse coefficient, and then based on the assumption that the HR and LR image blocks have the same sparse representation coefficient , using LR sparse coefficients and HR dictionary to reconstruct HR image patches. The algorithm can obtain sufficient prior knowledge and has good subjective visual effect, but the block effect is obvious, and the reconstruction effect has a great dependence on the learned dictionary, and the reconstruction takes a long time.

发明内容SUMMARY OF THE INVENTION

有鉴于此,本发明的目的是提出一种基于SAE字典学习和邻域回归的图像超分辨率方法,一方面提高字典的特征表达能力,减小重建结果对字典的依赖性;另一方面融入邻域回归理论,提高重建速度。In view of this, the purpose of the present invention is to propose an image super-resolution method based on SAE dictionary learning and neighborhood regression. Neighborhood regression theory to improve reconstruction speed.

本发明采用以下方案实现:一种基于SAE字典学习和邻域回归的图像超分辨率方法,具体包括以下步骤:The present invention adopts the following scheme to realize: an image super-resolution method based on SAE dictionary learning and neighborhood regression, which specifically includes the following steps:

针对字典学习模型SAE准备输入数据,并进行字典的构造与训练;Prepare input data for the dictionary learning model SAE, and construct and train the dictionary;

结合邻域回归理论和字典求解投影矩阵;Combine neighborhood regression theory and dictionary to solve projection matrix;

基于投影矩阵进行图像重建,获得高分辨率图像。Image reconstruction is performed based on the projection matrix to obtain high-resolution images.

进一步地,所述针对字典学习模型SAE准备输入数据具体为:Further, the described preparation input data for the dictionary learning model SAE is specifically:

将HR图像样本Ih进行下采样,得到LR图像Il,并将LR图像上采样得到中间图像Im The HR image sample Ih is down-sampled to obtain the LR image Il, and the LR image is upsampled to obtain the intermediate image Im ;

HR输入数据准备:将HR图像Ih与中间图像Im相减,得到差值图像Id,对差值图像Id进行分块与归一化,作为HR输入数据,记为

Figure BDA0002582207870000021
m为 HR输入数据的样本数;HR input data preparation: subtract the HR image I h from the intermediate image Im to obtain the difference image I d , and block and normalize the difference image I d as HR input data, denoted as
Figure BDA0002582207870000021
m is the number of samples of HR input data;

LR输入数据准备:对中间图像Im进行滤波,得到滤波图像,再对滤波图像进行归一化和图像分块,并对滤波图像块进行降维,记为

Figure BDA0002582207870000022
n为LR 输入数据的样本数;LR input data preparation: filter the intermediate image Im to obtain a filtered image, then normalize and block the filtered image, and reduce the dimension of the filtered image block, denoted as
Figure BDA0002582207870000022
n is the number of samples of LR input data;

将字典学习模型SAE的输入数据表示为S=[Sh,Sl]。Denote the input data of the dictionary learning model SAE as S=[ Sh , S l ].

进一步地,在LR输入数据准备时,采用主成分分析法PCA对滤波图像块进行降维。Further, when the LR input data is prepared, the principal component analysis method PCA is used to reduce the dimension of the filtered image block.

进一步地,所述进行字典的构造与训练具体为:Further, the construction and training of the dictionary are specifically:

结合字典学习的需要,在SAE的代价函数中,采用平均绝对值误差代替均方误差,得到改进SAE字典学习模型;Combined with the needs of dictionary learning, in the cost function of SAE, the mean absolute value error is used to replace the mean square error, and an improved SAE dictionary learning model is obtained;

输入S=[Sh,Sl],采用改进SAE字典学习模型进行学习,获得输入层与隐含层之间的权重W1,将权重转化为HR和LR字典对{Dh,Dl}。Input S=[S h , S l ], use the improved SAE dictionary learning model for learning, obtain the weight W 1 between the input layer and the hidden layer, and convert the weight into HR and LR dictionary pairs {D h , D l } .

进一步地,所述结合字典学习的需要,在SAE的代价函数中,采用平均绝对值误差代替均方误差,得到改进SAE字典学习模型具体为:Further, in combination with the needs of dictionary learning, in the cost function of SAE, the mean absolute value error is used to replace the mean square error, and the improved SAE dictionary learning model is specifically:

设si∈S为输入数据,oi∈O为输出数据,改进的SAE字典模型如下:Let s i ∈ S be the input data and o i ∈ O be the output data. The improved SAE dictionary model is as follows:

Figure BDA0002582207870000081
Figure BDA0002582207870000081

其中,第一项JMAE(θ)为重构误差项,此处采用平均绝对误差表示,m和n分别表示HR和LR输入数据的样本数;第二项Jweight(θ)为权重衰减项,用于减少权值的量级,防止过拟合,λ为该项的调节参数,

Figure BDA0002582207870000032
表示第l-1层节点i与第l层节点j 的连接权重,l表示网络的层数,Nl表示第l层的节点数,Nl+1第l+1层的节点数;第三项Jsparse(θ)为隐含层稀疏正则项,
Figure BDA0002582207870000033
为隐含层神经元的平均激活量,ρ为设定好的预期激活量,γ为调节参数,N2表示第2层的节点数;其中,
Figure BDA0002582207870000034
采用式(2)表示:Among them, the first term J MAE (θ) is the reconstruction error term, which is represented by the mean absolute error here, m and n represent the number of samples of HR and LR input data respectively; the second term J weight (θ) is the weight decay term , used to reduce the magnitude of the weights and prevent overfitting, λ is the adjustment parameter of this item,
Figure BDA0002582207870000032
Represents the connection weight between the l-1 layer node i and the l layer node j, l represents the number of layers of the network, N l represents the number of nodes in the l layer, N l+1 The number of nodes in the l+1 layer; the third The term J sparse (θ) is the hidden layer sparse regular term,
Figure BDA0002582207870000033
is the average activation of neurons in the hidden layer, ρ is the set expected activation, γ is the adjustment parameter, and N 2 represents the number of nodes in the second layer; among them,
Figure BDA0002582207870000034
It is expressed by formula (2):

Figure BDA0002582207870000035
Figure BDA0002582207870000035

进一步地,所述输入S=[Sh,Sl],采用改进SAE字典学习模型进行学习,获得输入层与隐含层之间的权重W1,将权重转化为HR和LR字典对{Dh,Dl}具体为:Further, the input S=[S h , S l ], the improved SAE dictionary learning model is used for learning, the weight W 1 between the input layer and the hidden layer is obtained, and the weight is converted into the HR and LR dictionary pair {D h , D l } are specifically:

在SAE字典学习模型的训练过程中,结合梯度下降法进行参数更新,最终获得输入层到隐含层的连接权重W1,其中W1={wi},i=1,2,...,m+n;根据网络权重与字典的关系,字典D等价于输入层与隐含层的链接权重W1,具体表示为HR字典 Dh={w1,w2,…,wm}、LR字典Dl={wm+1,wm+2,…,wm+n},字典对表示为D=(Dh,Dl),其中wi∈W1,且wi={w1,i,w2,i,...,wk,i},k为字典的维数,wk,i表示第k维第i个字典原子的权值。In the training process of the SAE dictionary learning model, the parameters are updated with the gradient descent method, and finally the connection weight W 1 from the input layer to the hidden layer is obtained, where W 1 ={ wi }, i=1,2,... ,m+n; according to the relationship between the network weight and the dictionary, the dictionary D is equivalent to the link weight W 1 of the input layer and the hidden layer, specifically expressed as HR dictionary D h ={w 1 ,w 2 ,...,w m } , LR dictionary D l ={w m+1 ,w m+2 ,...,w m+n }, the dictionary pair is expressed as D=(D h , D l ), where w i ∈ W 1 , and w i = {w 1,i ,w 2,i ,...,w k,i }, k is the dimension of the dictionary, w k,i is the weight of the i-th dictionary atom in the k-th dimension.

进一步地,所述结合邻域回归理论和字典求解投影矩阵具体为:首先,采用最近邻域方法计算HR和LR字典对{Dh,Dl}中每个原子在样本空间S=[Sh,Sl]的最近邻域映射关系{Nh,Nl};然后,基于映射关系{Nh,Nl},采用岭回归方法求解投影矩阵P。Further, the combination of the neighborhood regression theory and the dictionary to solve the projection matrix is specifically: first, adopt the nearest neighbor method to calculate the HR and LR dictionary pairs {D h , D l } in the sample space S=[S h for each atom in the , S l ] the nearest neighbor mapping relationship {N h , N l }; then, based on the mapping relationship {N h , N l }, the projection matrix P is solved by the ridge regression method.

进一步地,所述采用最近邻域方法计算HR和LR字典对{Dh,Dl}中每个原子在样本空间S=[Sh,Sl]的最近邻域映射关系{Nh,Nl}具体为:Further, the nearest neighbor method is used to calculate the nearest neighbor mapping relationship {N h ,N of each atom in the sample space S=[ Sh ,S l ] in the HR and LR dictionary pair {D h , D l } l }Specifically:

设S=[Sh,Sl]为训练样本,{Dh,Dl}为字典对,采用欧式距离计算原子

Figure BDA0002582207870000041
在LR 字典Dl中的K个最近邻域块集合Nl,q:Let S=[S h , S l ] be the training sample, {D h , D l } be the dictionary pair, and use the Euclidean distance to calculate the atoms
Figure BDA0002582207870000041
The set of K nearest neighbor blocks N l,q in the LR dictionary D l :

Figure BDA0002582207870000042
Figure BDA0002582207870000042

式中,

Figure BDA0002582207870000043
表示LR训练样本Sl中第p个训练样本,
Figure BDA0002582207870000044
表示LR字典的第q个原子;In the formula,
Figure BDA0002582207870000043
represents the p-th training sample in the LR training sample S l ,
Figure BDA0002582207870000044
Represents the qth atom of the LR dictionary;

从式(3)可计算出每一个LR字典原子对应的K个LR近邻图像块Nl,q,根据K个近邻图像块的位置,可从HR训练样本中获得相应的K个HR近邻图像块Nh,q,当遍历整个LR字典Dl的所有原子,可得所有原子的最近邻LR图像块组合而成的映射关系Nl,及其对应的最近邻HR图像块组合而成的最近邻映射关系Nh,最终获得{Nh,Nl}。From equation (3), the K LR neighbor image blocks N l,q corresponding to each LR dictionary atom can be calculated. According to the positions of the K neighbor image blocks, the corresponding K HR neighbor image blocks can be obtained from the HR training samples. N h,q , when traversing all atoms of the entire LR dictionary D l , the mapping relationship N l composed of the nearest neighbor LR image blocks of all atoms can be obtained, and the nearest neighbor composed of the corresponding nearest neighbor HR image blocks. The mapping relation N h is finally obtained {N h , N l }.

进一步地,所述基于映射关系{Nh,Nl},采用岭回归方法求解投影矩阵P具体为:Further, based on the mapping relationship {N h , N l }, using the ridge regression method to solve the projection matrix P is specifically:

用LR最近邻映射关系Nl取代字典Dl,则重建β的表达式为

Figure BDA0002582207870000051
Replacing the dictionary D l with the LR nearest neighbor mapping relation N l , the expression for reconstructing β is
Figure BDA0002582207870000051

Figure BDA0002582207870000052
Figure BDA0002582207870000052

式中,β表示系数矩阵,Y表示待重建低分辨率图像,η是权重系数,用于缓解奇异性问题并保证系数分解的稳定性;In the formula, β represents the coefficient matrix, Y represents the low-resolution image to be reconstructed, and η is the weight coefficient, which is used to alleviate the singularity problem and ensure the stability of the coefficient decomposition;

采用岭回归方法求解式(4),将系数

Figure BDA0002582207870000053
表示为The ridge regression method is used to solve equation (4), and the coefficients
Figure BDA0002582207870000053
Expressed as

Figure BDA0002582207870000054
Figure BDA0002582207870000054

重建的HR图像X通过映射关系Nh和系数

Figure BDA0002582207870000055
获得,The reconstructed HR image X is mapped by N h and coefficients
Figure BDA0002582207870000055
get,

Figure BDA0002582207870000056
Figure BDA0002582207870000056

记式(6)中的

Figure BDA0002582207870000057
为投影矩阵P。In Equation (6)
Figure BDA0002582207870000057
is the projection matrix P.

进一步地,所述基于投影矩阵进行图像重建,获得高分辨率图像具体为:Further, performing image reconstruction based on the projection matrix to obtain a high-resolution image is specifically:

首先,对待重建的图像Y进行预处理,得到LR测试特征图像 Yt={y1,y2,…,yi,…,yn};然后,采用欧式距离在LR字典Dl中寻找LR测试特征图像块yi对应的最近邻字典原子dk,表达式为

Figure BDA0002582207870000058
接着通过原子dk找到其对应的投影矩阵Pt,然后利用表达式xi=Ptyi,获得对应yi的HR图像块xi,以此重建所有的LR测试特征图像块,并组成HR图像X。First, preprocess the image Y to be reconstructed to obtain the LR test feature image Y t ={y 1 ,y 2 ,...,y i ,...,y n }; then, use Euclidean distance to find LR in the LR dictionary D l The nearest neighbor dictionary atom d k corresponding to the test feature image block y i is expressed as
Figure BDA0002582207870000058
Then find its corresponding projection matrix P t through atom d k , and then use the expression xi =P t y i to obtain the HR image block xi corresponding to yi , so as to reconstruct all LR test feature image blocks, and form HR image X.

与现有技术相比,本发明有以下有益效果:本发明首先对于字典的学习采用改进稀疏自动编码器,充分利用其突出的特征学习能力,增强字典的特征表达能力,提高图像的重建质量;然后,本发明在基于字典学习的超分辨率框架中融入邻域回归理论,避免原来框架中的稀疏编码过程,减少计算量,提高重建的速度。Compared with the prior art, the present invention has the following beneficial effects: the present invention firstly adopts an improved sparse auto-encoder for the learning of the dictionary, makes full use of its outstanding feature learning ability, enhances the feature expression ability of the dictionary, and improves the reconstruction quality of the image; Then, the present invention integrates the neighborhood regression theory into the super-resolution framework based on dictionary learning, avoids the sparse coding process in the original framework, reduces the amount of computation, and improves the speed of reconstruction.

附图说明Description of drawings

图1为本发明实施例的SAE输入数据预处理流程。FIG. 1 is an SAE input data preprocessing process according to an embodiment of the present invention.

图2为本发明实施例的方法流程示意图。FIG. 2 is a schematic flowchart of a method according to an embodiment of the present invention.

具体实施方式Detailed ways

下面结合附图及实施例对本发明做进一步说明。The present invention will be further described below with reference to the accompanying drawings and embodiments.

应该指出,以下详细说明都是示例性的,旨在对本申请提供进一步的说明。除非另有指明,本文使用的所有技术和科学术语具有与本申请所属技术领域的普通技术人员通常理解的相同含义。It should be noted that the following detailed description is exemplary and intended to provide further explanation of the application. Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.

需要注意的是,这里所使用的术语仅是为了描述具体实施方式,而非意图限制根据本申请的示例性实施方式。如在这里所使用的,除非上下文另外明确指出,否则单数形式也意图包括复数形式,此外,还应当理解的是,当在本说明书中使用术语“包含”和/或“包括”时,其指明存在特征、步骤、操作、器件、组件和/ 或它们的组合。It should be noted that the terminology used herein is for the purpose of describing specific embodiments only, and is not intended to limit the exemplary embodiments according to the present application. As used herein, unless the context clearly dictates otherwise, the singular is intended to include the plural as well, furthermore, it is to be understood that when the terms "comprising" and/or "including" are used in this specification, it indicates that There are features, steps, operations, devices, components and/or combinations thereof.

如图2所示,本实施例提供了一种基于SAE字典学习和邻域回归的图像超分辨率方法,具体包括以下步骤:As shown in Figure 2, this embodiment provides an image super-resolution method based on SAE dictionary learning and neighborhood regression, which specifically includes the following steps:

步骤S1:针对字典学习模型SAE准备输入数据,并进行字典的构造与训练;Step S1: prepare input data for the dictionary learning model SAE, and construct and train the dictionary;

步骤S2:结合邻域回归理论和字典求解投影矩阵;Step S2: combine the neighborhood regression theory and the dictionary to solve the projection matrix;

步骤S3:基于投影矩阵进行图像重建,获得高分辨率图像。Step S3: Perform image reconstruction based on the projection matrix to obtain a high-resolution image.

在本实施例中,如图1所示,步骤S1中所述针对字典学习模型SAE准备输入数据具体为:In this embodiment, as shown in FIG. 1 , the preparation of input data for the dictionary learning model SAE described in step S1 is specifically:

步骤S11:将HR图像样本Ih进行下采样,得到LR图像Il,并将LR图像上采样得到中间图像ImStep S11: down-sampling the HR image sample Ih to obtain the LR image Il , and up-sampling the LR image to obtain the intermediate image Im ;

步骤S12:HR输入数据准备:将HR图像Ih与中间图像Im相减,得到差值图像Id,对差值图像Id进行分块与归一化,作为HR输入数据,记为

Figure BDA0002582207870000071
m为HR输入数据的样本数;Step S12: HR input data preparation: subtract the HR image I h and the intermediate image I m to obtain a difference image I d , and block and normalize the difference image I d as HR input data, denoted as
Figure BDA0002582207870000071
m is the number of samples of HR input data;

步骤S13:LR输入数据准备:对中间图像Im进行滤波,得到滤波图像,再对滤波图像进行归一化和图像分块,并对滤波图像块进行降维,记为

Figure BDA0002582207870000072
n为LR输入数据的样本数;Step S13: LR input data preparation: filter the intermediate image Im to obtain a filtered image, then normalize and block the filtered image, and reduce the dimension of the filtered image block, denoted as
Figure BDA0002582207870000072
n is the number of samples of LR input data;

步骤S14:将字典学习模型SAE的输入数据表示为S=[Sh,Sl]。Step S14: Denote the input data of the dictionary learning model SAE as S=[S h , S l ].

其中,在LR输入数据准备时,采用主成分分析法PCA对滤波图像块进行降维。Among them, when the LR input data is prepared, the principal component analysis method PCA is used to reduce the dimension of the filtered image block.

在本实施例中,步骤S1中的所述进行字典的构造与训练具体为:In this embodiment, the construction and training of the dictionary in step S1 is specifically:

步骤S15:结合字典学习的需要,在SAE的代价函数中,采用平均绝对值误差代替均方误差,得到改进SAE字典学习模型;Step S15: Combined with the needs of dictionary learning, in the cost function of SAE, the mean absolute value error is used instead of the mean square error to obtain an improved SAE dictionary learning model;

设si∈S为输入数据,oi∈O为输出数据,改进的SAE字典模型如下:Let s i ∈ S be the input data and o i ∈ O be the output data. The improved SAE dictionary model is as follows:

Figure BDA0002582207870000081
Figure BDA0002582207870000081

其中,第一项JMAE(θ)为重构误差项,此处采用平均绝对误差(Mean SquaredError, MSE)表示,m和n分别表示HR和LR输入数据的样本数;第二项Jweight(θ)为权重衰减项,用于减少权值的量级,防止过拟合,λ为该项的调节参数,

Figure BDA0002582207870000082
表示第l-1 层节点i与第l层节点j的连接权重,l表示网络的层数,Nl表示第l层的节点数, Nl+1第l+1层的节点数;第三项Jsparse(θ)为隐含层稀疏正则项,
Figure BDA0002582207870000083
为隐含层神经元的平均激活量,ρ为设定好的预期激活量,其值接近于0,γ为该项的调节参数, N2表示第2层的节点数;对于
Figure BDA0002582207870000084
显著偏离ρ的情况,一般采用相对熵来惩罚,如式(2)所示:Among them, the first item J MAE (θ) is the reconstruction error term, which is represented by the mean absolute error (Mean SquaredError, MSE) here, m and n represent the number of samples of HR and LR input data respectively; the second item J weight ( θ) is the weight decay term, which is used to reduce the magnitude of the weight value and prevent over-fitting, and λ is the adjustment parameter of this term,
Figure BDA0002582207870000082
Represents the connection weight between the l-1 layer node i and the l layer node j, l represents the number of layers of the network, N l represents the number of nodes in the l layer, N l+1 The number of nodes in the l+1 layer; the third The term J sparse (θ) is the hidden layer sparse regular term,
Figure BDA0002582207870000083
is the average activation of neurons in the hidden layer, ρ is the set expected activation, its value is close to 0, γ is the adjustment parameter of this item, N 2 represents the number of nodes in the second layer; for
Figure BDA0002582207870000084
In the case of significant deviation from ρ, relative entropy is generally used to punish, as shown in formula (2):

Figure BDA0002582207870000085
Figure BDA0002582207870000085

步骤S16:输入S=[Sh,Sl],采用改进SAE字典学习模型进行学习,获得输入层与隐含层之间的权重W1,将权重转化为HR和LR字典对{Dh,Dl}。Step S16: Input S=[S h , S l ], use the improved SAE dictionary learning model for learning, obtain the weight W 1 between the input layer and the hidden layer, and convert the weight into HR and LR dictionary pairs {D h , D l }.

在SAE字典学习模型的训练过程中,结合梯度下降法进行参数更新,最终获得输入层到隐含层的连接权重W1,其中W1={wi},i=1,2,...,m+n;在字典学习中,可以通过字典矩阵和稀疏表示输入数据,而在SAE中,可以通过隐含层的表示与学习权重表示输入数据,根据二者之间的联系可知字典D等价于输入层与隐含层的链接权重W1,具体表示为HR字典Dh={w1,w2,…,wm}、LR字典 Dl={wm+1,wm+2,…,wm+n},字典对表示为D=(Dh,Dl),其中wi∈W1,且wi={w1,i,w2,i,...,wk,i},k为字典的维数,wk,i表示第k维第i个字典原子的权值。In the training process of the SAE dictionary learning model, the parameters are updated with the gradient descent method, and finally the connection weight W 1 from the input layer to the hidden layer is obtained, where W 1 ={ wi }, i=1,2,... ,m+n; in dictionary learning, the input data can be represented by dictionary matrix and sparse representation, while in SAE, the input data can be represented by the representation of the hidden layer and the learning weight, according to the relationship between the two, the dictionary D, etc. It is equivalent to the link weight W 1 between the input layer and the hidden layer, specifically expressed as HR dictionary D h ={w 1 ,w 2 ,...,w m }, LR dictionary D l ={w m+1 ,w m+2 , . _ _ _ _ _ _ k,i }, k is the dimension of the dictionary, w k,i represents the weight of the i-th dictionary atom in the k-th dimension.

在本实施例中,步骤S2具体为:In this embodiment, step S2 is specifically:

步骤S21:采用最近邻域方法计算HR和LR字典对{Dh,Dl}中每个原子在样本空间S=[Sh,Sl]的最近邻域映射关系{Nh,Nl};即,Step S21: Use the nearest neighbor method to calculate the nearest neighbor mapping relationship {N h ,N l } of each atom in the HR and LR dictionary pair {D h ,D l } in the sample space S=[ Sh ,S l ] ;which is,

设S=[Sh,Sl]为训练样本,{Dh,Dl}为字典对,采用欧式距离计算原子

Figure BDA0002582207870000091
在LR 字典Dl中的K个最近邻域块集合Nl,q:Let S=[S h , S l ] be the training sample, {D h , D l } be the dictionary pair, and use the Euclidean distance to calculate the atoms
Figure BDA0002582207870000091
The set of K nearest neighbor blocks N l,q in the LR dictionary D l :

Figure BDA0002582207870000092
Figure BDA0002582207870000092

式中,

Figure BDA0002582207870000093
表示LR训练样本Sl中第p个训练样本,
Figure BDA0002582207870000094
表示LR字典的第q个原子;In the formula,
Figure BDA0002582207870000093
represents the p-th training sample in the LR training sample S l ,
Figure BDA0002582207870000094
Represents the qth atom of the LR dictionary;

从式(3)可计算出每一个LR字典原子对应的K个LR近邻图像块Nl,q,根据K个近邻图像块的位置,可从HR训练样本中获得相应的K个HR近邻图像块Nh,q,当遍历整个LR字典Dl的所有原子,可得所有原子的最近邻LR图像块组合而成的映射关系Nl,及其对应的最近邻HR图像块组合而成的最近邻映射关系Nh,最终获得{Nh,Nl}。From equation (3), the K LR neighbor image blocks N l,q corresponding to each LR dictionary atom can be calculated. According to the positions of the K neighbor image blocks, the corresponding K HR neighbor image blocks can be obtained from the HR training samples. N h,q , when traversing all atoms of the entire LR dictionary D l , the mapping relationship N l composed of the nearest neighbor LR image blocks of all atoms can be obtained, and the nearest neighbor composed of the corresponding nearest neighbor HR image blocks. The mapping relation N h is finally obtained {N h , N l }.

步骤S22:基于映射关系{Nh,Nl},采用岭回归方法求解投影矩阵P,即,Step S22: Based on the mapping relationship {N h , N l }, use the ridge regression method to solve the projection matrix P, that is,

用LR最近邻映射关系Nl取代字典Dl,则重建β的表达式为

Figure BDA0002582207870000095
Replacing the dictionary D l with the LR nearest neighbor mapping relation N l , the expression for reconstructing β is
Figure BDA0002582207870000095

Figure BDA0002582207870000096
Figure BDA0002582207870000096

式中,β表示系数矩阵,Y表示待重建低分辨率图像,η是权重系数,用于缓解奇异性问题并保证系数分解的稳定性;In the formula, β represents the coefficient matrix, Y represents the low-resolution image to be reconstructed, and η is the weight coefficient, which is used to alleviate the singularity problem and ensure the stability of the coefficient decomposition;

采用岭回归方法求解式(4),将系数

Figure BDA0002582207870000097
表示为The ridge regression method is used to solve equation (4), and the coefficients
Figure BDA0002582207870000097
Expressed as

Figure BDA0002582207870000098
Figure BDA0002582207870000098

重建的HR图像X通过映射关系Nh和系数

Figure BDA0002582207870000099
获得,The reconstructed HR image X is mapped by N h and coefficients
Figure BDA0002582207870000099
get,

Figure BDA0002582207870000101
Figure BDA0002582207870000101

记式(6)中的

Figure BDA0002582207870000102
为投影矩阵P。In Equation (6)
Figure BDA0002582207870000102
is the projection matrix P.

在本实施例中,步骤S3具体为:In this embodiment, step S3 is specifically:

首先,对待重建的图像Y进行预处理,得到LR测试特征图像 Yt={y1,y2,…,yi,…,yn};然后,采用欧式距离在LR字典Dl中寻找LR测试特征图像块yi对应的最近邻字典原子dk,表达式为

Figure BDA0002582207870000103
接着通过原子dk找到其对应的投影矩阵Pt,然后利用表达式xi=Ptyi,获得对应yi的HR图像块xi,以此重建所有的LR测试特征图像块,并组成HR图像X。First, preprocess the image Y to be reconstructed to obtain the LR test feature image Y t ={y 1 ,y 2 ,...,y i ,...,y n }; then, use Euclidean distance to find LR in the LR dictionary D l The nearest neighbor dictionary atom d k corresponding to the test feature image block y i is expressed as
Figure BDA0002582207870000103
Then find its corresponding projection matrix P t through atom d k , and then use the expression xi =P t y i to obtain the HR image block xi corresponding to yi , so as to reconstruct all LR test feature image blocks, and form HR image X.

接下来,本实施例采用以下仿真实验来进行进一步的说明。Next, this embodiment uses the following simulation experiments for further description.

本实施例所采用的仿真工具为MATLAB,评价指标为峰值信噪比PSNR和结构相似度SSIM,其中PSNR越大,SSIM越接近1,则超分辨率效果越好。The simulation tool used in this embodiment is MATLAB, and the evaluation indicators are peak signal-to-noise ratio (PSNR) and structural similarity (SSIM). The larger the PSNR, the closer the SSIM is to 1, the better the super-resolution effect.

仿真实验具体设置如下:The specific settings of the simulation experiment are as follows:

数据准备:为了保证实验的客观性,字典学习采用91张通用标准HR训练样本,测试图像来源于标准测试库Set5和Set14。为了定量地评价重建图像的质量,将这些测试图像作为HR参考图像,通过下采样获取待处理的LR图像。Data preparation: In order to ensure the objectivity of the experiment, dictionary learning adopts 91 general standard HR training samples, and the test images come from standard test libraries Set5 and Set14. In order to quantitatively evaluate the quality of the reconstructed images, these test images are used as HR reference images, and the LR images to be processed are obtained by downsampling.

对比算法:Bicubic、L1SR(Super Resolution with L1 Regression)、SISR(Single Image Super Resolution)、ANR(Anchored Neighborhood Regression)、SRCNN(Super Resolution using Convolution Neural Network)等5种SR算法作对比。Comparison algorithms: 5 SR algorithms, including Bicubic, L1SR (Super Resolution with L1 Regression), SISR (Single Image Super Resolution), ANR (Anchored Neighborhood Regression), and SRCNN (Super Resolution using Convolution Neural Network), are compared.

重要参数设置:采样因子为3,SAE隐含层的节点数为1024。Important parameter settings: the sampling factor is 3, and the number of nodes in the SAE hidden layer is 1024.

仿真实验主要分为2组,具体如下。The simulation experiments are mainly divided into two groups, as follows.

第1组实验:与不同SR方法比较。Group 1 experiments: Comparison with different SR methods.

表1列出了不同SR算法得到的重建图像对应的PSNR和SSIM,其中最后一栏标出的数值表示在相应的评价指标下本实施例对应算法的性能最优,10幅图像分别来自Set5和Set14。由表1可见,本实施例方法得到的PSNR和SSIM值大体上都是最优的,表明其重建效果更优。Table 1 lists the PSNR and SSIM corresponding to the reconstructed images obtained by different SR algorithms. The values marked in the last column indicate that the performance of the corresponding algorithm in this embodiment is the best under the corresponding evaluation indicators. The 10 images are from Set5 and Set14. It can be seen from Table 1 that the PSNR and SSIM values obtained by the method in this embodiment are generally optimal, indicating that the reconstruction effect is better.

表1不同SR方法的PSNR(dB)和SSIM值的比较Table 1 Comparison of PSNR (dB) and SSIM values for different SR methods

Figure BDA0002582207870000111
Figure BDA0002582207870000111

第2组实验:重建速度比较。Experiment 2: Comparison of reconstruction speed.

本组实验将不同算法在同样的设备和环境中运行,用于验证在基于字典学习的SR算法中融入邻域回归思想能提升的重建速度的效果。表2中列出了Set5和 Set14测试集合在不同SR算法下的平均重建时间。从中可以看出,本发明速度明显高于其他SR算法。In this group of experiments, different algorithms were run in the same equipment and environment to verify the effect of improving the reconstruction speed by incorporating the idea of neighborhood regression into the dictionary learning-based SR algorithm. Table 2 lists the average reconstruction time of Set5 and Set14 test sets under different SR algorithms. It can be seen from this that the speed of the present invention is significantly higher than other SR algorithms.

表2不同SR算平均重建时间(s)的比较Table 2 Comparison of the average reconstruction time (s) of different SR algorithms

测试图像库test image library L1SRL1SR SISRSISR SRCNNSRCNN 本方法this method Set 5Set 5 14.2814.28 0.960.96 2.982.98 0.330.33 Set 14Set 14 31.9331.93 1.931.93 8.038.03 0.65 0.65

本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。As will be appreciated by those skilled in the art, the embodiments of the present application may be provided as a method, a system, or a computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.

本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the present application. It will be understood that each flow and/or block in the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to the processor of a general purpose computer, special purpose computer, embedded processor or other programmable data processing device to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing device produce Means for implementing the functions specified in a flow or flow of a flowchart and/or a block or blocks of a block diagram.

这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/ 或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions The apparatus implements the functions specified in the flow or flows of the flowcharts and/or the block or blocks of the block diagrams.

这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded on a computer or other programmable data processing device to cause a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process such that The instructions provide steps for implementing the functions specified in the flow or blocks of the flowcharts and/or the block or blocks of the block diagrams.

以上所述,仅是本发明的较佳实施例而已,并非是对本发明作其它形式的限制,任何熟悉本专业的技术人员可能利用上述揭示的技术内容加以变更或改型为等同变化的等效实施例。但是凡是未脱离本发明技术方案内容,依据本发明的技术实质对以上实施例所作的任何简单修改、等同变化与改型,仍属于本发明技术方案的保护范围。The above are only preferred embodiments of the present invention, and are not intended to limit the present invention in other forms. Any person skilled in the art may use the technical content disclosed above to make changes or modifications to equivalent changes. Example. However, any simple modifications, equivalent changes and modifications made to the above embodiments according to the technical essence of the present invention without departing from the content of the technical solutions of the present invention still belong to the protection scope of the technical solutions of the present invention.

Claims (10)

1. An image super-resolution method based on SAE dictionary learning and neighborhood regression is characterized by comprising the following steps:
preparing input data aiming at a dictionary learning model SAE, and constructing and training a dictionary;
solving a projection matrix by combining a neighborhood regression theory and a dictionary;
and reconstructing the image based on the projection matrix to obtain a high-resolution image.
2. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 1, wherein the preparation of input data for the dictionary learning model SAE is specifically as follows:
HR image sample IhDownsampling to obtain LR image IlAnd up-sampling the LR image to obtain an intermediate image Im
HR input data preparation: HR image IhAnd an intermediate image ImSubtracting to obtain a difference image IdFor difference image IdPartitioning and normalizing to obtain HR input data
Figure FDA0002582207860000011
m is the number of samples of HR input data;
LR input data preparation: for intermediate image ImFiltering to obtain a filtered image, normalizing and partitioning the filtered image, and reducing dimensions of the filtered image block, which is recorded as
Figure FDA0002582207860000012
n is the number of samples of LR input data;
input data of the dictionary learning model SAE is represented as S ═ Sh,Sl]。
3. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 2, wherein in LR input data preparation, a Principal Component Analysis (PCA) method is used to reduce the dimension of the filtering image block.
4. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 1, wherein said constructing and training of the dictionary specifically comprises:
combining with the requirement of dictionary learning, in the cost function of SAE, adopting average absolute value error to replace mean square error to obtain an improved SAE dictionary learning model;
input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,Dl}。
5. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 4, wherein said combination of the requirement of dictionary learning, in the cost function of SAE, the mean absolute value error is used to replace the mean square error, and the improved SAE dictionary learning model is specifically:
let siE.g. S as input data, oiE O is taken as output data, and the improved SAE dictionary model is as follows:
Figure DEST_PATH_BDA0002582207870000081
wherein the first item JMAE(θ) is a reconstruction error term, here expressed in mean absolute error, m and n represent HR and HR, respectivelyNumber of samples of LR input data; second item Jweight(theta) is a weight decay term used to reduce the magnitude of the weights, preventing overfitting, lambda is the tuning parameter of this term,
Figure FDA0002582207860000022
represents the connection weight of the l-1 layer node i and the l layer node j, wherein l represents the layer number of the network, and NlIndicates the number of nodes of the l-th layer, Nl+1Number of nodes of layer l + 1; third item Jsparse(theta) is a hidden layer sparsity regularization term,
Figure FDA0002582207860000023
for the mean activation of the neurons in the hidden layer, ρ is the set expected activation, γ is the regulatory parameter, N2Represents the number of nodes of layer 2; wherein,
Figure FDA0002582207860000024
expressed by formula (2):
Figure FDA0002582207860000025
6. the image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 4, wherein said input S ═ Sh,Sl]Learning by adopting an improved SAE dictionary learning model to obtain the weight W between the input layer and the hidden layer1The weights are converted into HR and LR dictionary pairs { Dh,DlThe concrete steps are as follows:
in the training process of the SAE dictionary learning model, updating parameters by combining a gradient descent method, and finally obtaining the connection weight W from the input layer to the hidden layer1Wherein W is1={wi1,2, i, m + n; according to the relation between the network weight and the dictionary, the dictionary D is equivalent to the link weight W of the input layer and the hidden layer1Denoted HR dictionary Dh={w1,w2,…,wmR, LR dictionary Dl={wm+1,wm+2,…,wm+nDenoted D ═ D for dictionary pairh,Dl) Wherein w isi∈W1And w isi={w1,i,w2,i,...,wk,iK is the dimension of the dictionary, wk,iRepresenting the weight of the ith dictionary atom in the kth dimension.
7. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 1, wherein said solving of projection matrix in combination with neighborhood regression theory and dictionary specifically comprises: first, the HR and LR dictionary pair { D is calculated by adopting a nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,Nl}; then, based on the mapping { Nh,NlAnd solving a projection matrix P by using a ridge regression method.
8. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 7, wherein said computing HR and LR dictionary pair { D ] by nearest neighbor methodh,DlEach atom in the lattice space S ═ Sh,Sl]Nearest neighbor mapping of { N }h,NlThe concrete steps are as follows:
let S be ═ Sh,Sl]For training samples, { Dh,DlThe atom is calculated by Euclidean distance
Figure FDA0002582207860000031
In LR dictionary DlK nearest neighbor domain block sets N in (1)l,q
Figure FDA0002582207860000032
In the formula,
Figure FDA0002582207860000041
represents the LR training sample SlThe p-th training sample of (1),
Figure FDA0002582207860000042
representation of LR dictionary DlThe qth atom of (1);
from equation (3), K LR neighboring image blocks N corresponding to each LR dictionary atom can be calculatedl,qAccording to the positions of the K adjacent image blocks, obtaining corresponding K HR adjacent image blocks N from HR training samplesh,qWhen traversing the entire LR dictionary DlObtaining the mapping relation N formed by combining the nearest neighbor LR image blocks of all atomslAnd a nearest neighbor mapping relation N formed by combining the nearest neighbor HR image blocks corresponding to the nearest neighbor HR image blockshFinally obtain { Nh,Nl}。
9. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 7, wherein said mapping relation { N }h,NlSolving the projection matrix P by using a ridge regression method specifically comprises the following steps:
using LR nearest neighbor mapping relation NlReplacing dictionary DlThen the expression for the reconstruction of beta is
Figure FDA0002582207860000043
Figure FDA0002582207860000044
In the formula, beta represents a coefficient matrix, Y represents a low-resolution image to be reconstructed, and eta is a weight coefficient and is used for relieving the singularity problem and ensuring the stability of coefficient decomposition;
solving the formula (4) by using a ridge regression method, and calculating the coefficient
Figure FDA0002582207860000045
Is shown as
Figure FDA0002582207860000046
The reconstructed HR image X passes through the mapping relation NhSum coefficient
Figure FDA0002582207860000047
The method comprises the steps of (1) obtaining,
Figure FDA0002582207860000048
note N in the formula (6)h(Nl TNl+ηI)-1Nl TIs a projection matrix P.
10. The image super-resolution method based on SAE dictionary learning and neighborhood regression as claimed in claim 7, wherein said image reconstruction based on projection matrix to obtain high resolution image specifically comprises:
firstly, preprocessing an image Y to be reconstructed to obtain an LR test characteristic image Yt={y1,y2,…,yi,…,yn}; then, the Euclidean distance is adopted to be in LR dictionary DlFinding LR test characteristic image block y iniCorresponding nearest neighbor dictionary atom dkThe expression is
Figure FDA0002582207860000051
Then through atom dkFind its corresponding projection matrix PtThen using the expression xi=PtyiObtain a correspondence yiHR image block xiAll LR test characteristic image blocks are reconstructed in this way, and an HR image X is formed.
CN202010670836.8A 2020-07-13 2020-07-13 Image super-resolution method based on SAE dictionary learning and neighborhood regression Active CN111833252B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010670836.8A CN111833252B (en) 2020-07-13 2020-07-13 Image super-resolution method based on SAE dictionary learning and neighborhood regression

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010670836.8A CN111833252B (en) 2020-07-13 2020-07-13 Image super-resolution method based on SAE dictionary learning and neighborhood regression

Publications (2)

Publication Number Publication Date
CN111833252A true CN111833252A (en) 2020-10-27
CN111833252B CN111833252B (en) 2022-06-14

Family

ID=72922812

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010670836.8A Active CN111833252B (en) 2020-07-13 2020-07-13 Image super-resolution method based on SAE dictionary learning and neighborhood regression

Country Status (1)

Country Link
CN (1) CN111833252B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109146785A (en) * 2018-08-02 2019-01-04 华侨大学 A kind of image super-resolution method based on the sparse autocoder of improvement
CN109712074A (en) * 2018-12-20 2019-05-03 黑龙江大学 Remote sensing image super-resolution reconstruction method based on two-parameter beta process joint dictionary
CN110097503A (en) * 2019-04-12 2019-08-06 浙江师范大学 The super-resolution method returned based on neighborhood
EP3591608A1 (en) * 2018-07-06 2020-01-08 Tata Consultancy Services Limited Method and system for image reconstruction using deep dictionary learning (ddl)

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3591608A1 (en) * 2018-07-06 2020-01-08 Tata Consultancy Services Limited Method and system for image reconstruction using deep dictionary learning (ddl)
CN109146785A (en) * 2018-08-02 2019-01-04 华侨大学 A kind of image super-resolution method based on the sparse autocoder of improvement
CN109712074A (en) * 2018-12-20 2019-05-03 黑龙江大学 Remote sensing image super-resolution reconstruction method based on two-parameter beta process joint dictionary
CN110097503A (en) * 2019-04-12 2019-08-06 浙江师范大学 The super-resolution method returned based on neighborhood

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
詹曙等: "基于耦合特征空间下改进字典学习的图像超分辨率重建", 《电子学报》 *
黄德天等: "正则化技术和低秩矩阵在稀疏表示超分辨率算法中的应用", 《计算机辅助设计与图形学学报》 *

Also Published As

Publication number Publication date
CN111833252B (en) 2022-06-14

Similar Documents

Publication Publication Date Title
CN110119780B (en) A Generative Adversarial Network-Based Super-resolution Reconstruction Method for Hyperspectral Images
CN111754403B (en) An Image Super-Resolution Reconstruction Method Based on Residual Learning
CN110570353B (en) Densely connected generative adversarial network single image super-resolution reconstruction method
CN110120011B (en) A video super-resolution method based on convolutional neural network and mixed resolution
CN103093445B (en) Unified feature space image super-resolution reconstruction method based on joint sparse constraint
CN111768342A (en) A face super-resolution method based on attention mechanism and multi-level feedback supervision
CN110111256B (en) Image super-resolution reconstruction method based on residual distillation network
CN109671022B (en) A super-resolution method for image texture enhancement based on deep feature translation network
CN109741256A (en) Image super-resolution reconstruction method based on sparse representation and deep learning
CN111932461A (en) Convolutional neural network-based self-learning image super-resolution reconstruction method and system
CN112001843B (en) A deep learning-based infrared image super-resolution reconstruction method
CN111612695B (en) Super-resolution reconstruction method for low-resolution face image
CN115984117B (en) Channel attention-based variation self-coding image super-resolution method and system
CN112950505B (en) Image processing method, system and medium based on generation countermeasure network
CN108805802B (en) Constraint condition-based front face reconstruction system and method of stacked stepping self-encoder
CN111861884A (en) A super-resolution reconstruction method of satellite cloud images based on deep learning
CN116612009A (en) Multi-scale connection generation countermeasure network medical image super-resolution reconstruction method
CN109146785A (en) A kind of image super-resolution method based on the sparse autocoder of improvement
CN116594061B (en) A seismic data denoising method based on multi-scale U-shaped attention network
CN109949217A (en) Video Super-Resolution Reconstruction Method Based on Residual Learning and Implicit Motion Compensation
CN118212129A (en) Multi-source fusion super-resolution method for hyperspectral remote sensing images based on bilinear unmixing
CN117764862A (en) Hyperspectral image reconstruction method based on space spectrum double priori frequency domain enhancement
CN116681592A (en) Image super-resolution method based on multi-scale adaptive non-local attention network
CN113628114B (en) A dual-channel sparse coding method for image super-resolution reconstruction
CN113920015A (en) An edge-preserving super-resolution reconstruction method for infrared images based on generative adversarial networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant