CN111833252A - Image Super-Resolution Method Based on SAE Dictionary Learning and Neighborhood Regression - Google Patents
Image Super-Resolution Method Based on SAE Dictionary Learning and Neighborhood Regression Download PDFInfo
- Publication number
- CN111833252A CN111833252A CN202010670836.8A CN202010670836A CN111833252A CN 111833252 A CN111833252 A CN 111833252A CN 202010670836 A CN202010670836 A CN 202010670836A CN 111833252 A CN111833252 A CN 111833252A
- Authority
- CN
- China
- Prior art keywords
- dictionary
- image
- sae
- input data
- dictionary learning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 55
- 239000011159 matrix material Substances 0.000 claims abstract description 27
- 238000013507 mapping Methods 0.000 claims description 20
- 238000012549 training Methods 0.000 claims description 20
- 238000012360 testing method Methods 0.000 claims description 14
- 238000002360 preparation method Methods 0.000 claims description 11
- 230000006870 function Effects 0.000 claims description 9
- 230000004913 activation Effects 0.000 claims description 6
- 238000005070 sampling Methods 0.000 claims description 4
- 238000000354 decomposition reaction Methods 0.000 claims description 3
- 238000011478 gradient descent method Methods 0.000 claims description 3
- 210000002569 neuron Anatomy 0.000 claims description 3
- 238000007781 pre-processing Methods 0.000 claims description 2
- 238000001914 filtration Methods 0.000 claims 2
- 238000000513 principal component analysis Methods 0.000 claims 2
- 238000000638 solvent extraction Methods 0.000 claims 2
- 230000001105 regulatory effect Effects 0.000 claims 1
- 238000004590 computer program Methods 0.000 description 7
- 238000010586 diagram Methods 0.000 description 6
- 230000000694 effects Effects 0.000 description 5
- 238000012545 processing Methods 0.000 description 5
- 238000002474 experimental method Methods 0.000 description 4
- 238000004088 simulation Methods 0.000 description 4
- 101100365548 Caenorhabditis elegans set-14 gene Proteins 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000003860 storage Methods 0.000 description 3
- 238000010276 construction Methods 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 238000012847 principal component analysis method Methods 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4053—Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2135—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/28—Determining representative reference patterns, e.g. by averaging or distorting; Generating dictionaries
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
本发明涉及基于SAE字典学习和邻域回归的图像超分辨率方法,首先针对字典学习模型SAE准备输入数据,并进行字典的构造与训练;然后结合邻域回归理论和字典求解投影矩阵;最后基于投影矩阵进行图像重建,获得高分辨率图像。本发明一方面提高字典的特征表达能力,减小重建结果对字典的依赖性;另一方面融入邻域回归理论,提高重建速度。
The present invention relates to an image super-resolution method based on SAE dictionary learning and neighborhood regression. First, input data is prepared for the dictionary learning model SAE, and a dictionary is constructed and trained; then, the projection matrix is solved by combining the neighborhood regression theory and the dictionary; finally, the image is reconstructed based on the projection matrix to obtain a high-resolution image. On the one hand, the present invention improves the feature expression ability of the dictionary and reduces the dependence of the reconstruction result on the dictionary; on the other hand, the neighborhood regression theory is incorporated to improve the reconstruction speed.
Description
技术领域technical field
本发明涉及图像超分辨率方法设计领域,特别是一种基于SAE字典学习和邻域回归的图像超分辨率方法。The invention relates to the field of image super-resolution method design, in particular to an image super-resolution method based on SAE dictionary learning and neighborhood regression.
背景技术Background technique
在现实中,由于图像采集设备的限制、场景变化以及光源等因素,往往不能得到高质量的图像,当图像的分辨率较低就无法满足实际应用的要求。图像超分辨率(Super-Resolution,SR)方法利用图像信号处理技术将已有的单幅或多幅低分辨率(LowResolution,LR)图像重建成高分辨率(High Resolution,HR)图像,其关键在于在重建过程中加入一定的附加信息来弥补图像降质过程中损失的细节信息。由于SR重建能够突破成像器件自身固有分辨率的限制实现图像分辨率的提升,因此其在遥感、医疗、视频监控等领域都具有重要的应用价值。In reality, due to factors such as the limitation of image acquisition equipment, scene changes, and light sources, high-quality images are often not obtained. When the resolution of the image is low, it cannot meet the requirements of practical applications. Image Super-Resolution (SR) method uses image signal processing technology to reconstruct existing single or multiple low-resolution (LowResolution, LR) images into high-resolution (High Resolution, HR) images. The key It is to add some additional information in the reconstruction process to make up for the detailed information lost in the process of image degradation. Since SR reconstruction can break through the limitation of the inherent resolution of the imaging device and improve the image resolution, it has important application value in remote sensing, medical treatment, video surveillance and other fields.
目前,SR方法主要分为3种类型:基于插值、基于重建和基于学习的SR方法。其中,基于学习的SR方法是近年来的热点方向,而基于字典学习的SR方法是基于学习的SR方法中最为流行的一种,最早由Yang等提出。该算法基于压缩感知理论而提出,采用字典联合学习法学习HR和LR字典对,先利用LR图像块和LR字典求得LR稀疏系数,然后基于HR和LR图像块具有相同的稀疏表示系数的假设,利用LR稀疏系数和HR字典重建HR图像块。该算法能够获取到充分的先验知识,具有较好的主观视觉效果,但是块效应较为明显,而且重建的效果对学习的字典具有很大的依赖性,而且重建耗时较长。Currently, SR methods are mainly divided into three types: interpolation-based, reconstruction-based, and learning-based SR methods. Among them, the learning-based SR method is a hot topic in recent years, and the dictionary learning-based SR method is the most popular one among the learning-based SR methods, which was first proposed by Yang et al. The algorithm is based on the compressed sensing theory. It adopts the dictionary joint learning method to learn the HR and LR dictionary pairs. First, the LR image block and the LR dictionary are used to obtain the LR sparse coefficient, and then based on the assumption that the HR and LR image blocks have the same sparse representation coefficient , using LR sparse coefficients and HR dictionary to reconstruct HR image patches. The algorithm can obtain sufficient prior knowledge and has good subjective visual effect, but the block effect is obvious, and the reconstruction effect has a great dependence on the learned dictionary, and the reconstruction takes a long time.
发明内容SUMMARY OF THE INVENTION
有鉴于此,本发明的目的是提出一种基于SAE字典学习和邻域回归的图像超分辨率方法,一方面提高字典的特征表达能力,减小重建结果对字典的依赖性;另一方面融入邻域回归理论,提高重建速度。In view of this, the purpose of the present invention is to propose an image super-resolution method based on SAE dictionary learning and neighborhood regression. Neighborhood regression theory to improve reconstruction speed.
本发明采用以下方案实现:一种基于SAE字典学习和邻域回归的图像超分辨率方法,具体包括以下步骤:The present invention adopts the following scheme to realize: an image super-resolution method based on SAE dictionary learning and neighborhood regression, which specifically includes the following steps:
针对字典学习模型SAE准备输入数据,并进行字典的构造与训练;Prepare input data for the dictionary learning model SAE, and construct and train the dictionary;
结合邻域回归理论和字典求解投影矩阵;Combine neighborhood regression theory and dictionary to solve projection matrix;
基于投影矩阵进行图像重建,获得高分辨率图像。Image reconstruction is performed based on the projection matrix to obtain high-resolution images.
进一步地,所述针对字典学习模型SAE准备输入数据具体为:Further, the described preparation input data for the dictionary learning model SAE is specifically:
将HR图像样本Ih进行下采样,得到LR图像Il,并将LR图像上采样得到中间图像Im; The HR image sample Ih is down-sampled to obtain the LR image Il, and the LR image is upsampled to obtain the intermediate image Im ;
HR输入数据准备:将HR图像Ih与中间图像Im相减,得到差值图像Id,对差值图像Id进行分块与归一化,作为HR输入数据,记为m为 HR输入数据的样本数;HR input data preparation: subtract the HR image I h from the intermediate image Im to obtain the difference image I d , and block and normalize the difference image I d as HR input data, denoted as m is the number of samples of HR input data;
LR输入数据准备:对中间图像Im进行滤波,得到滤波图像,再对滤波图像进行归一化和图像分块,并对滤波图像块进行降维,记为n为LR 输入数据的样本数;LR input data preparation: filter the intermediate image Im to obtain a filtered image, then normalize and block the filtered image, and reduce the dimension of the filtered image block, denoted as n is the number of samples of LR input data;
将字典学习模型SAE的输入数据表示为S=[Sh,Sl]。Denote the input data of the dictionary learning model SAE as S=[ Sh , S l ].
进一步地,在LR输入数据准备时,采用主成分分析法PCA对滤波图像块进行降维。Further, when the LR input data is prepared, the principal component analysis method PCA is used to reduce the dimension of the filtered image block.
进一步地,所述进行字典的构造与训练具体为:Further, the construction and training of the dictionary are specifically:
结合字典学习的需要,在SAE的代价函数中,采用平均绝对值误差代替均方误差,得到改进SAE字典学习模型;Combined with the needs of dictionary learning, in the cost function of SAE, the mean absolute value error is used to replace the mean square error, and an improved SAE dictionary learning model is obtained;
输入S=[Sh,Sl],采用改进SAE字典学习模型进行学习,获得输入层与隐含层之间的权重W1,将权重转化为HR和LR字典对{Dh,Dl}。Input S=[S h , S l ], use the improved SAE dictionary learning model for learning, obtain the weight W 1 between the input layer and the hidden layer, and convert the weight into HR and LR dictionary pairs {D h , D l } .
进一步地,所述结合字典学习的需要,在SAE的代价函数中,采用平均绝对值误差代替均方误差,得到改进SAE字典学习模型具体为:Further, in combination with the needs of dictionary learning, in the cost function of SAE, the mean absolute value error is used to replace the mean square error, and the improved SAE dictionary learning model is specifically:
设si∈S为输入数据,oi∈O为输出数据,改进的SAE字典模型如下:Let s i ∈ S be the input data and o i ∈ O be the output data. The improved SAE dictionary model is as follows:
其中,第一项JMAE(θ)为重构误差项,此处采用平均绝对误差表示,m和n分别表示HR和LR输入数据的样本数;第二项Jweight(θ)为权重衰减项,用于减少权值的量级,防止过拟合,λ为该项的调节参数,表示第l-1层节点i与第l层节点j 的连接权重,l表示网络的层数,Nl表示第l层的节点数,Nl+1第l+1层的节点数;第三项Jsparse(θ)为隐含层稀疏正则项,为隐含层神经元的平均激活量,ρ为设定好的预期激活量,γ为调节参数,N2表示第2层的节点数;其中,采用式(2)表示:Among them, the first term J MAE (θ) is the reconstruction error term, which is represented by the mean absolute error here, m and n represent the number of samples of HR and LR input data respectively; the second term J weight (θ) is the weight decay term , used to reduce the magnitude of the weights and prevent overfitting, λ is the adjustment parameter of this item, Represents the connection weight between the l-1 layer node i and the l layer node j, l represents the number of layers of the network, N l represents the number of nodes in the l layer, N l+1 The number of nodes in the l+1 layer; the third The term J sparse (θ) is the hidden layer sparse regular term, is the average activation of neurons in the hidden layer, ρ is the set expected activation, γ is the adjustment parameter, and N 2 represents the number of nodes in the second layer; among them, It is expressed by formula (2):
进一步地,所述输入S=[Sh,Sl],采用改进SAE字典学习模型进行学习,获得输入层与隐含层之间的权重W1,将权重转化为HR和LR字典对{Dh,Dl}具体为:Further, the input S=[S h , S l ], the improved SAE dictionary learning model is used for learning, the weight W 1 between the input layer and the hidden layer is obtained, and the weight is converted into the HR and LR dictionary pair {D h , D l } are specifically:
在SAE字典学习模型的训练过程中,结合梯度下降法进行参数更新,最终获得输入层到隐含层的连接权重W1,其中W1={wi},i=1,2,...,m+n;根据网络权重与字典的关系,字典D等价于输入层与隐含层的链接权重W1,具体表示为HR字典 Dh={w1,w2,…,wm}、LR字典Dl={wm+1,wm+2,…,wm+n},字典对表示为D=(Dh,Dl),其中wi∈W1,且wi={w1,i,w2,i,...,wk,i},k为字典的维数,wk,i表示第k维第i个字典原子的权值。In the training process of the SAE dictionary learning model, the parameters are updated with the gradient descent method, and finally the connection weight W 1 from the input layer to the hidden layer is obtained, where W 1 ={ wi }, i=1,2,... ,m+n; according to the relationship between the network weight and the dictionary, the dictionary D is equivalent to the link weight W 1 of the input layer and the hidden layer, specifically expressed as HR dictionary D h ={w 1 ,w 2 ,...,w m } , LR dictionary D l ={w m+1 ,w m+2 ,...,w m+n }, the dictionary pair is expressed as D=(D h , D l ), where w i ∈ W 1 , and w i = {w 1,i ,w 2,i ,...,w k,i }, k is the dimension of the dictionary, w k,i is the weight of the i-th dictionary atom in the k-th dimension.
进一步地,所述结合邻域回归理论和字典求解投影矩阵具体为:首先,采用最近邻域方法计算HR和LR字典对{Dh,Dl}中每个原子在样本空间S=[Sh,Sl]的最近邻域映射关系{Nh,Nl};然后,基于映射关系{Nh,Nl},采用岭回归方法求解投影矩阵P。Further, the combination of the neighborhood regression theory and the dictionary to solve the projection matrix is specifically: first, adopt the nearest neighbor method to calculate the HR and LR dictionary pairs {D h , D l } in the sample space S=[S h for each atom in the , S l ] the nearest neighbor mapping relationship {N h , N l }; then, based on the mapping relationship {N h , N l }, the projection matrix P is solved by the ridge regression method.
进一步地,所述采用最近邻域方法计算HR和LR字典对{Dh,Dl}中每个原子在样本空间S=[Sh,Sl]的最近邻域映射关系{Nh,Nl}具体为:Further, the nearest neighbor method is used to calculate the nearest neighbor mapping relationship {N h ,N of each atom in the sample space S=[ Sh ,S l ] in the HR and LR dictionary pair {D h , D l } l }Specifically:
设S=[Sh,Sl]为训练样本,{Dh,Dl}为字典对,采用欧式距离计算原子在LR 字典Dl中的K个最近邻域块集合Nl,q:Let S=[S h , S l ] be the training sample, {D h , D l } be the dictionary pair, and use the Euclidean distance to calculate the atoms The set of K nearest neighbor blocks N l,q in the LR dictionary D l :
式中,表示LR训练样本Sl中第p个训练样本,表示LR字典的第q个原子;In the formula, represents the p-th training sample in the LR training sample S l , Represents the qth atom of the LR dictionary;
从式(3)可计算出每一个LR字典原子对应的K个LR近邻图像块Nl,q,根据K个近邻图像块的位置,可从HR训练样本中获得相应的K个HR近邻图像块Nh,q,当遍历整个LR字典Dl的所有原子,可得所有原子的最近邻LR图像块组合而成的映射关系Nl,及其对应的最近邻HR图像块组合而成的最近邻映射关系Nh,最终获得{Nh,Nl}。From equation (3), the K LR neighbor image blocks N l,q corresponding to each LR dictionary atom can be calculated. According to the positions of the K neighbor image blocks, the corresponding K HR neighbor image blocks can be obtained from the HR training samples. N h,q , when traversing all atoms of the entire LR dictionary D l , the mapping relationship N l composed of the nearest neighbor LR image blocks of all atoms can be obtained, and the nearest neighbor composed of the corresponding nearest neighbor HR image blocks. The mapping relation N h is finally obtained {N h , N l }.
进一步地,所述基于映射关系{Nh,Nl},采用岭回归方法求解投影矩阵P具体为:Further, based on the mapping relationship {N h , N l }, using the ridge regression method to solve the projection matrix P is specifically:
用LR最近邻映射关系Nl取代字典Dl,则重建β的表达式为 Replacing the dictionary D l with the LR nearest neighbor mapping relation N l , the expression for reconstructing β is
式中,β表示系数矩阵,Y表示待重建低分辨率图像,η是权重系数,用于缓解奇异性问题并保证系数分解的稳定性;In the formula, β represents the coefficient matrix, Y represents the low-resolution image to be reconstructed, and η is the weight coefficient, which is used to alleviate the singularity problem and ensure the stability of the coefficient decomposition;
采用岭回归方法求解式(4),将系数表示为The ridge regression method is used to solve equation (4), and the coefficients Expressed as
重建的HR图像X通过映射关系Nh和系数获得,The reconstructed HR image X is mapped by N h and coefficients get,
记式(6)中的为投影矩阵P。In Equation (6) is the projection matrix P.
进一步地,所述基于投影矩阵进行图像重建,获得高分辨率图像具体为:Further, performing image reconstruction based on the projection matrix to obtain a high-resolution image is specifically:
首先,对待重建的图像Y进行预处理,得到LR测试特征图像 Yt={y1,y2,…,yi,…,yn};然后,采用欧式距离在LR字典Dl中寻找LR测试特征图像块yi对应的最近邻字典原子dk,表达式为接着通过原子dk找到其对应的投影矩阵Pt,然后利用表达式xi=Ptyi,获得对应yi的HR图像块xi,以此重建所有的LR测试特征图像块,并组成HR图像X。First, preprocess the image Y to be reconstructed to obtain the LR test feature image Y t ={y 1 ,y 2 ,...,y i ,...,y n }; then, use Euclidean distance to find LR in the LR dictionary D l The nearest neighbor dictionary atom d k corresponding to the test feature image block y i is expressed as Then find its corresponding projection matrix P t through atom d k , and then use the expression xi =P t y i to obtain the HR image block xi corresponding to yi , so as to reconstruct all LR test feature image blocks, and form HR image X.
与现有技术相比,本发明有以下有益效果:本发明首先对于字典的学习采用改进稀疏自动编码器,充分利用其突出的特征学习能力,增强字典的特征表达能力,提高图像的重建质量;然后,本发明在基于字典学习的超分辨率框架中融入邻域回归理论,避免原来框架中的稀疏编码过程,减少计算量,提高重建的速度。Compared with the prior art, the present invention has the following beneficial effects: the present invention firstly adopts an improved sparse auto-encoder for the learning of the dictionary, makes full use of its outstanding feature learning ability, enhances the feature expression ability of the dictionary, and improves the reconstruction quality of the image; Then, the present invention integrates the neighborhood regression theory into the super-resolution framework based on dictionary learning, avoids the sparse coding process in the original framework, reduces the amount of computation, and improves the speed of reconstruction.
附图说明Description of drawings
图1为本发明实施例的SAE输入数据预处理流程。FIG. 1 is an SAE input data preprocessing process according to an embodiment of the present invention.
图2为本发明实施例的方法流程示意图。FIG. 2 is a schematic flowchart of a method according to an embodiment of the present invention.
具体实施方式Detailed ways
下面结合附图及实施例对本发明做进一步说明。The present invention will be further described below with reference to the accompanying drawings and embodiments.
应该指出,以下详细说明都是示例性的,旨在对本申请提供进一步的说明。除非另有指明,本文使用的所有技术和科学术语具有与本申请所属技术领域的普通技术人员通常理解的相同含义。It should be noted that the following detailed description is exemplary and intended to provide further explanation of the application. Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.
需要注意的是,这里所使用的术语仅是为了描述具体实施方式,而非意图限制根据本申请的示例性实施方式。如在这里所使用的,除非上下文另外明确指出,否则单数形式也意图包括复数形式,此外,还应当理解的是,当在本说明书中使用术语“包含”和/或“包括”时,其指明存在特征、步骤、操作、器件、组件和/ 或它们的组合。It should be noted that the terminology used herein is for the purpose of describing specific embodiments only, and is not intended to limit the exemplary embodiments according to the present application. As used herein, unless the context clearly dictates otherwise, the singular is intended to include the plural as well, furthermore, it is to be understood that when the terms "comprising" and/or "including" are used in this specification, it indicates that There are features, steps, operations, devices, components and/or combinations thereof.
如图2所示,本实施例提供了一种基于SAE字典学习和邻域回归的图像超分辨率方法,具体包括以下步骤:As shown in Figure 2, this embodiment provides an image super-resolution method based on SAE dictionary learning and neighborhood regression, which specifically includes the following steps:
步骤S1:针对字典学习模型SAE准备输入数据,并进行字典的构造与训练;Step S1: prepare input data for the dictionary learning model SAE, and construct and train the dictionary;
步骤S2:结合邻域回归理论和字典求解投影矩阵;Step S2: combine the neighborhood regression theory and the dictionary to solve the projection matrix;
步骤S3:基于投影矩阵进行图像重建,获得高分辨率图像。Step S3: Perform image reconstruction based on the projection matrix to obtain a high-resolution image.
在本实施例中,如图1所示,步骤S1中所述针对字典学习模型SAE准备输入数据具体为:In this embodiment, as shown in FIG. 1 , the preparation of input data for the dictionary learning model SAE described in step S1 is specifically:
步骤S11:将HR图像样本Ih进行下采样,得到LR图像Il,并将LR图像上采样得到中间图像Im;Step S11: down-sampling the HR image sample Ih to obtain the LR image Il , and up-sampling the LR image to obtain the intermediate image Im ;
步骤S12:HR输入数据准备:将HR图像Ih与中间图像Im相减,得到差值图像Id,对差值图像Id进行分块与归一化,作为HR输入数据,记为 m为HR输入数据的样本数;Step S12: HR input data preparation: subtract the HR image I h and the intermediate image I m to obtain a difference image I d , and block and normalize the difference image I d as HR input data, denoted as m is the number of samples of HR input data;
步骤S13:LR输入数据准备:对中间图像Im进行滤波,得到滤波图像,再对滤波图像进行归一化和图像分块,并对滤波图像块进行降维,记为 n为LR输入数据的样本数;Step S13: LR input data preparation: filter the intermediate image Im to obtain a filtered image, then normalize and block the filtered image, and reduce the dimension of the filtered image block, denoted as n is the number of samples of LR input data;
步骤S14:将字典学习模型SAE的输入数据表示为S=[Sh,Sl]。Step S14: Denote the input data of the dictionary learning model SAE as S=[S h , S l ].
其中,在LR输入数据准备时,采用主成分分析法PCA对滤波图像块进行降维。Among them, when the LR input data is prepared, the principal component analysis method PCA is used to reduce the dimension of the filtered image block.
在本实施例中,步骤S1中的所述进行字典的构造与训练具体为:In this embodiment, the construction and training of the dictionary in step S1 is specifically:
步骤S15:结合字典学习的需要,在SAE的代价函数中,采用平均绝对值误差代替均方误差,得到改进SAE字典学习模型;Step S15: Combined with the needs of dictionary learning, in the cost function of SAE, the mean absolute value error is used instead of the mean square error to obtain an improved SAE dictionary learning model;
设si∈S为输入数据,oi∈O为输出数据,改进的SAE字典模型如下:Let s i ∈ S be the input data and o i ∈ O be the output data. The improved SAE dictionary model is as follows:
其中,第一项JMAE(θ)为重构误差项,此处采用平均绝对误差(Mean SquaredError, MSE)表示,m和n分别表示HR和LR输入数据的样本数;第二项Jweight(θ)为权重衰减项,用于减少权值的量级,防止过拟合,λ为该项的调节参数,表示第l-1 层节点i与第l层节点j的连接权重,l表示网络的层数,Nl表示第l层的节点数, Nl+1第l+1层的节点数;第三项Jsparse(θ)为隐含层稀疏正则项,为隐含层神经元的平均激活量,ρ为设定好的预期激活量,其值接近于0,γ为该项的调节参数, N2表示第2层的节点数;对于显著偏离ρ的情况,一般采用相对熵来惩罚,如式(2)所示:Among them, the first item J MAE (θ) is the reconstruction error term, which is represented by the mean absolute error (Mean SquaredError, MSE) here, m and n represent the number of samples of HR and LR input data respectively; the second item J weight ( θ) is the weight decay term, which is used to reduce the magnitude of the weight value and prevent over-fitting, and λ is the adjustment parameter of this term, Represents the connection weight between the l-1 layer node i and the l layer node j, l represents the number of layers of the network, N l represents the number of nodes in the l layer, N l+1 The number of nodes in the l+1 layer; the third The term J sparse (θ) is the hidden layer sparse regular term, is the average activation of neurons in the hidden layer, ρ is the set expected activation, its value is close to 0, γ is the adjustment parameter of this item, N 2 represents the number of nodes in the second layer; for In the case of significant deviation from ρ, relative entropy is generally used to punish, as shown in formula (2):
步骤S16:输入S=[Sh,Sl],采用改进SAE字典学习模型进行学习,获得输入层与隐含层之间的权重W1,将权重转化为HR和LR字典对{Dh,Dl}。Step S16: Input S=[S h , S l ], use the improved SAE dictionary learning model for learning, obtain the weight W 1 between the input layer and the hidden layer, and convert the weight into HR and LR dictionary pairs {D h , D l }.
在SAE字典学习模型的训练过程中,结合梯度下降法进行参数更新,最终获得输入层到隐含层的连接权重W1,其中W1={wi},i=1,2,...,m+n;在字典学习中,可以通过字典矩阵和稀疏表示输入数据,而在SAE中,可以通过隐含层的表示与学习权重表示输入数据,根据二者之间的联系可知字典D等价于输入层与隐含层的链接权重W1,具体表示为HR字典Dh={w1,w2,…,wm}、LR字典 Dl={wm+1,wm+2,…,wm+n},字典对表示为D=(Dh,Dl),其中wi∈W1,且wi={w1,i,w2,i,...,wk,i},k为字典的维数,wk,i表示第k维第i个字典原子的权值。In the training process of the SAE dictionary learning model, the parameters are updated with the gradient descent method, and finally the connection weight W 1 from the input layer to the hidden layer is obtained, where W 1 ={ wi }, i=1,2,... ,m+n; in dictionary learning, the input data can be represented by dictionary matrix and sparse representation, while in SAE, the input data can be represented by the representation of the hidden layer and the learning weight, according to the relationship between the two, the dictionary D, etc. It is equivalent to the link weight W 1 between the input layer and the hidden layer, specifically expressed as HR dictionary D h ={w 1 ,w 2 ,...,w m }, LR dictionary D l ={w m+1 ,w m+2 , . _ _ _ _ _ _ k,i }, k is the dimension of the dictionary, w k,i represents the weight of the i-th dictionary atom in the k-th dimension.
在本实施例中,步骤S2具体为:In this embodiment, step S2 is specifically:
步骤S21:采用最近邻域方法计算HR和LR字典对{Dh,Dl}中每个原子在样本空间S=[Sh,Sl]的最近邻域映射关系{Nh,Nl};即,Step S21: Use the nearest neighbor method to calculate the nearest neighbor mapping relationship {N h ,N l } of each atom in the HR and LR dictionary pair {D h ,D l } in the sample space S=[ Sh ,S l ] ;which is,
设S=[Sh,Sl]为训练样本,{Dh,Dl}为字典对,采用欧式距离计算原子在LR 字典Dl中的K个最近邻域块集合Nl,q:Let S=[S h , S l ] be the training sample, {D h , D l } be the dictionary pair, and use the Euclidean distance to calculate the atoms The set of K nearest neighbor blocks N l,q in the LR dictionary D l :
式中,表示LR训练样本Sl中第p个训练样本,表示LR字典的第q个原子;In the formula, represents the p-th training sample in the LR training sample S l , Represents the qth atom of the LR dictionary;
从式(3)可计算出每一个LR字典原子对应的K个LR近邻图像块Nl,q,根据K个近邻图像块的位置,可从HR训练样本中获得相应的K个HR近邻图像块Nh,q,当遍历整个LR字典Dl的所有原子,可得所有原子的最近邻LR图像块组合而成的映射关系Nl,及其对应的最近邻HR图像块组合而成的最近邻映射关系Nh,最终获得{Nh,Nl}。From equation (3), the K LR neighbor image blocks N l,q corresponding to each LR dictionary atom can be calculated. According to the positions of the K neighbor image blocks, the corresponding K HR neighbor image blocks can be obtained from the HR training samples. N h,q , when traversing all atoms of the entire LR dictionary D l , the mapping relationship N l composed of the nearest neighbor LR image blocks of all atoms can be obtained, and the nearest neighbor composed of the corresponding nearest neighbor HR image blocks. The mapping relation N h is finally obtained {N h , N l }.
步骤S22:基于映射关系{Nh,Nl},采用岭回归方法求解投影矩阵P,即,Step S22: Based on the mapping relationship {N h , N l }, use the ridge regression method to solve the projection matrix P, that is,
用LR最近邻映射关系Nl取代字典Dl,则重建β的表达式为 Replacing the dictionary D l with the LR nearest neighbor mapping relation N l , the expression for reconstructing β is
式中,β表示系数矩阵,Y表示待重建低分辨率图像,η是权重系数,用于缓解奇异性问题并保证系数分解的稳定性;In the formula, β represents the coefficient matrix, Y represents the low-resolution image to be reconstructed, and η is the weight coefficient, which is used to alleviate the singularity problem and ensure the stability of the coefficient decomposition;
采用岭回归方法求解式(4),将系数表示为The ridge regression method is used to solve equation (4), and the coefficients Expressed as
重建的HR图像X通过映射关系Nh和系数获得,The reconstructed HR image X is mapped by N h and coefficients get,
记式(6)中的为投影矩阵P。In Equation (6) is the projection matrix P.
在本实施例中,步骤S3具体为:In this embodiment, step S3 is specifically:
首先,对待重建的图像Y进行预处理,得到LR测试特征图像 Yt={y1,y2,…,yi,…,yn};然后,采用欧式距离在LR字典Dl中寻找LR测试特征图像块yi对应的最近邻字典原子dk,表达式为接着通过原子dk找到其对应的投影矩阵Pt,然后利用表达式xi=Ptyi,获得对应yi的HR图像块xi,以此重建所有的LR测试特征图像块,并组成HR图像X。First, preprocess the image Y to be reconstructed to obtain the LR test feature image Y t ={y 1 ,y 2 ,...,y i ,...,y n }; then, use Euclidean distance to find LR in the LR dictionary D l The nearest neighbor dictionary atom d k corresponding to the test feature image block y i is expressed as Then find its corresponding projection matrix P t through atom d k , and then use the expression xi =P t y i to obtain the HR image block xi corresponding to yi , so as to reconstruct all LR test feature image blocks, and form HR image X.
接下来,本实施例采用以下仿真实验来进行进一步的说明。Next, this embodiment uses the following simulation experiments for further description.
本实施例所采用的仿真工具为MATLAB,评价指标为峰值信噪比PSNR和结构相似度SSIM,其中PSNR越大,SSIM越接近1,则超分辨率效果越好。The simulation tool used in this embodiment is MATLAB, and the evaluation indicators are peak signal-to-noise ratio (PSNR) and structural similarity (SSIM). The larger the PSNR, the closer the SSIM is to 1, the better the super-resolution effect.
仿真实验具体设置如下:The specific settings of the simulation experiment are as follows:
数据准备:为了保证实验的客观性,字典学习采用91张通用标准HR训练样本,测试图像来源于标准测试库Set5和Set14。为了定量地评价重建图像的质量,将这些测试图像作为HR参考图像,通过下采样获取待处理的LR图像。Data preparation: In order to ensure the objectivity of the experiment, dictionary learning adopts 91 general standard HR training samples, and the test images come from standard test libraries Set5 and Set14. In order to quantitatively evaluate the quality of the reconstructed images, these test images are used as HR reference images, and the LR images to be processed are obtained by downsampling.
对比算法:Bicubic、L1SR(Super Resolution with L1 Regression)、SISR(Single Image Super Resolution)、ANR(Anchored Neighborhood Regression)、SRCNN(Super Resolution using Convolution Neural Network)等5种SR算法作对比。Comparison algorithms: 5 SR algorithms, including Bicubic, L1SR (Super Resolution with L1 Regression), SISR (Single Image Super Resolution), ANR (Anchored Neighborhood Regression), and SRCNN (Super Resolution using Convolution Neural Network), are compared.
重要参数设置:采样因子为3,SAE隐含层的节点数为1024。Important parameter settings: the sampling factor is 3, and the number of nodes in the SAE hidden layer is 1024.
仿真实验主要分为2组,具体如下。The simulation experiments are mainly divided into two groups, as follows.
第1组实验:与不同SR方法比较。Group 1 experiments: Comparison with different SR methods.
表1列出了不同SR算法得到的重建图像对应的PSNR和SSIM,其中最后一栏标出的数值表示在相应的评价指标下本实施例对应算法的性能最优,10幅图像分别来自Set5和Set14。由表1可见,本实施例方法得到的PSNR和SSIM值大体上都是最优的,表明其重建效果更优。Table 1 lists the PSNR and SSIM corresponding to the reconstructed images obtained by different SR algorithms. The values marked in the last column indicate that the performance of the corresponding algorithm in this embodiment is the best under the corresponding evaluation indicators. The 10 images are from Set5 and Set14. It can be seen from Table 1 that the PSNR and SSIM values obtained by the method in this embodiment are generally optimal, indicating that the reconstruction effect is better.
表1不同SR方法的PSNR(dB)和SSIM值的比较Table 1 Comparison of PSNR (dB) and SSIM values for different SR methods
第2组实验:重建速度比较。Experiment 2: Comparison of reconstruction speed.
本组实验将不同算法在同样的设备和环境中运行,用于验证在基于字典学习的SR算法中融入邻域回归思想能提升的重建速度的效果。表2中列出了Set5和 Set14测试集合在不同SR算法下的平均重建时间。从中可以看出,本发明速度明显高于其他SR算法。In this group of experiments, different algorithms were run in the same equipment and environment to verify the effect of improving the reconstruction speed by incorporating the idea of neighborhood regression into the dictionary learning-based SR algorithm. Table 2 lists the average reconstruction time of Set5 and Set14 test sets under different SR algorithms. It can be seen from this that the speed of the present invention is significantly higher than other SR algorithms.
表2不同SR算平均重建时间(s)的比较Table 2 Comparison of the average reconstruction time (s) of different SR algorithms
本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。As will be appreciated by those skilled in the art, the embodiments of the present application may be provided as a method, a system, or a computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the present application. It will be understood that each flow and/or block in the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to the processor of a general purpose computer, special purpose computer, embedded processor or other programmable data processing device to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing device produce Means for implementing the functions specified in a flow or flow of a flowchart and/or a block or blocks of a block diagram.
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/ 或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions The apparatus implements the functions specified in the flow or flows of the flowcharts and/or the block or blocks of the block diagrams.
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded on a computer or other programmable data processing device to cause a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process such that The instructions provide steps for implementing the functions specified in the flow or blocks of the flowcharts and/or the block or blocks of the block diagrams.
以上所述,仅是本发明的较佳实施例而已,并非是对本发明作其它形式的限制,任何熟悉本专业的技术人员可能利用上述揭示的技术内容加以变更或改型为等同变化的等效实施例。但是凡是未脱离本发明技术方案内容,依据本发明的技术实质对以上实施例所作的任何简单修改、等同变化与改型,仍属于本发明技术方案的保护范围。The above are only preferred embodiments of the present invention, and are not intended to limit the present invention in other forms. Any person skilled in the art may use the technical content disclosed above to make changes or modifications to equivalent changes. Example. However, any simple modifications, equivalent changes and modifications made to the above embodiments according to the technical essence of the present invention without departing from the content of the technical solutions of the present invention still belong to the protection scope of the technical solutions of the present invention.
Claims (10)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010670836.8A CN111833252B (en) | 2020-07-13 | 2020-07-13 | Image super-resolution method based on SAE dictionary learning and neighborhood regression |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010670836.8A CN111833252B (en) | 2020-07-13 | 2020-07-13 | Image super-resolution method based on SAE dictionary learning and neighborhood regression |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111833252A true CN111833252A (en) | 2020-10-27 |
CN111833252B CN111833252B (en) | 2022-06-14 |
Family
ID=72922812
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010670836.8A Active CN111833252B (en) | 2020-07-13 | 2020-07-13 | Image super-resolution method based on SAE dictionary learning and neighborhood regression |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111833252B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109146785A (en) * | 2018-08-02 | 2019-01-04 | 华侨大学 | A kind of image super-resolution method based on the sparse autocoder of improvement |
CN109712074A (en) * | 2018-12-20 | 2019-05-03 | 黑龙江大学 | Remote sensing image super-resolution reconstruction method based on two-parameter beta process joint dictionary |
CN110097503A (en) * | 2019-04-12 | 2019-08-06 | 浙江师范大学 | The super-resolution method returned based on neighborhood |
EP3591608A1 (en) * | 2018-07-06 | 2020-01-08 | Tata Consultancy Services Limited | Method and system for image reconstruction using deep dictionary learning (ddl) |
-
2020
- 2020-07-13 CN CN202010670836.8A patent/CN111833252B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3591608A1 (en) * | 2018-07-06 | 2020-01-08 | Tata Consultancy Services Limited | Method and system for image reconstruction using deep dictionary learning (ddl) |
CN109146785A (en) * | 2018-08-02 | 2019-01-04 | 华侨大学 | A kind of image super-resolution method based on the sparse autocoder of improvement |
CN109712074A (en) * | 2018-12-20 | 2019-05-03 | 黑龙江大学 | Remote sensing image super-resolution reconstruction method based on two-parameter beta process joint dictionary |
CN110097503A (en) * | 2019-04-12 | 2019-08-06 | 浙江师范大学 | The super-resolution method returned based on neighborhood |
Non-Patent Citations (2)
Title |
---|
詹曙等: "基于耦合特征空间下改进字典学习的图像超分辨率重建", 《电子学报》 * |
黄德天等: "正则化技术和低秩矩阵在稀疏表示超分辨率算法中的应用", 《计算机辅助设计与图形学学报》 * |
Also Published As
Publication number | Publication date |
---|---|
CN111833252B (en) | 2022-06-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110119780B (en) | A Generative Adversarial Network-Based Super-resolution Reconstruction Method for Hyperspectral Images | |
CN111754403B (en) | An Image Super-Resolution Reconstruction Method Based on Residual Learning | |
CN110570353B (en) | Densely connected generative adversarial network single image super-resolution reconstruction method | |
CN110120011B (en) | A video super-resolution method based on convolutional neural network and mixed resolution | |
CN103093445B (en) | Unified feature space image super-resolution reconstruction method based on joint sparse constraint | |
CN111768342A (en) | A face super-resolution method based on attention mechanism and multi-level feedback supervision | |
CN110111256B (en) | Image super-resolution reconstruction method based on residual distillation network | |
CN109671022B (en) | A super-resolution method for image texture enhancement based on deep feature translation network | |
CN109741256A (en) | Image super-resolution reconstruction method based on sparse representation and deep learning | |
CN111932461A (en) | Convolutional neural network-based self-learning image super-resolution reconstruction method and system | |
CN112001843B (en) | A deep learning-based infrared image super-resolution reconstruction method | |
CN111612695B (en) | Super-resolution reconstruction method for low-resolution face image | |
CN115984117B (en) | Channel attention-based variation self-coding image super-resolution method and system | |
CN112950505B (en) | Image processing method, system and medium based on generation countermeasure network | |
CN108805802B (en) | Constraint condition-based front face reconstruction system and method of stacked stepping self-encoder | |
CN111861884A (en) | A super-resolution reconstruction method of satellite cloud images based on deep learning | |
CN116612009A (en) | Multi-scale connection generation countermeasure network medical image super-resolution reconstruction method | |
CN109146785A (en) | A kind of image super-resolution method based on the sparse autocoder of improvement | |
CN116594061B (en) | A seismic data denoising method based on multi-scale U-shaped attention network | |
CN109949217A (en) | Video Super-Resolution Reconstruction Method Based on Residual Learning and Implicit Motion Compensation | |
CN118212129A (en) | Multi-source fusion super-resolution method for hyperspectral remote sensing images based on bilinear unmixing | |
CN117764862A (en) | Hyperspectral image reconstruction method based on space spectrum double priori frequency domain enhancement | |
CN116681592A (en) | Image super-resolution method based on multi-scale adaptive non-local attention network | |
CN113628114B (en) | A dual-channel sparse coding method for image super-resolution reconstruction | |
CN113920015A (en) | An edge-preserving super-resolution reconstruction method for infrared images based on generative adversarial networks |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |