CN110223234A - Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion - Google Patents

Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion Download PDF

Info

Publication number
CN110223234A
CN110223234A CN201910506498.1A CN201910506498A CN110223234A CN 110223234 A CN110223234 A CN 110223234A CN 201910506498 A CN201910506498 A CN 201910506498A CN 110223234 A CN110223234 A CN 110223234A
Authority
CN
China
Prior art keywords
layer
network
image
shrinkage
sub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910506498.1A
Other languages
Chinese (zh)
Inventor
杨勇
张东阳
黄淑英
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN201910506498.1A priority Critical patent/CN110223234A/en
Publication of CN110223234A publication Critical patent/CN110223234A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution

Abstract

The invention discloses a kind of depth residual error network image super resolution ratio reconstruction methods based on cascade shrinkage expansion, it include: to obtain low resolution observed image sequence first, bicubic difference operation is carried out to piece image any in the sequence, obtains the initial valuation of high-definition picture;It is then based on the depth residual error neural network of cascade shrinkage expansion, the training of depth residual error neural network is carried out to initial valuation image, by trained characteristic pattern and initial valuation image addition, recovers the corresponding high-definition picture of low-resolution image.The present invention efficiently solves the problems, such as network structure redundancy, training complexity, image reconstruction details deficiency in the prior art, realizes the better detail recovery of super-resolution image, and algorithm is simple, the speed of service is fast, practical.

Description

Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion
Technical field
The present invention relates to multimedia technology fields, particularly belong to digital picture and digital technical field of video processing, especially It is related to a kind of depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion.
Background technique
The super-resolution rebuilding technology of image is widely applied to every field, including medical image, satellite image, video Monitoring, HD video and detection and identification mission.In satellite image field, due to by remote sensing satellite launch cost, at image quality Amount etc. conditions restriction, while there is also satellite image transmission process by uncertain atmospheric perturbation and the outer space electricity Magnetic signal interference, the resolution ratio of the remote sensing satellite image of acquisition are often unable to reach the demand of many tasks.Use image reconstruction Technology improves HR level, can effectively save satellite resource cost, it helps improve identification and the reconnaissance capability of remote sensing satellite. The purpose of image super-resolution is in the case where given one or several low-resolution images, and reconstruct obtains a super-resolution Image.In the past more than ten years, researcher proposes many image super-resolution rebuilding algorithms, according to low-resolution image Number, image super-resolution rebuilding algorithm can be divided into the super-resolution rebuilding algorithm of multiple image and the oversubscription of single image Resolution algorithm for reconstructing.
For the super-resolution rebuilding of single image, presently, there are many different algorithms.Based on traditional interpolation side Method is all by a basic function, such as bilinear interpolation, bicubic interpolation and nearest neighbor algorithm, it is generally the case that these Method is all simple and effective, but since the picture element interpolation of these algorithms operates, will lead to edge and aliasing effect occurs in high-frequency region It should be with fuzzy distortion.
The image super-resolution rebuilding technology based on deep learning method achieves very big development and effect in recent years, It rebuilds in effect and time efficiency and has surmounted traditional method based on rebuilding and based on rarefaction representation.But it is based on depth The super-resolution rebuilding technology of habit still can not preferably handle practical problem at present, therefore there are also very big rooms for promotion.Mesh Before, there are the following problems for the image super-resolution rebuilding technology based on depth convolutional neural networks:
1) there are redundancies for the structure of reconstruction network:
The super resolution ratio reconstruction method for being currently based on deep learning still has many problems, one of them is due to depth The level of network structure is deep, relatively more, the bigger problem of calculation amount that leads to that there are parameters.Image super-resolution weight at present Build that frame is widely used is the convolutional neural networks based on residual error study, and has been achieved for preferable effect.But such as What design residual error structure, does not increase with abundant super-resolution rebuilding details and network size, is still problem to be solved.
2) reconstruction image details is insufficient:
Image super-resolution rebuilding is an ill-conditioning problem, because of the height corresponding for each low-resolution image The possible solution domain of image in different resolution is possible very big.Currently based on the single image high-resolution of depth convolutional neural networks The accuracy and speed that rate is rebuild have breakthrough, and traditional image super-resolution rebuilding algorithm usually makes for increasing receptive field It is realized with more convolutional layers or using expansion convolution.But using more layers mean network it is huge and complicated and It is more difficult to train.In addition, frequently can lead to grid effect using expansion convolution in image super-resolution rebuilding.
Summary of the invention
It is provided the purpose of the present invention is reducing the limitation that cost overcomes hardware by Image Reconstruction Technology, while for people A kind of effective and feasible algorithm solution proposes a kind of depth residual error network image oversubscription based on cascade shrinkage expansion Resolution method for reconstructing realizes the better detail recovery of super-resolution image, algorithm to solve the above-mentioned problems of the prior art Simply, the speed of service is fast, practical.
To achieve the above object, the present invention provides a kind of depth residual error network image super-resolution based on cascade shrinkage expansion Rate method for reconstructing, specific technical solution are as follows:
A kind of depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion, comprising:
Step S1 obtains low resolution observed image sequence, and it is poor to carry out bicubic to piece image any in the sequence It is worth operation, obtains the initial valuation of high-definition picture;
Step S2 carries out depth residual error to initial valuation image based on the depth residual error neural network of cascade shrinkage expansion Neural network training, by trained characteristic pattern and initial valuation image addition, recovers the corresponding high score of low-resolution image Resolution image.
Preferably, in step S2, the training of depth residual error neural network is carried out to initial valuation image, comprising: use first It shrinks sub-network and shrinks initial valuation characteristics of image, then rebuild by the characteristics of image of contraction by extension sub-network defeated Out, the corresponding high-definition picture of low-resolution image is recovered finally by reconstruction sub-network.
Preferably, shrinking sub-network collapse step is to be input to initial valuation image and shrink sub-network first order convolution Layer obtains the first shrinkage layer C1;First shrinkage layer C1 is operated by down-sampling, is shunk using the study of convolutional layer and is obtained the Two shrinkage layer C2;Second shrinkage layer C2 is operated by down-sampling, learns to obtain third shrinkage layer C3 using a convolutional layer;The Three shrinkage layer C3 are operated by down-sampling, learn to obtain the 4th shrinkage layer C4 using cubic convolution layer, the 4th shrinkage layer C4 makees For the first extension layer D1 in extension sub-network.
Preferably, extension sub-network spread step be, the first extension layer D1 through up-sampling operation after with third shrinkage layer C3 It is added, obtains the second extension layer D2;Second extension layer D2 is received after a convolutional layer learns and up-samples operation, then with second Contracting layer C2 is added, and obtains third extension layer D3;Third extension layer D3 by a convolutional layer learns and up-sample operation after, then with First shrinkage layer C1 is added, and obtains the 4th extension layer D4, the 4th extension layer D4 is by a convolutional layer study, repeated contraction step And spread step.
Preferably, rebuilding sub-network reconstruction procedures is, in repetitive extension sub-network, the second extension layer D2 passes through on twice Sampling operation exports third feature figure R3 to the 4th extension layer D4;Third extension layer D3 is by primary up-sampling operation to the 4th Extension layer D4 exports second feature figure R2;By extending sub-network spread step, obtained the 4th extension layer D4 output first is special Sign figure R1;Fisrt feature figure R1, second feature figure R2 are added to obtain trained characteristic pattern with third feature figure R3, trained After characteristic pattern carries out convolution operation, then with initial valuation image addition, obtain the corresponding high-definition picture of low-resolution image.
Preferably, the convolution down-sampling operation for executing that step-length is 2 in sub-network is shunk, executing step-length in extension sub-network is 2 deconvolution up-samples operation, and each layer in network is made of 3 × 3 filters and activation primitive operation.
Preferably, shrink sub-network in, the second shrinkage layer C2 and the first shrinkage layer C1 sampling after the completion of, by fast connecting Through extension sub-network respective layer reaches the next stage shrinkage layer for shrinking sub-network.
Compared with prior art, beneficial effects of the present invention are specific as follows:
It is first the invention discloses a kind of depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion First to input low-resolution image, bicubic difference operation is carried out, the initial valuation of high-definition picture is obtained, then passes through contraction The operation of sub-network down-sampling carries out feature shrink, increases receptive field, and the shrinkage characteristic that will be finally obtained is more by extension sub-network Grade up-sampling realizes that details is rebuild and multi-stage characteristics export, and does not increase for feature-rich and computation burden, by fast connecting It connects and combines the Feature Mapping for shrinking sub-network with the Feature Mapping of extension sub-network, merged finally by sub-network is rebuild The feature of different scale, recovers the corresponding high-definition picture of low-resolution image, and the present invention can learn plan according to residual error Slightly realize that residual information is transmitted using sub-network multi-grade remnant learning structure is shunk, meanwhile, using expansion sub-network multistage weight The mode of building exports different levels detail textures, preferably recovers the details of low-resolution image.And inventive algorithm and its He compares mainstream algorithm, with good performance and efficiency, and when obtaining optimal performance and efficiency, this algorithm is compared with other mainstreams Algorithm time-consuming is shorter, and this algorithm network structure is simple, and the computation complexity speed of service is fast, while reducing calculating cost Preferable performance is achieved, resultant effect has preferable promotion, practical.
Detailed description of the invention
It in order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, below will be to institute in embodiment Attached drawing to be used is needed to be briefly described, it should be apparent that, the accompanying drawings in the following description is only some implementations of the invention Example, for those of ordinary skill in the art, without any creative labor, can also be according to these attached drawings Obtain other attached drawings.
Fig. 1 is inventive algorithm frame diagram;
Fig. 2 is the single neuronal structure figure of the present invention;
Fig. 3 is that the present invention is based on the image super-resolution rebuilding frames of the depth residual error network of cascade shrinkage expansion;
Fig. 4 is that inventive algorithm and current main-stream algorithm rebuild picture comparison diagram;
Fig. 5 is test run time analysis figure;
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete Site preparation description, it is clear that described embodiments are only a part of the embodiments of the present invention, instead of all the embodiments.It is based on Embodiment in the present invention, it is obtained by those of ordinary skill in the art without making creative efforts every other Embodiment shall fall within the protection scope of the present invention.
In order to make the foregoing objectives, features and advantages of the present invention clearer and more comprehensible, with reference to the accompanying drawing and specific real Applying mode, the present invention is described in further detail.
Referring to Fig. 2, a kind of depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion of the present invention, Include:
1. the nonlinear fitting of neural network
Neural network is exactly the multiple neurons connected according to certain rule, each neuron node therein in fact Receive input value of the output valve of upper one layer of neuron as this neuron, and input value is passed into next layer, input layer mind Attribute value can be directly passed to next layer (or output layer) through first node.In multilayer neural network, upper layer node There is a functional relation, this function is known as activation primitive between output and the input of lower level node.Single neuronal structure As shown in Fig. 2, single neuron knot mathematic(al) representation can be described as it is as follows:
For formula (1), f represents the neuron, wherein xjFor j-th of input of neuron, wjFor the jth of corresponding input A weight, b are bias term, and y is the output of neuron, which includes n input and an output.One neural network There is weights in each connection, these weights are exactly the parameter of neural network model in fact, that is, model to be learnt Thing.However, these parameters such as the connection type of a neural network, the number of plies of network, every layer of number of nodes, learning rate, then It is artificial set in advance.Hyper parameter (Hyper-Parameters) is normally referred to as these parameters being artificially arranged.
2. propagated forward
The input that propagated forward algorithm may be summarized to be each layer in neural network is weighted with corresponding connection weight And operation, result add a bias term, but in some neural networks bias term it is not necessary to.Then, then By a nonlinear activation function, such as ReLu, Sigmoid, Tanh function, obtained each layer of output result.Finally Constantly constantly operation forward by this method, obtains output layer result.For propagated forward, regardless of how high dimension is, Its process can be indicated with following formula:
Wherein,Indicating the weight inputted for j-th of l layers of i-th cell, f indicates activation primitive,It indicates The activation output valve of l layers of i-th cell.Since relatively simple, forward direction is propagated in the output therefore preceding paragraph that only need to calculate each layer Then start the backpropagation of network after propagation.
3. backpropagation
In order to which the parameter to network is updated, must just carry out gradient backpropagation (Back-Propagation, BP amplitude) is updated to calculate each parameter, back propagation is specific implementation of the gradient descent method on depth network.It is logical Often, gradient is calculated the objective function of depth network, and minimizing target function gradient descent method is a kind of Neural Network Optimization Algorithm, central idea are minimum (or most to wish to reach objective function along the direction undated parameter value of target function gradient Greatly).Gradient descent method is the most common optimization algorithm of deep learning network.
Back-propagation algorithm is usually used to trained CNN, i.e., is updated by backpropagation loss to each layer of gradient The corresponding parameter of network, network parameter update iteration is complete primary, is further continued for calculating the loss of next iteration by propagated forward, Then undated parameter is repeated, until reaching setting the number of iterations or reaching setting penalty values, so that the e-learning of training arrives Mapping relations as close possible to real goal map.Calculate the chain type derivation formula expression of gradient are as follows:
It include fixed sample the collection { (x of m sample equipped with one(1),y(1)),…,(x(m),y(m)), use batch gradient Descent method (Batch Gradient Descent) solves neural network.Specifically, to single sample (x, y) if using normal Two norms calculate its loss, then its cost function indicates are as follows:
In order to reduce the amplitude of weight, over-fitting is prevented, by the way that regularization term is added, wherein weight attenuation parameter λ is used for Control two in formula relative importances, loss function statement are as follows:
M indicates the batch size of batch gradient descent method, updates loss by constantly iterating to calculate, target is for parameter W and b asks the smallest cost J (W, b).Each iteration is updated W and b all in accordance with following formula in gradient descent method:
(W, b)=(W(1),b(1),W(2),b(2)…,W(n),b(n))………………8
Wherein α is learning rate, by calculating separately loss function to the partial derivative of weight W and biasing b multiplied by study Rate has then obtained the paces that parameter needs to update every time.Wherein, formula (8) is to need Optimal Parameters set, i.e. weight W and partially B is set, so far, gradient back-propagation algorithm is returning to propagated forward step after having updated all parameters, constantly with this Reciprocation cycle terminates until training.
Under normal conditions, one layer of active coating can be all followed by after convolutional layer.Activation primitive be for be added it is non-linear because Element, because the ability to express of linear model cannot achieve complicated Nonlinear Mapping.Output after convolution is passed through into activation letter Number is mapped to the value of activation primitive above threshold, and activation primitive exports the Nonlinear Mapping of feature in order to realize.
Correcting linear unit (Rectified Linear unit, ReLU) is also a common nonlinear activation function, ReLU function is linear in Zheng Banqu, that is, realize input mapping be itself, in the area Fu Ban, the functional value is then 0, derivative namely It is 0.The function is substantially better than the function of front two, is present most popular function.The function is relatively simple, reversed derivation Also very easy, mathematic(al) representation is as follows:
F (x)=max (0, x) ... ... ... ... ... ... ... 9
Non-linear behavior based on activation primitive, neural network may be implemented advanced by combined use activation primitive Nonlinear Mapping.Above-mentioned common activation primitive also include the more advanced activation primitive of some variations such as: Leaky It is the activation primitive extended that ReLU function, ELU function, MaxOut function, PReLU function etc., which are with ReLU, and Softmax function is Polytypic activation primitive is used for the extension of Sigmoid function.
Construction step such as Fig. 3 of depth residual error network algorithm based on cascade shrinkage expansion, comprising: shrink sub-network, expand It opens up sub-network and rebuilds sub-network.It shrinks and extension sub-network is respectively all comprising there are four the features of level to describe.It uses first Low-resolution image after bicubic interpolation operation is by shrinking sub-network increase receptive field and being used to expand by shrinkage characteristic expression Sub-network is opened up, the feature shrink of multistage down-sampling layer is then constructed;Extension sub-network constructs and to shrink sub-network corresponding The Fusion Features of multistage up-sampling layer are to realize that image detail restores;Subnet will be shunk by fast connection in rebuilding sub-network Network combines with the Feature Mapping of extension sub-network to realize that multistage reconstruction features export.
Algorithm implementation of the invention as shown in figure 3, each rectangular block represents one layer of convolutional layer in figure, Downsampling indicate execute step-length be 2 convolution realize down-sampling operation, Upsampling indicate execute step-length be 2 it is anti- Convolution realizes up-sampling operation, and each layer in network is made of 3 × 3 filters and activation primitive operation.Activation primitive uses Linear amending unit ReLU is parameterized, the expression ability of the activation primitive energy maximization network does not use the last layer then Activation primitive is to keep the layer as linear convergent rate.Specific each section is described as follows:
(1) sub-network is shunk
Sub-network is shunk in building;Shown in its structure such as Fig. 3 left hand half point, contraction sub-network is shunk every time is all by step-length 2 convolutional layer realizes down-sampling operation, is first input to initial valuation image and shrinks sub-network first order convolutional layer and obtain the One shrinkage layer C1;First shrinkage layer C1 is operated by down-sampling, is shunk using a convolutional layer study and is obtained the second shrinkage layer C2;Second shrinkage layer C2 is operated by down-sampling, learns to obtain third shrinkage layer C3 using a convolutional layer;Third shrinkage layer C3 is operated by down-sampling, learns to obtain the 4th shrinkage layer C4 using cubic convolution layer, the 4th shrinkage layer C4 is as extension The first extension layer D1 in network.
(2) sub-network is extended
Building extension sub-network;As shown in the half part to the right of the left side Fig. 3, extension sub-network extends all by step its structure every time A length of 2 deconvolution executes up-sampling operation, and realization gradually restores characteristic pattern spatial resolution, first by the first extension Layer D1 is added after up-sampling operation with third shrinkage layer C3, obtains the second extension layer D2;Second extension layer D2 passes through a secondary volume It after lamination study and up-sampling operation, then is added with the second shrinkage layer C2, obtains third extension layer D3;Third extension layer D3 passes through It after the study of convolutional layer and up-sampling operation, then is added with the first shrinkage layer C1, obtains the 4th extension layer D4, the 4th extension layer D4 is by a convolutional layer study, repeated contraction step and spread step, and in repetitive extension step, the first order extends sub-network Phase add operation in upper sampling process has only used the first order and has shunk the block that sub-network down-sampling corresponds to size, expanded in the second level It opens up during sub-network up-sampling is added, had not only used the second level to shrink corresponding piece of sub-network down-sampling, but also used the The block of level-one contraction sub-network down-sampling.
(3) sub-network is rebuild
In repetitive extension sub-network, the second extension layer D2 is by up-sampling operation twice to the 4th extension layer D4, output Third feature figure R3;Third extension layer D3, to the 4th extension layer D4, exports second feature figure R2 by primary up-sampling operation;Through Extension sub-network spread step is crossed, obtained the 4th extension layer D4 output fisrt feature figure R1;Fisrt feature figure R1, second feature Figure R2 is added to obtain trained characteristic pattern with third feature figure R3, after trained characteristic pattern carries out convolution operation, then and just Beginning valuation image addition obtains the corresponding high-definition picture of low-resolution image.
The training of depth residual error network algorithm and test based on cascade shrinkage expansion
(1) algorithm training
The image reconstruction of the multiple dimensioned factor may be implemented in algorithm for reconstructing of the present invention.Using with multiple scale factor models Benefit is that all parameters of network can be shared on different amplification factors.It is similar to VDSR, EDSR and DRNN method, we Model can rebuild the HR image that amplification factor is 2,3 and 4 times.It needs compared to other algorithms for each scale factor training list Only model, our method reduce many network parameters, can reduce the ruler of image reconstruction model in practical applications It is very little.In order to preferably train network, the network of proposition, DIV2K data are trained using a large-scale image reconstruction data set Collection contains the training image and authentication image of 1000 high-resolution image constructions.Usually by the cromogram in all data sets As being converted to YCbCr space from rgb space, an extract light intensity level is trained and tests, and color component is inserted by bicubic Value amplification obtains.Network will be used as loss function using mean square error function (MSE), while the constraint to parameter is added to prevent Network over-fitting:
Wherein, Θ is network parameter, and n is training set image number, y(i)For HR image, x(i)For LR image, Y (x(i)) be Network output, λ is regularization coefficient.This part implementation steps is as follows: carrying out random Gaussian core blur degradation to training image collection Operation, then carries out down-sampling and obtains multiple LR images, and HR image construction training image pair corresponding in training set, and uses number Expand training set according to enhancing;By obtained LR and HR image to being sent in contraction expansion depth residual error network, to train weight Establishing network;The Color Channel after the luminance channel and interpolation after reconstruction is finally transformed into rgb space again.
Usually in data prediction, training image is cut into the block that size is 64 × 64 preferably to accelerate to instruct Practice.Training data is expanded by two ways: by 90 °, 180 ° or 270 ° of image Random-Rotation and horizontal and vertical being turned over Turn image.Since the picture size in DIV2K data set is larger, we only randomly choose a kind of enhancing processing and expand training set To original 2 times.In addition to the last layer, each layer is all by the convolution algorithm of 64 3 × 3 filters and an activation primitive operation Composition.So filter initialization mode, which uses, is based on He-normal method.Using batch gradient descent algorithm instruction in training process Practice network, every batch of image number of blocks is set as 64.Training also uses learning rate decaying strategy, and initial learning rate is set as 0.0001,1/10th are decayed to every 15 wheels, decays four times in total until 60 wheel of training terminates.All experiments exist It is realized on TensorFlow platform using Python3.6 environment.
(2) test of heuristics
In order to further illustrate the validity of invention algorithm, we are compared this method and other image rebuilding methods Compared with objectively evaluating index, Y-PSNR (Peak Signal to Noise Ratio) abbreviation PSNR using common image And measure structural similarity (Structural Similarity index) abbreviation SSIM of two images index of similarity.
Experimental result has published the test code from different authors, including Bicubic, SRCNN, VDSR, DRCN, LapSRN, ARN and DWSR.On BSD100 and Urban100 assessment data set, method and existing some depth to proposition The image rebuilding method of study is tested, using the HR image by bicubic linear interpolation arithmetic as initial estimation, Luminance channel calculates PSNR the and SSIM value of reconstruction image, objectively evaluates the performance of image reconstruction algorithm, table 1 gives 2 bases The average PSNR and SSIM value of 7 kinds of algorithm for reconstructing, can be seen that algorithm for reconstructing of the present invention from the data of table 1 and exists on quasi- data set The PSNR value of the test result of four disclosed image reconstruction evaluation data sets is high compared with other mainstreams algorithms, obtained PSNR and SSIM value has reached preferable effect, and image synthesis effect is improved;
1 inventive algorithm of table and current main-stream algorithm are compared in the PSNR and SSIM of different data collection
Comparing result on open test data set is as shown in figure 4, can be with by the detail section amplification to reconstruction image Find out, algorithm for reconstructing of the invention can preferably rebuild the information such as detail textures, realize the better details of super-resolution image Restore, it is practical.
The average PSNR and mean test time of Urban100 test data set when amplifying 4 times as shown in Figure 5 Tetra- Riming time of algorithm of VDSR, DRCN, LAPSRN, DWSR of relationship, algorithm for reconstructing of the present invention and deep learning method carry out Comparison, algorithm for reconstructing of the present invention is with good performance and efficiency, when obtaining optimal performance and efficiency, inventive algorithm Time close to DWSR and LAPSR algorithm, still, the PSNR value of inventive algorithm is higher than DWSR and LAPSR algorithm.And Inventive algorithm is more shorter than other three kinds of algorithm time-consumings, and PSNR is apparently higher than other three kinds of algorithms, especially in SCRNN Inventive algorithm reduce calculate cost while achieve preferable performance, the low speed of service of computation complexity is fast, algorithm and Network structure is simple.
Embodiment described above is only that the preferred embodiment of invention is described, and is not defined to the range of invention, Without departing from the spirit of the design of the present invention, those of ordinary skill in the art make technical solution of the present invention various Modification and improvement are each fallen in the protection scope that claims of the present invention determines.

Claims (7)

1. a kind of depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion, characterized by comprising:
Step S1 obtains low resolution observed image sequence, carries out bicubic difference fortune to piece image any in the sequence It calculates, obtains the initial valuation of high-definition picture;
Step S2 carries out depth residual error nerve to initial valuation image based on the depth residual error neural network of cascade shrinkage expansion Network training, by trained characteristic pattern and initial valuation image addition, recovers the corresponding high-resolution of low-resolution image Image.
2. the depth residual error network image super resolution ratio reconstruction method according to claim 1 based on cascade shrinkage expansion, It is characterized by: carrying out the training of depth residual error neural network to initial valuation image, comprising: first using contraction in step S2 Sub-network shrinks initial valuation characteristics of image, the characteristics of image of contraction is then extended output by extension sub-network, most The corresponding high-definition picture of low-resolution image is recovered by rebuilding sub-network afterwards.
3. the depth residual error network image super resolution ratio reconstruction method according to claim 2 based on cascade shrinkage expansion, It is characterized by: shrinking sub-network collapse step is that initial valuation image is input to contraction sub-network first order convolutional layer and is obtained To the first shrinkage layer C1;First shrinkage layer C1 is operated by down-sampling, is shunk using a convolutional layer study and is obtained the second receipts Contracting layer C2;Second shrinkage layer C2 is operated by down-sampling, learns to obtain third shrinkage layer C3 using a convolutional layer;Third is received Contracting layer C3 is operated by down-sampling, learns to obtain the 4th shrinkage layer C4 using cubic convolution layer, the 4th shrinkage layer C4 is as expansion Open up the first extension layer D1 in sub-network.
4. the depth residual error network image super resolution ratio reconstruction method according to claim 2 based on cascade shrinkage expansion, It is characterized by: extension sub-network spread step is, the first extension layer D1 is added after up-sampling operation with third shrinkage layer C3, Obtain the second extension layer D2;Second extension layer D2 by a convolutional layer learns and up-sample operation after, then with the second shrinkage layer C2 is added, and obtains third extension layer D3;Third extension layer D3 is after a convolutional layer learns and up-samples operation, then with first Shrinkage layer C1 is added, and obtains the 4th extension layer D4, the 4th extension layer D4 is by a convolutional layer study, repeated contraction step and expansion Open up step.
5. the depth residual error network image super resolution ratio reconstruction method according to claim 2 based on cascade shrinkage expansion, It is characterized by: rebuilding sub-network reconstruction procedures is that in repetitive extension sub-network, the second extension layer D2 by up-sampling twice It operates to the 4th extension layer D4, exports third feature figure R3;Third extension layer D3 is by primary up-sampling operation to the 4th extension Layer D4, exports second feature figure R2;By extending sub-network spread step, obtained the 4th extension layer D4 output fisrt feature figure R1;Fisrt feature figure R1, second feature figure R2 are added to obtain trained characteristic pattern, trained feature with third feature figure R3 After figure carries out convolution operation, then with initial valuation image addition, obtain the corresponding high-definition picture of low-resolution image.
6. the depth residual error network image super-resolution rebuilding side based on cascade shrinkage expansion according to claim 3-4 Method, it is characterised in that: shrink the convolution down-sampling operation for executing that step-length is 2 in sub-network, executing step-length in extension sub-network is 2 Deconvolution up-sample operation, each layer in network is made of 3 × 3 filters and activation primitive operation.
7. the depth residual error network image super resolution ratio reconstruction method according to claim 3 based on cascade shrinkage expansion, It is characterized by: shrink sub-network in, the second shrinkage layer C2 and the first shrinkage layer C1 sampling after the completion of, it is through by fast connecting It extends sub-network respective layer or reaches the next stage shrinkage layer for shrinking sub-network.
CN201910506498.1A 2019-06-12 2019-06-12 Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion Pending CN110223234A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910506498.1A CN110223234A (en) 2019-06-12 2019-06-12 Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910506498.1A CN110223234A (en) 2019-06-12 2019-06-12 Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion

Publications (1)

Publication Number Publication Date
CN110223234A true CN110223234A (en) 2019-09-10

Family

ID=67816658

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910506498.1A Pending CN110223234A (en) 2019-06-12 2019-06-12 Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion

Country Status (1)

Country Link
CN (1) CN110223234A (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110648282A (en) * 2019-09-29 2020-01-03 燕山大学 Image super-resolution reconstruction method and system based on width neural network
CN110930327A (en) * 2019-11-18 2020-03-27 浙江工商大学 Video denoising method based on cascade depth residual error network
CN110991337A (en) * 2019-12-02 2020-04-10 山东浪潮人工智能研究院有限公司 Vehicle detection method based on self-adaptive double-path detection network
CN111028147A (en) * 2019-11-25 2020-04-17 天津大学 Three-dimensional medical image super-resolution reconstruction method based on cascade cavity convolution
CN111275686A (en) * 2020-01-20 2020-06-12 中山大学 Method and device for generating medical image data for artificial neural network training
CN111382845A (en) * 2020-03-12 2020-07-07 成都信息工程大学 Template reconstruction method based on self-attention mechanism
CN111563843A (en) * 2020-04-30 2020-08-21 苏州大学 Image super-resolution reconstruction method, system and related device
CN111915481A (en) * 2020-06-08 2020-11-10 北京大米未来科技有限公司 Image processing method, image processing apparatus, electronic device, and medium
CN111915490A (en) * 2020-08-14 2020-11-10 深圳清研智城科技有限公司 License plate image super-resolution reconstruction model and method based on multi-scale features
CN112508013A (en) * 2020-12-02 2021-03-16 哈尔滨市科佳通用机电股份有限公司 Lock catch loss fault detection method, system and device
CN113033704A (en) * 2021-04-22 2021-06-25 江西理工大学 Intelligent judging method for copper converter converting copper making period end point based on pattern recognition

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107358575A (en) * 2017-06-08 2017-11-17 清华大学 A kind of single image super resolution ratio reconstruction method based on depth residual error network
CN109146784A (en) * 2018-07-27 2019-01-04 徐州工程学院 A kind of image super-resolution rebuilding method based on multiple dimensioned generation confrontation network
CN109559297A (en) * 2018-11-15 2019-04-02 大连大学 A method of generating the Lung neoplasm detection of network based on 3D region
CN109767386A (en) * 2018-12-22 2019-05-17 昆明理工大学 A kind of rapid image super resolution ratio reconstruction method based on deep learning

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107358575A (en) * 2017-06-08 2017-11-17 清华大学 A kind of single image super resolution ratio reconstruction method based on depth residual error network
CN109146784A (en) * 2018-07-27 2019-01-04 徐州工程学院 A kind of image super-resolution rebuilding method based on multiple dimensioned generation confrontation network
CN109559297A (en) * 2018-11-15 2019-04-02 大连大学 A method of generating the Lung neoplasm detection of network based on 3D region
CN109767386A (en) * 2018-12-22 2019-05-17 昆明理工大学 A kind of rapid image super resolution ratio reconstruction method based on deep learning

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
KANGFU MEI等: "An Effective Single-Image Super-Resolution Model Using Squeeze-and-Excitation Networks", 《10.48550/ARXIV.1810.01831》 *
KANGFU MEI等: "An Effective Single-Image Super-Resolution Model Using Squeeze-and-Excitation Networks", 《10.48550/ARXIV.1810.01831》, 31 October 2018 (2018-10-31), pages 1 - 12 *
房雪键: "基于深度学习的图像分类算法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》, pages 13 - 17 *
郑启财: "基于深度学习的图像检索技术的研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》, pages 21 - 23 *
雷晓静: "基于深度学习的图像分类研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110648282B (en) * 2019-09-29 2021-03-23 燕山大学 Image super-resolution reconstruction method and system based on width neural network
CN110648282A (en) * 2019-09-29 2020-01-03 燕山大学 Image super-resolution reconstruction method and system based on width neural network
CN110930327A (en) * 2019-11-18 2020-03-27 浙江工商大学 Video denoising method based on cascade depth residual error network
CN110930327B (en) * 2019-11-18 2022-09-23 浙江工商大学 Video denoising method based on cascade depth residual error network
CN111028147A (en) * 2019-11-25 2020-04-17 天津大学 Three-dimensional medical image super-resolution reconstruction method based on cascade cavity convolution
CN111028147B (en) * 2019-11-25 2024-02-02 天津大学 Three-dimensional medical image super-resolution reconstruction method based on cascade cavity convolution
CN110991337A (en) * 2019-12-02 2020-04-10 山东浪潮人工智能研究院有限公司 Vehicle detection method based on self-adaptive double-path detection network
CN110991337B (en) * 2019-12-02 2023-08-25 山东浪潮科学研究院有限公司 Vehicle detection method based on self-adaptive two-way detection network
CN111275686B (en) * 2020-01-20 2023-05-26 中山大学 Method and device for generating medical image data for artificial neural network training
CN111275686A (en) * 2020-01-20 2020-06-12 中山大学 Method and device for generating medical image data for artificial neural network training
CN111382845B (en) * 2020-03-12 2022-09-02 成都信息工程大学 Template reconstruction method based on self-attention mechanism
CN111382845A (en) * 2020-03-12 2020-07-07 成都信息工程大学 Template reconstruction method based on self-attention mechanism
CN111563843A (en) * 2020-04-30 2020-08-21 苏州大学 Image super-resolution reconstruction method, system and related device
CN111563843B (en) * 2020-04-30 2023-11-28 苏州大学 Image super-resolution reconstruction method, system and related device
CN111915481A (en) * 2020-06-08 2020-11-10 北京大米未来科技有限公司 Image processing method, image processing apparatus, electronic device, and medium
CN111915481B (en) * 2020-06-08 2024-03-29 北京大米未来科技有限公司 Image processing method, device, electronic equipment and medium
CN111915490A (en) * 2020-08-14 2020-11-10 深圳清研智城科技有限公司 License plate image super-resolution reconstruction model and method based on multi-scale features
CN112508013A (en) * 2020-12-02 2021-03-16 哈尔滨市科佳通用机电股份有限公司 Lock catch loss fault detection method, system and device
CN112508013B (en) * 2020-12-02 2021-07-16 哈尔滨市科佳通用机电股份有限公司 Lock catch loss fault detection method, system and device
CN113033704A (en) * 2021-04-22 2021-06-25 江西理工大学 Intelligent judging method for copper converter converting copper making period end point based on pattern recognition
CN113033704B (en) * 2021-04-22 2023-11-07 江西理工大学 Intelligent judging method and system for copper converter converting copper-making final point based on pattern recognition

Similar Documents

Publication Publication Date Title
CN110223234A (en) Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion
Gao et al. Image super-resolution using knowledge distillation
CN107123089B (en) Remote sensing image super-resolution reconstruction method and system based on depth convolution network
CN110363716B (en) High-quality reconstruction method for generating confrontation network composite degraded image based on conditions
CN112489164B (en) Image coloring method based on improved depth separable convolutional neural network
CN110570363A (en) Image defogging method based on Cycle-GAN with pyramid pooling and multi-scale discriminator
CN111275171B (en) Small target detection method based on parameter sharing multi-scale super-division reconstruction
Chen et al. Single image super-resolution using deep CNN with dense skip connections and inception-resnet
CN107423747A (en) A kind of conspicuousness object detection method based on depth convolutional network
CN109598676A (en) A kind of single image super-resolution method based on Hadamard transform
CN109636721A (en) Video super-resolution method based on confrontation study and attention mechanism
CN108734675A (en) Image recovery method based on mixing sparse prior model
CN104408697B (en) Image Super-resolution Reconstruction method based on genetic algorithm and canonical prior model
CN107240136A (en) A kind of Still Image Compression Methods based on deep learning model
CN113538234A (en) Remote sensing image super-resolution reconstruction method based on lightweight generation model
CN113870124B (en) Weak supervision-based double-network mutual excitation learning shadow removing method
CN111461325A (en) Multi-target layered reinforcement learning algorithm for sparse rewarding environment problem
CN116822382B (en) Sea surface temperature prediction method and network based on space-time multiple characteristic diagram convolution
CN116311186A (en) Plant leaf lesion identification method based on improved transducer model
CN112560624A (en) High-resolution remote sensing image semantic segmentation method based on model depth integration
CN115601236A (en) Remote sensing image super-resolution reconstruction method based on characteristic information distillation network
Zhou et al. MSAR‐DefogNet: Lightweight cloud removal network for high resolution remote sensing images based on multi scale convolution
CN112686817B (en) Image completion method based on uncertainty estimation
CN115797181A (en) Image super-resolution reconstruction method for mine fuzzy environment
CN116206214A (en) Automatic landslide recognition method, system, equipment and medium based on lightweight convolutional neural network and double attention

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190910