CN112508792A - Single-image super-resolution method and system of deep neural network integration model based on online knowledge migration - Google Patents

Single-image super-resolution method and system of deep neural network integration model based on online knowledge migration Download PDF

Info

Publication number
CN112508792A
CN112508792A CN202011531087.7A CN202011531087A CN112508792A CN 112508792 A CN112508792 A CN 112508792A CN 202011531087 A CN202011531087 A CN 202011531087A CN 112508792 A CN112508792 A CN 112508792A
Authority
CN
China
Prior art keywords
learners
image
base
model
input
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011531087.7A
Other languages
Chinese (zh)
Inventor
张泊宇
罗喜伶
金晨
王雪檬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Innovation Research Institute of Beihang University
Original Assignee
Hangzhou Innovation Research Institute of Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Innovation Research Institute of Beihang University filed Critical Hangzhou Innovation Research Institute of Beihang University
Priority to CN202011531087.7A priority Critical patent/CN112508792A/en
Publication of CN112508792A publication Critical patent/CN112508792A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a single-image super-resolution method and a single-image super-resolution system of a deep neural network integration model based on online knowledge migration.A plurality of DNN-based base learners are used for constructing the integration model, the integration model comprises a plurality of DNN-based base learners, and each base learner is provided with an up-sampling module and a refinement network; the upsampling modules of different base learners adopt different upsampling methods; the basis learners are connected by an expanded Cross-batch unit, and the outputs of the basis learners are combined into the final output HR image by a learnable weight. The invention can actively introduce the difference into the base learners, namely actively enhance the difference among the base learners, and ensure that the integrated model can obtain good performance; the extended Cross-stick unit allows the beneficial inputs from other base learners to be strengthened, while the unfavorable inputs are weakened; the learnable weights can be adaptively combined according to the outputs of different basis learners, thereby achieving better results.

Description

Single-image super-resolution method and system of deep neural network integration model based on online knowledge migration
Technical Field
The invention relates to a single-image super-resolution method and a single-image super-resolution system of a deep neural network integration model based on online knowledge migration, which are particularly suitable for scenes with high requirements on details of reconstructed high-resolution images.
Background
Single Image Super Resolution (SISR) is a classical problem in the field of computer vision, aiming at reconstructing a High Resolution (HR) Image with larger size and richer details from a given Single Low Resolution (LR) Image. The technology is widely applied to remote sensing image imaging, video security and medical image imaging. A large number of SISR methods have been proposed, and can be specifically classified into interpolation, reconstruction, and learning methods. The interpolation method uses a domain interpolation-based method to reconstruct the HR image from the LR image, such as nearest neighbor interpolation, bicubic interpolation, etc., which are very fast and direct, but lack accuracy. Reconstruction-based methods constrain the possible solution space of the reconstructed HR image by defining fine explicit prior information (in the form of distributions, energy functions, etc.). However, the performance of these methods drops off rapidly as the upsampling factor increases. The learning-based approach learns a mapping function from LR to HR images using pairs of LR images and corresponding HR images. Among current SISR methods, learning-based methods, particularly Deep Neural Network (DNN) -based SISR models, are widely used and outperform interpolation or reconstruction-based models in performance.
Since SISR is a process of recovering a large amount of missing information according to limited information, it is an ill-defined problem that the solution is not unique, i.e. SISR methods using different recovery mechanisms will generate high resolution images containing different details for the same low resolution image. Combining the high resolution images produced by these different SISR methods using an integration model can provide high resolution images with finer details. The key to the excellent performance of the integrated model is that the individual base learners are excellent and have sufficient variability between the base learners. The invention provides an integrated model framework for end-to-end training of a SISR method, wherein a difference introduction module and an online knowledge migration module are innovatively designed, the difference introduction module can actively enhance the difference between base learners, and the online knowledge migration module can promote the performances of all the base learners in the training process, so that the performances of the integrated model are improved.
The ensemble model has been applied to the SISR problem and shows better performance than a single learner, such as ESCN and MSCN. The methods firstly initialize a plurality of DNNs with the same structure at random, train the DNNs by utilizing a training set respectively, and construct an integrated model by taking the DNNs as a base learner. In reconstructing the HR image, as shown in FIG. 1, the model first upsamples the LR image I with an upsampling module (Up, usually bicubic interpolation)LRAmplifying to a target size, and using the amplified image as input of each base learner to obtain a set of output
Figure BDA0002852138630000021
Finally, the group of outputs are combined into a final HR image I in an equal weight modeHR. However, the DNN-based integration model has the following problems: 1) DNN, the learning of an individual-base learner of an integrated model is a highly complex non-convex optimization problem, and the base learner learned by adopting a gradient descent method is easy to fall into bad local minimum values, so that the effect of the whole integrated model is influenced; 2) even if each base learner is well learned, the strong learning power of DNN will reduce the variability between different base learners, which is a guarantee that the integrated model achieves good performance. Therefore, these problems limit the integration model to make a significant breakthrough in the effect of SISR.
Disclosure of Invention
The invention aims to provide a novel integrated model framework based on DNN (discrete cosine transformation) for solving the problem of SISR (Single-input Single-output) and can solve the problems of insufficient difference among base learners and difficulty in training the base learners existing in the traditional DNN integrated model for solving the SISR, and the effect of reconstructing an HR image by the model is improved.
The technical scheme of the invention is as follows:
the invention firstly discloses a single-image super-resolution method of a deep neural network integration model based on online knowledge migration, which comprises the following steps:
1) constructing a DNN-based integrated model, wherein the integrated model comprises a plurality of DNN-based base learners, and each base learner is provided with an up-sampling module and a refinement network; the upsampling modules of different base learners adopt different upsampling methods; the base learners are connected by an expanded Cross-batch unit, and the activation graphs { x ] of all the base learners at the same layer are linearly combined1…xNGet the input of the next layer of these base learners
Figure BDA0002852138630000022
The outputs of the basis learners are combined into the final output HR image through learnable weights;
2) training the integrated model by using a training set containing LR images and HR images corresponding to the LR images as labels, and obtaining the integrated model capable of performing super resolution to high resolution images through training;
3) and carrying out super-resolution processing on the low-resolution images in the test data set by using the trained integrated model.
The invention also discloses a deep neural network integration system based on online knowledge migration for single image super-resolution processing, which comprises the following steps:
a plurality of DNN-based base learners; each base learner is provided with an up-sampling module and a refinement network; the up-sampling modules of different base learners adopt different up-sampling methods to up-sample the input LR image; the refinement network is composed of a feature extraction network, a plurality of residual blocks, a reconstruction network and a jump layer connection from a coarse input to an output, wherein the feature extraction network uses a convolution layer to extract features from the coarse input; each residual block comprises a convolutional layer with a Leaky ReLU activation function and a convolutional layer without an activation function; the reconstruction net uses the two convolution layers to reconstruct a residual image;
an extended Cross-stick cell; it linearly combines activation maps { x) of all base learners in the same layer1…xNGet the input of the next layer of these base learners
Figure BDA0002852138630000031
A learnable base learner output merging unit; which combines the outputs of the base learners into the final output HR image with learnable weights.
Compared with the prior art, the invention has the following beneficial effects:
(1) according to the invention, each base learner is additionally provided with an up-sampling module, and each up-sampling module adopts different up-sampling methods to generate different reconstructed image details, so that the difference can be actively introduced into the base learners, namely, the difference between the base learners is actively enhanced, and the integrated model can be ensured to obtain good performance; the up-sampling method is used as an up-sampling module instead of the traditional interpolation method, so that the refined network can obtain coarse input with higher quality, and the training difficulty of the refined network is reduced.
(2) The invention expands a Cross-batch unit as an online knowledge migration module; the extended Cross-stick unit linearly combines the activation maps { x } of all DNNs at the same level1…xNGet the input of the next layer of these DNNs
Figure BDA0002852138630000032
The invention can strengthen the beneficial input from other base learners and weaken the unfavorable input by the method.
(3) Compared with the traditional equal weight method, the learnable basis learner output merging mechanism is adopted, and the learnable weight w can be adaptively combined according to the outputs of different basis learners, so that a better effect is obtained. Meanwhile, the whole integrated model can be trained in an end-to-end mode, and the whole optimal solution can be found more conveniently.
Drawings
FIG. 1 is a conventional DNN-based integration model framework;
FIG. 2 is an integrated model framework for DNN proposed by the present invention;
FIG. 3 is a DNN integration model based on online knowledge migration;
FIG. 4 is a network structure of a base learner;
FIG. 5 is an expanded Cross-stick cell;
FIG. 6 is an integration model building process;
FIG. 7 is a comparison of the performance of KTDE and KTDE-VDSR, KTDE-DRRN, KTDE-MS-LapSRN;
FIG. 8 is a graph of KTDE and KTDE-w/o KT Training Loss (Training Loss) during Training;
FIG. 9 is a graph of PSNR changes on Set5 for KTDE and KTDE-w/o KT after different degrees of Gaussian noise have been added to the model;
FIG. 10 is a comparison of KTDE and KTDE-w/o KT in picture restoration detail after different degrees of Gaussian noise have been added to the model.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
As shown in fig. 1 and 2, compared with the traditional DNN-based integrated model framework (fig. 1), the integrated model framework designed by the present invention (fig. 2) is added with a "difference upsampling module" and an "online knowledge migration module". Wherein the former is given a different upsampling module per base learner
Figure BDA0002852138630000041
Thereby providing each base learner with inputs having different details, i.e., differences between the active reinforcement base learners; the latter connects two different base learners to each other, so that the error gradient signal can be from one base in the training processThe learner passes to different base learners, helping to improve overall performance.
Based on the framework, the invention designs a DNN integration model KTDE (FIG. 3) which can be trained end to verify the validity of the framework. The model comprises 3 base learners with 'difference up-sampling modules', the base learners are connected by 'online knowledge migration modules', and each base learner is combined into a final output, namely a reconstructed HR image, through a learnable weight w.
The present invention uses three DNN-based basis learners to build our integration model KTDE. In the proposed framework, each base learner has an upsampling module and a refinement network. Fig. 4 shows the network structure of the base learner we have designed. When given the input LR image ILRThe base learner first uses a unique upsampling module UpkIt is scaled up to the coarse input l 'of target size'LR. This module can actively introduce diversity into the basis learner since different up-sampling methods can produce different reconstructed image details. In this invention, three basis learners respectively adopt pre-trained DNN-based SISR models, namely VDSR, DRRN and MS-LapSRN, as an upsampling module. By using the advanced upsampling methods instead of the traditional interpolation method as the upsampling module, the refined network can obtain coarse input with higher quality, so that the training difficulty of the refined network is reduced.
The next refinement network (fig. 4) consists of a feature extraction network (FENet), three residual blocks (ResBlock), a reconstruction network (RecNet) and a hop connection from coarse input to output. Specifically, FENet uses a convolutional layer (Conv) to extract features from the coarse input. Each residual block is composed of a convolutional layer with a Leaky ReLU activation function and an no activation function. Finally, the model uses two convolution layers to reconstruct the residual image. The layer jump connection not only alleviates the training difficulty, but also preserves the diversity introduced by the "difference upsampling module". All convolutional layers have 64 channels (except the last convolutional layer which has only one channel for reconstruction), the filter size is set to 3 x 3, and the step size is set to 1 to ensure the input and output sizes are the same.
For online knowledge transfer between multiple basic learners, we extend the Cross-batch unit (i.misra, a.shurivastava, a.gupta, and m.hebert, "Cross-batch networks for multi-task learning," in Computer Vision and Pattern Recognition,2016, pp.3994-4003) to enable it to be applied to more than two DNNs. The extended Cross-latch cell (FIG. 5) linearly combines activation maps { x } for all DNNs at the same level1…xNGet the input of the next layer of these DNNs
Figure BDA0002852138630000051
As shown in equation (1). In training these DNNs, each DNN can obtain information from a different DNN as it propagates forward or backward, helping to obtain a better solution for each other.
Figure BDA0002852138630000052
Through the back propagation algorithm learning matrix A, the Cross-sticch unit can adaptively decide how much the output of a certain hidden layer of each base learner is transmitted to other base learners. Thus, for a base learner, the beneficial inputs from other base learners are strengthened, and the unfavorable inputs are weakened.
For the input LR image ILROutput of all base learners
Figure BDA0002852138630000053
Combined together by way of linear weighting, the weights w can be updated in the training using a back-propagation algorithm, as shown in equation (2). Compared with the traditional equal weight method, the learning w can be adaptively combined according to the outputs of different base learners, so that a better effect is obtained. Meanwhile, the whole integrated model can be trained in an end-to-end mode, and the whole optimal solution can be found more conveniently.
Figure BDA0002852138630000061
The present invention will be described in further detail with reference to fig. 6 and an exemplary embodiment. This example describes how to build an integration model and train. The method comprises the following specific steps:
the method comprises the following steps: a training data set is constructed. The model KTDE designed by the invention is a supervised learning model, so a data set which comprises an LR image and an HR image corresponding to the LR image as tags needs to be constructed. Since such paired LR images and HR images are difficult to directly obtain, the present example constitutes a training data set by collecting a large number of high-quality natural images as HR images and down-sampling them by m times respectively by bicubic interpolation to obtain LR images. Meanwhile, in order to improve the generalization capability of the model, the training data is augmented in the following way: 1) rotating each pair of images by 90, 180 and 270 degrees simultaneously, so that the data volume is expanded to 4 times; 2) and horizontally overturning the data expanded in the last step, and further expanding the data volume to 8 times. This example simultaneously collects images (from the public data sets Set5, Set14, BSDS100 and Urban100) that are not duplicative of the training data Set, and constructs test sets in the same manner (but without data augmentation) for testing the performance of the trained model.
Step two: and constructing an integration model and initializing. The algorithm firstly obtains pre-training models of VDSR, DRRN and MS-LapSRN through an open source community, and then constructs an integrated model shown in figure 2. Meanwhile, since the model designed by the invention adopts the Leaky ReLU activation function, the model parameter initialization method suitable for the activation function is used in the example to initialize the model parameters of each base learner. For matrix a, to make the training process more stable, the present example initializes its parameters on the diagonal to 0.8, with the remaining parameters being 0.1. For the integrated weight w, the present example treats each base learner as equally important at initialization, i.e., both are initialized to 1/3.
Step three: and (5) training a model. This example sets the MAE as a loss function and trains the model through a stochastic gradient descent optimization algorithm Adam. Learning of Adam's algorithmThe rate is set to 10 for the model parameters-3For matrix A, 10 is set-2To speed up convergence. Momentum term beta of simultaneous Adam algorithm1And beta2Set to 0.9 and 0.999 respectively. The model was trained on a 4 × 10 total5Step, the batch size of the training data for each step is 32. The pre-trained model does not participate in the training of the model.
Step four: and (5) testing the model. After the model training is finished, the model is used for reconstructing the LR images in the test set to obtain the HR images, and the similarity between the LR images and the real HR images is evaluated through the PSNR and SSIM indexes. Wherein the PSNR (Peak Signal to Noise Ratio) formula is shown in (3), the index range is (0, + ∞), and the larger the value is, the closer the reconstructed image quality is to the real HR image; SSIM (Structural Similarity) formula is shown in (4), the index range is [0,1], and the closer the value is to 1, the closer the reconstructed image quality is to the true HR image.
Figure BDA0002852138630000071
Figure BDA0002852138630000072
Wherein n is the number of pixel bits, generally 8 is taken, and MSE is the mean square error of the current image x and the reference image y; mu.sxAnd muyRepresenting the mean, σ, of the images x and yxAnd σyRepresenting the variance, σ, of the images x and y, respectivelyxyRepresenting the covariance of images x and y. c. C1=(k1L)2,c2=(k2L)2Is a constant used to maintain stability, and L is the dynamic range of pixel values, generally taken as 255, k1=0.01,k2=0.03。
The importance of the differential upsampling module of the present invention is illustrated below:
three integrated models KTDE-VDSR, KTDE-DRRN and KTDE-MS-LapSRN are additionally designed for comparison in the experiment, and the three integrated models are the same as the model designed by the inventionThe models are identical, except that the three models each use the same upsampling module (VDSR, DRRN and MS-LapSRN). In the experiment, under the condition that the up-sampling multiple is 2, the Chi-square distance (Chi-square distance) is adopted to calculate the output of different base learners aiming at the same test set and the same integrated model
Figure BDA0002852138630000073
The average difference between them, i.e. equation (5).
Figure BDA0002852138630000074
The experimental results are shown in table 1, and the degree of difference between the model outputs of different upsampling modules adopted by each base learner is significantly higher than that of the integrated model adopting the same upsampling model.
Further, this experiment compared the performance change of 4 models on the Set5 data Set during training, and the results are shown in fig. 7, where the curves represent the PSNR change with x 2 as the upsampling multiple on the Set5 data Set during training of each model. As can be seen from fig. 7, KTDE is superior in performance to other integrated models that use the same upsampling module. This experiment verifies the importance of the "difference upsampling module". Table 1KTDE and KTDE-VDSR, KTDE-DRRN, KTDE-MS-LapSRN mean chi-square distances between the outputs of the respective basis learners on different test sets
KTDE-VDSR KTDE-DRRN KTDE-MS-LapSRN KTDE
Set5 2.07×10-6 2.44×10-6 2.43×10-6 1.89×10-2
Set14 8.27×10-6 9.63×10-6 9.53×10-6 2.07×10-2
BSDS100 2.37×10-6 2.63×10-6 2.57×10-6 1.95×10-2
Urban100 1.79×10-5 2.01×10-5 4.90×10-5 1.20×10-2
The importance of the online knowledge migration module is explained below:
secondly, this experiment will verify the importance of the "online knowledge migration module". Specifically, the effectiveness of the model KTDE-w/o KT is verified by comparing the performance of the model KTDE and the model KTDE-w/o KT without the online knowledge migration module during training and testing. As can be seen from FIG. 8, KTDE with an "online knowledge migration Module" can converge faster than KTDE-w/o KT at training.
Meanwhile, in the experiment, Gaussian noise with the standard deviation sigma of 0-0.01 is added to the trained KTDE and KTDE-w/o KT, and the influence degree of the Gaussian noise on the PSNR index of the model on the Set5 is tested, so that the robustness of the model is verified. As shown in fig. 9, KTDE-w/o KT performance decreases significantly as the noise level increases; in contrast, KTDE performance declines slowly. The experiment also shows the anti-interference capability of the model to Gaussian noise in a visual mode. As shown in fig. 10, as the noise level increases, in the HR picture reconstructed by KTDE-w/o KT, a ring-shaped ghost appears at the edge; in contrast, the KTDE model has no obvious change in reconstruction effect. These experiments demonstrate the importance of the "online knowledge migration module".
The method of the invention is compared with the prior art methods as follows:
finally, the experiment compares the model with other SISR models (including SISR ensemble model, single learner model) at different magnifications for 4 test sets. The comparison result is shown in table 2, the model designed by the invention is superior to other models in both PSNR and SSIM indexes, and the result proves the effectiveness of the model.
Table 2 KTDE compares PSNR/SSIM of the existing SISR integrated model or single learner model under different test sets and different upsampling multiples
Figure BDA0002852138630000081
The above-mentioned embodiments only express several embodiments of the present invention, and the description thereof is more specific and detailed, but not construed as limiting the scope of the present invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the inventive concept, which falls within the scope of the present invention. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (7)

1. A single-image super-resolution method of a deep neural network integration model based on online knowledge migration is characterized by comprising the following steps:
1) constructing a DNN-based integrated model, wherein the integrated model comprises a plurality of DNN-based base learners, and each base learner is provided with an up-sampling module and a refinement network; the upsampling modules of different base learners adopt different upsampling methods; the base learners are connected by an expanded Cross-batch unit, and all the base learners are positioned in the same hidden layer by linearly combining the activation graphs { x1…xNGet the input of the next hidden layer of these base learners
Figure FDA0002852138620000011
The outputs of the basis learners are combined into the final output HR image through learnable weights;
2) training the integrated model by using a training set containing LR images and HR images corresponding to the LR images as labels, and obtaining the integrated model capable of performing super resolution to high resolution images through training;
3) and carrying out super-resolution processing on the low-resolution images in the test data set by using the trained integrated model.
2. The on-line knowledge transfer-based deep neural network integrated model single-image super-resolution method as claimed in claim 1, wherein for a given input LR image ILRThe up-sampling module of each base learner respectively adopts different up-sampling methods to obtain coarse input images I 'with different target sizes'LRAs input to its refinement network.
3. The on-line knowledge migration-based single-image super-resolution method for the deep neural network integration model, as set forth in claim 1, wherein the refinement network consists of a feature extraction network, a plurality of residual blocks, a reconstruction network and a jump layer connection from a coarse input to an output; the feature extraction network uses a convolutional layer to extract features from the coarse input; each residual block comprises a convolutional layer with a Leaky ReLU activation function and a convolutional layer without an activation function; finally, the reconstruction net uses the two convolution layers to reconstruct the residual image.
4. The on-line knowledge migration-based single-image super-resolution method for deep neural network integration model according to claim 1, wherein the expanded Cross-stich unit linearly combines activation maps { x } of all DNNs in the same hidden layer1…xNGet the input of the next hidden layers of DNN
Figure FDA0002852138620000012
As shown in equation (1);
Figure FDA0002852138620000013
by learning the matrix A through the back propagation algorithm, the expanded Cross-batch unit can adaptively determine how much the output of a certain hidden layer of each base learner is transmitted to other base learners.
5. The on-line knowledge transfer-based single-image super-resolution method for deep neural network integration model according to claim 1, wherein for input LR image ILROutput of all base learners
Figure FDA0002852138620000021
The weights w can be updated in the training by using a back propagation algorithm, as shown in formula (2);
Figure FDA0002852138620000022
6. the on-line knowledge transfer-based single-image super-resolution method for the deep neural network integration model, as set forth in claim 1, wherein the up-sampling module adopts a pre-trained model, and the pre-trained model does not participate in the integration model training.
7. An online knowledge migration-based deep neural network integration system for super-resolution processing of a single image, characterized by comprising:
a plurality of DNN-based base learners; each base learner is provided with an up-sampling module and a refinement network; the up-sampling modules of different base learners adopt different up-sampling methods to up-sample the input LR image; the refinement network is composed of a feature extraction network, a plurality of residual blocks, a reconstruction network and a jump layer connection from a coarse input to an output, wherein the feature extraction network uses a convolution layer to extract features from the coarse input; each residual block comprises a convolutional layer with a Leaky ReLU activation function and a convolutional layer without an activation function; the reconstruction net uses the two convolution layers to reconstruct a residual image;
an extended Cross-stick cell; it linearly combines activation maps { x) of all base learners in the same layer1…xNGet the input of the next layer of these base learners
Figure FDA0002852138620000023
A learnable base learner output merging unit; which combines the outputs of the base learners into the final output HR image with learnable weights.
CN202011531087.7A 2020-12-22 2020-12-22 Single-image super-resolution method and system of deep neural network integration model based on online knowledge migration Pending CN112508792A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011531087.7A CN112508792A (en) 2020-12-22 2020-12-22 Single-image super-resolution method and system of deep neural network integration model based on online knowledge migration

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011531087.7A CN112508792A (en) 2020-12-22 2020-12-22 Single-image super-resolution method and system of deep neural network integration model based on online knowledge migration

Publications (1)

Publication Number Publication Date
CN112508792A true CN112508792A (en) 2021-03-16

Family

ID=74922022

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011531087.7A Pending CN112508792A (en) 2020-12-22 2020-12-22 Single-image super-resolution method and system of deep neural network integration model based on online knowledge migration

Country Status (1)

Country Link
CN (1) CN112508792A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114779241A (en) * 2022-06-27 2022-07-22 北京航空航天大学杭州创新研究院 Package security check method and device, storage medium and electronic equipment

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102402690A (en) * 2011-09-28 2012-04-04 南京师范大学 Data classification method based on intuitive fuzzy integration and system
US20180075581A1 (en) * 2016-09-15 2018-03-15 Twitter, Inc. Super resolution using a generative adversarial network
CN109903225A (en) * 2019-01-29 2019-06-18 中南大学 A kind of medical image Enhancement Method based on deep learning
US20190206026A1 (en) * 2018-01-02 2019-07-04 Google Llc Frame-Recurrent Video Super-Resolution
CN111754403A (en) * 2020-06-15 2020-10-09 南京邮电大学 Image super-resolution reconstruction method based on residual learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102402690A (en) * 2011-09-28 2012-04-04 南京师范大学 Data classification method based on intuitive fuzzy integration and system
US20180075581A1 (en) * 2016-09-15 2018-03-15 Twitter, Inc. Super resolution using a generative adversarial network
US20190206026A1 (en) * 2018-01-02 2019-07-04 Google Llc Frame-Recurrent Video Super-Resolution
CN109903225A (en) * 2019-01-29 2019-06-18 中南大学 A kind of medical image Enhancement Method based on deep learning
CN111754403A (en) * 2020-06-15 2020-10-09 南京邮电大学 Image super-resolution reconstruction method based on residual learning

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
ISHAN MISRA 等: "Cross-stitch networks for multi-task learning", 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), pages 3996 - 3997 *
LINGFENG WANG等: "Ensemble based deep networks for image super-resolution", 《PATTERN RECOGNITION》, vol. 68, pages 192 - 194 *
RIBEIRO, E: "Exploring Texture Transfer Learning via Convolutional Neural Networks for Iris Super Resolution", 2017 INTERNATIONAL CONFERENCE OF THE BIOMETRICS SPECIAL INTEREST GROUP (BIOSIG) *
ZHANG, BY等: "Image scanning microscopy with a long depth of focus generated by an annular radially polarized beam", 《OPTICS EXPRESS》, vol. 28, no. 26, pages 39288 - 39298 *
孙超等: "基于迁移学习的红外图像超分辨率方法研究", 《激光与红外》, vol. 47, no. 12, pages 1559 - 1564 *
李伟;张旭东;: "基于卷积神经网络的深度图像超分辨率重建方法", 电子测量与仪器学报, no. 12 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114779241A (en) * 2022-06-27 2022-07-22 北京航空航天大学杭州创新研究院 Package security check method and device, storage medium and electronic equipment
CN114779241B (en) * 2022-06-27 2022-09-30 北京航空航天大学杭州创新研究院 Package security check method and device, storage medium and electronic equipment

Similar Documents

Publication Publication Date Title
CN111583109B (en) Image super-resolution method based on generation of countermeasure network
CN111047515B (en) Attention mechanism-based cavity convolutional neural network image super-resolution reconstruction method
CN109146784B (en) Image super-resolution reconstruction method based on multi-scale generation countermeasure network
CN110378844B (en) Image blind motion blur removing method based on cyclic multi-scale generation countermeasure network
CN107968962B (en) A kind of video generation method of the non-conterminous image of two frames based on deep learning
CN109035142B (en) Satellite image super-resolution method combining countermeasure network with aerial image prior
CN110458750B (en) Unsupervised image style migration method based on dual learning
CN112348743B (en) Image super-resolution method fusing discriminant network and generation network
CN111080513B (en) Attention mechanism-based human face image super-resolution method
CN109685716B (en) Image super-resolution reconstruction method for generating countermeasure network based on Gaussian coding feedback
CN106447626A (en) Blurred kernel dimension estimation method and system based on deep learning
CN102156875A (en) Image super-resolution reconstruction method based on multitask KSVD (K singular value decomposition) dictionary learning
CN112699844B (en) Image super-resolution method based on multi-scale residual hierarchy close-coupled network
CN110363068B (en) High-resolution pedestrian image generation method based on multiscale circulation generation type countermeasure network
CN111861886A (en) Image super-resolution reconstruction method based on multi-scale feedback network
CN114331830B (en) Super-resolution reconstruction method based on multi-scale residual error attention
CN115936985A (en) Image super-resolution reconstruction method based on high-order degradation cycle generation countermeasure network
CN111768340A (en) Super-resolution image reconstruction method and system based on dense multi-path network
CN112950480A (en) Super-resolution reconstruction method integrating multiple receptive fields and dense residual attention
Wang et al. A new blind image denoising method based on asymmetric generative adversarial network
CN113743301B (en) Solid-state nanopore sequencing electric signal noise reduction processing method based on residual self-encoder convolutional neural network
CN112508792A (en) Single-image super-resolution method and system of deep neural network integration model based on online knowledge migration
CN114494003A (en) Ancient character generation method combining shape transformation and texture transformation
CN110223224A (en) A kind of Image Super-resolution realization algorithm based on information filtering network
Feng et al. U 2-Former: Nested U-shaped Transformer for Image Restoration via Multi-view Contrastive Learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination