CN111382845B - Template reconstruction method based on self-attention mechanism - Google Patents
Template reconstruction method based on self-attention mechanism Download PDFInfo
- Publication number
- CN111382845B CN111382845B CN202010171371.1A CN202010171371A CN111382845B CN 111382845 B CN111382845 B CN 111382845B CN 202010171371 A CN202010171371 A CN 202010171371A CN 111382845 B CN111382845 B CN 111382845B
- Authority
- CN
- China
- Prior art keywords
- image
- level
- layer
- sampling
- template
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000000034 method Methods 0.000 title claims abstract description 32
- 230000007246 mechanism Effects 0.000 title claims abstract description 21
- 238000005070 sampling Methods 0.000 claims abstract description 65
- 238000000605 extraction Methods 0.000 claims abstract description 8
- 238000013507 mapping Methods 0.000 claims abstract description 7
- 238000004364 calculation method Methods 0.000 claims description 31
- 238000012549 training Methods 0.000 claims description 10
- 238000010606 normalization Methods 0.000 claims description 3
- 238000012545 processing Methods 0.000 claims description 3
- 238000001514 detection method Methods 0.000 description 19
- 230000008569 process Effects 0.000 description 6
- 238000013527 convolutional neural network Methods 0.000 description 5
- 230000007547 defect Effects 0.000 description 5
- 230000000694 effects Effects 0.000 description 5
- 238000011426 transformation method Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 230000003321 amplification Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000013508 migration Methods 0.000 description 1
- 230000005012 migration Effects 0.000 description 1
- 238000012821 model calculation Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 230000001575 pathological effect Effects 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 238000000513 principal component analysis Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/751—Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Evolutionary Computation (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention provides a template reconstruction method based on a self-attention mechanism, which comprises the steps of firstly carrying out down-sampling on an input image layer by layer to obtain a down-sampling image, carrying out feature extraction on the down-sampling image, then respectively carrying out table look-up mapping on the features of each layer by adopting the self-attention mechanism to obtain a residual error corresponding to each layer, finally carrying out up-sampling on the down-sampling image layer by layer, simultaneously fusing the residual error information of each layer to generate an up-sampling image corresponding to each layer, and finally generating an up-sampling image which is a reconstructed template image.
Description
Technical Field
The invention belongs to the field of machine vision detection, and particularly relates to a template reconstruction method based on a self-attention mechanism.
Background
The defect detection method based on template comparison is widely applied to the field of machine vision detection due to the advantages of high speed, high precision and the like. In the defect detection process based on template comparison, in order to improve the detection rate and reduce the false detection rate, the template reconstruction is usually required to be performed for preprocessing operation, so that the individuation difference of the object to be detected is eliminated. In general, template reconstruction can be divided into a rigid transformation method and a non-rigid transformation method, and the non-rigid transformation method has been receiving more and more attention in recent years due to its wider application scenarios.
The non-rigid transformation template reconstruction method can be divided into a method based on image registration and a method based on characteristic reconstruction according to different reconstruction principles, and the invention belongs to the non-rigid template reconstruction method based on the characteristic reconstruction. In this technical field, conventional methods include a method based on Kernel Principal Component Analysis (KPCA), and the like, and with the recent popularization and application of deep learning techniques, the method based on deep learning has started to make its way, wherein the application of an image restoration technique based on deep self-encoding (AutoEncoder) to template reconstruction is mainly represented.
In the template reconstruction method using the depth self-coding as the core technology, a Convolutional Neural Network (CNN) is utilized to construct two parts, namely an encoding network (encoder) for down-sampling of an image and a decoding network (decoder) for up-sampling of features, low-dimensional features of the input image are extracted through the encoding network, the reconstruction process from the features to the image is completed through the decoding network, the constrained reconstructed image and the input image are approximately the same in content, individualized differences such as displacement, chromatic aberration and defects are eliminated, and the reconstructed image is compared with the input image, so that defect content possibly contained in the input image is detected.
The above prior art has the following disadvantages:
1. in a deep self-coding network, low-dimensional features extracted through a coding network generally cannot contain all detail features beneficial to reconstruction, so that reconstructed images are blurred to different degrees, and detection false alarm is caused. Although means such as a generative countermeasure network (GAN) have been proposed in recent years to alleviate the problem of blurring, the principle is based on statistical information from other training images, rather than the image to be reconstructed itself, and therefore the false alarm problem cannot be completely solved.
2. For a decoding network, feature upsampling is usually performed by using a transposed-conv/deconv (transposed-conv/deconv) operation, and since the transposed convolution is a pathological problem, a checkerboard artifact is often generated on a final reconstruction result, thereby affecting a detection result. Although methods such as sub-pixel convolution (sub conv) have been proposed to alleviate this phenomenon, their effect cannot be eradicated.
3. Due to the fact that the interpretability of the convolutional neural network is poor, feedback from observation of reconstruction effects to design of a deep self-coding network is difficult, selection of hyper-parameters such as low-dimensional characteristic dimensions and the number of convolutional kernels is usually set based on experience, optimal balance on effects and efficiency is difficult to obtain, and migration of a reconstruction method between different scene problems is hindered.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a template reconstruction method based on a self-attention mechanism, which carries out table look-up mapping through the self-attention mechanism, calculates and obtains a residual error according to the characteristics extracted from an image, and realizes the reconstruction of a template image through the residual error. The invention reduces the problem of false detection, avoids the checkerboard artifact phenomenon and enables the observation of the reconstructed result to be more visual and controllable.
The specific implementation content of the invention is as follows:
a template reconstruction method based on a self-attention mechanism comprises the steps of firstly carrying out down-sampling on an input image layer by layer to obtain a down-sampling image, carrying out feature extraction on the down-sampling image, then respectively carrying out table look-up mapping on features of each layer by adopting the self-attention mechanism to obtain residual errors corresponding to each layer, finally carrying out up-sampling on the down-sampling image layer by layer, simultaneously fusing residual error information of each layer to generate an up-sampling image corresponding to each layer, and finally generating an up-sampling image which is a reconstructed template image.
In order to better implement the present invention, the specific steps of "performing table look-up mapping on the features of each level by using an attention mechanism to obtain a residual error corresponding to each level" further include:
s1, defining each level of characteristics of the input image as a three-dimensional tensor with a c dimension, wherein each space position in each level of characteristics is a vector with the length of c;
s2, defining a two-dimensional tensor query table with the size of c x n, and cross-multiplying any one spatial position of each level of characteristics with the two-dimensional tensor query table with the size of c x n to obtain a vector with the length of n;
s3, performing softmax normalization processing on the vector with the length of n obtained in the step S2 to obtain another vector with the length of n, wherein n elements contained in the obtained another vector with the length of n are all larger than or equal to 0, and the sum of the n elements is equal to 1; the n elements are attention weights;
s4, carrying out weighted summation on n codebook data with the same space size by using the attention weight obtained in the step S3 to obtain a residual error of a corresponding hierarchy of the codebook data; the residual error comprises high-frequency detail information which is missed from an up-sampling image of the next stage after up-sampling.
In order to better implement the present invention, further, the two-dimensional tensor look-up table and the codebook data are both automatically learned in the network training as network parameters.
In order to better implement the present invention, the step of "down-sampling the input image layer by layer to obtain a down-sampled image, and extracting features of the down-sampled image" specifically includes: performing down-sampling calculation on an input image for h times to obtain an h-layer down-sampled image; then, carrying out feature extraction on the input image and the downsampling image except the h layer; the feature A extracted from the input image is the feature of the level corresponding to the input image; and the features of the corresponding levels of the down-sampling images except the h-th level are as follows: and fusing the features extracted from the downsampled image of the current level with the features of the previous level.
To better implement the present invention, further, the down-sampling calculation includes nearest neighbor interpolation, bilinear interpolation, or bicubic interpolation calculation.
In order to better implement the present invention, further, the specific generating method of the reconstructed template image is as follows: carrying out up-sampling calculation level by level from the down-sampling image of the h layer; after the down-sampling image of the h layer is subjected to up-sampling calculation, the down-sampling image of the h layer and the residual error of the h-1 layer are added pixel by pixel to generate an up-sampling image of the h-1 layer; similarly, the upsampled image of each level from the h-1 th level to the top level is subjected to upsampling calculation and then is added with the residual error of the previous level pixel by pixel to generate the upsampled image of the previous level, and the upsampled image subjected to the upsampling calculation and the residual error addition for h times is the reconstructed template image and corresponds to the level of the input image.
Compared with the prior art, the method has the following advantages and beneficial effects:
1. in the process of utilizing the characteristics to calculate the residual error, the invention adopts a mode of combining an attention mechanism with a codebook to finish the calculation, thereby avoiding the ill-conditioned problem of solving the transposed convolution and further avoiding the occurrence of checkerboard artifact phenomenon;
2. high-frequency information contained in the codebook is easy to visualize, the physical significance of the number n of the codebook is very clear and easy to adjust, and the dimensionality c of the lookup table is obviously and positively related to n, so that the network structure, namely the hyper-parameter setting is more controllable, and visual adjustment can be conveniently carried out through observation of a reconstruction result; in general, the dimension super parameters such as c and n can achieve a good reconstruction effect when the value is small, such as less than 64, compared with the dimension value which is hundreds of times in the traditional reconstruction model, the network parameters and the calculated amount are remarkably reduced, the model calculation is rapid, and the model is easy to deploy on low-cost hardware with limited calculation capability;
3. the template is reconstructed by adopting a mode of up-sampling step by step, and a mode of residual correction is adopted in the reconstruction process, namely, low-frequency information of the image after reduction and amplification is utilized, so that the model only needs to estimate the residual, namely high-frequency detail information, instead of the traditional mode of directly estimating the image by using the model, the high-frequency and low-frequency information need to be estimated simultaneously, therefore, the problem is simplified, the model effect is better, the definition of the reconstructed template is higher, and the probability of false alarm detection is further reduced.
Drawings
FIG. 1 is a schematic flow diagram of the present invention;
fig. 2 is a block diagram of a template reconstruction system according to the present invention.
Detailed Description
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it should be understood that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments, and therefore should not be considered as a limitation to the scope of protection. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, are within the scope of the present invention.
Example 1:
a template reconstruction method based on a self-attention mechanism comprises the steps of firstly carrying out down-sampling on an input image layer by layer to obtain a down-sampling image, carrying out feature extraction on the down-sampling image, then respectively carrying out table look-up mapping on features of each layer by adopting the self-attention mechanism to obtain residual errors corresponding to each layer, finally carrying out up-sampling on the down-sampling image layer by layer, simultaneously fusing residual error information of each layer to generate an up-sampling image corresponding to each layer, and finally generating an up-sampling image which is a reconstructed template image.
Example 2:
based on the above embodiment 1, as shown in fig. 1, taking a network subjected to four times of down-sampling as an example, the template reconstruction network integrally presents a U-shaped structure: wherein, starting from the input image, the down-sampling is performed step by step in the left wing, and starting from the minimum down-sampling image, the up-sampling is performed step by step in the right wing; in the left wing down-sampling process, a thin arrow pointing downwards represents simple image down-sampling calculation, such as nearest neighbor calculation, bilinear calculation, bicubic interpolation calculation and the like; in the process of sampling on the right wing, a thin arrow with an upward direction represents image up-sampling calculation; as shown in fig. 1, the input image is subjected to a first downsampling calculation to generate 1/2 downsampled images, the 1/2 downsampled image is subjected to a second downsampling calculation to generate 1/4 downsampled images, the 1/4 downsampled image is subjected to a third downsampling calculation to generate 1/8 downsampled images, and finally the 1/8 downsampled image is subjected to a fourth downsampling calculation to generate 1/16 downsampled images;
in addition, feature extraction is performed on each level from the input image to 1/8 down-sampling images, as shown in fig. 1, the feature extracted from the input image is feature a, the feature of each level from the input image downwards is the fusion of the feature extracted from the down-sampling image of the level and the feature of the previous level, for example, feature B of the level where 1/2 down-sampling images are located is the fusion of feature a and the feature extracted from the 1/2 down-sampling image;
after 1/16 downsampled images and the characteristics of each hierarchy are obtained, residual errors corresponding to the characteristics of each hierarchy need to be calculated through an attention mechanism, the 1/16 downsampled images are subjected to four times of upsampling, and the residual errors of each hierarchy are combined to finally generate reconstructed template images; specifically, the operation of generating the reconstructed template image is as follows: carrying out up-sampling calculation on 1/16 down-sampled images, and simultaneously adding residuals D of corresponding levels of 1/8 down-sampled images pixel by pixel to generate 1/8 up-sampled images; performing upsampling calculation on 1/8 upsampled images, and simultaneously adding residuals C of corresponding levels of 1/4 downsampled images pixel by pixel to generate 1/4 upsampled images; in the same way, a reconstructed image can be generated through two operations;
the specific operation of calculating the residual error of each level through the self-attention mechanism is as follows:
1. defining each level of feature as a three-dimensional tensor with a certain typical dimension c, such as 8 or 16, and the like, that is, each spatial position in the feature is a vector with the length c;
2. a two-dimensional tensor query table with the size of c x n is arranged, the feature of each space position is cross-multiplied with the two-dimensional tensor query table to obtain a vector with the length of n, and the operation can be efficiently realized in a network through convolution operation with the size of 1x 1;
3. performing softmax normalization processing on the vector with the length n obtained in the step 2 to obtain another vector with the length n, wherein the sum of n elements is 1 and is greater than or equal to 0, and the vector is called attention weight;
4. carrying out weighted summation on n pieces of codebook data with the same spatial size by using the n-dimensional attention weight obtained in the step 3 to obtain a residual error for the level reconstruction; the residual mainly contains high-frequency detail information which is lost after the up-sampling calculation of the reconstructed image from the next stage.
Other parts of this embodiment are the same as those of embodiment 1, and thus are not described again.
Example 3:
based on the method, as shown in fig. 2, the apparatus includes a training module and a detection module connected to each other; the training module comprises a training data set and a reconstruction model module which are connected in sequence; the detection module comprises a sensor, an input image module and a differential detection module which are connected in sequence; the detection module also comprises a reconstruction template module connected with the input image module and the seal checking detection module; the training module and the detection module are connected through a reconstruction model module and a reconstruction template module;
the training module is an external server and is used for obtaining a reconstruction model through a series of training set learning. The detection module is a main part of the device, thin arrows in the figure show simple data transmission and calculation, and hollow thick arrows show calculation performed by using a convolutional neural network; the main workflow of the detection module is as follows:
1. acquiring an image of an object to be detected by using a sensor such as a camera, a camera and the like to obtain an input image;
2. the main principle and the flow for obtaining the reconstruction model by using the input image and the reconstruction model obtained by the training module are as described in embodiments 1 and 2, so that the description is omitted, and the reconstruction image, namely the template, is obtained by performing inference calculation on the obtained reconstruction model through a convolutional neural network;
3. and carrying out difference on the input image and the reconstruction template, taking an absolute value, carrying out appropriate post-processing such as high-low threshold filtering, connected domain analysis and the like, and finally obtaining a detection result and finishing output.
The above description is only a preferred embodiment of the present invention, and is not intended to limit the present invention in any way, and any simple modifications and equivalent variations of the above embodiment according to the technical spirit of the present invention are within the scope of the present invention.
Claims (4)
1. A template reconstruction method based on a self-attention mechanism is characterized in that firstly, down-sampling is carried out on an input image layer by layer to obtain a down-sampling image, feature extraction is carried out on the down-sampling image, then table look-up mapping is carried out on the features of each layer by adopting the self-attention mechanism to obtain residual errors corresponding to each layer, finally, the down-sampling image is up-sampled level by level, the residual error information of each layer is fused to generate an up-sampling image corresponding to each layer, and finally, the generated up-sampling image is a reconstructed template image;
the specific steps of performing table look-up mapping on the characteristics of each level by adopting an attention mechanism to obtain the residual error corresponding to each level are as follows:
s1, defining each level of characteristics of the input image as a three-dimensional tensor with a c dimension, wherein each space position in each level of characteristics is a vector with the length of c;
s2, defining a two-dimensional tensor query table with the size of c x n, and cross-multiplying any one spatial position of each level of characteristics with the two-dimensional tensor query table with the size of c x n to obtain a vector with the length of n;
s3, performing softmax normalization processing on the vector with the length of n obtained in the step S2 to obtain another vector with the length of n, wherein n elements contained in the obtained another vector with the length of n are all larger than or equal to 0, and the sum of the n elements is equal to 1; the n elements are attention weights;
s4, carrying out weighted summation on n pieces of codebook data with the same space size by using the attention weight obtained in the step S3 to obtain a residual error of a corresponding level of the codebook data; the residual error comprises high-frequency detail information which is lost after the upsampling of an upsampled image of the next stage;
the step of performing layer-by-layer level down-sampling on the input image to obtain a down-sampled image and performing feature extraction on the down-sampled image specifically includes: performing down-sampling calculation on an input image for h times to obtain an h-layer down-sampled image; then, performing feature extraction on the input image and the downsampled images except for the h layer; the feature A extracted from the input image is the feature of the level corresponding to the input image; and the features of the corresponding levels of the down-sampling images except the h-th level are as follows: and fusing the features extracted from the downsampled image of the current level with the features of the previous level.
2. The template reconstruction method based on the self-attention mechanism as claimed in claim 1, wherein the two-dimensional tensor look-up table and the codebook data are both automatically learned in network training as network parameters.
3. The self-attention mechanism-based template reconstruction method of claim 1, wherein the downsampling calculation comprises nearest neighbor interpolation, bilinear interpolation or bicubic interpolation calculation.
4. The template reconstruction method based on the self-attention mechanism as claimed in claim 1, wherein the specific generation method of the reconstructed template image is as follows: performing up-sampling calculation level by level from the down-sampling image of the h layer; after the down-sampling image of the h layer is subjected to up-sampling calculation, the down-sampling image of the h layer and the residual error of the h-1 layer are added pixel by pixel to generate an up-sampling image of the h-1 layer; similarly, the upsampled image of each level from the h-1 th level to the top level is subjected to upsampling calculation and then is added with the residual error of the previous level pixel by pixel to generate the upsampled image of the previous level, and the upsampled image subjected to the upsampling calculation and the residual error addition for h times is the reconstructed template image and corresponds to the level of the input image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010171371.1A CN111382845B (en) | 2020-03-12 | 2020-03-12 | Template reconstruction method based on self-attention mechanism |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010171371.1A CN111382845B (en) | 2020-03-12 | 2020-03-12 | Template reconstruction method based on self-attention mechanism |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111382845A CN111382845A (en) | 2020-07-07 |
CN111382845B true CN111382845B (en) | 2022-09-02 |
Family
ID=71219046
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010171371.1A Expired - Fee Related CN111382845B (en) | 2020-03-12 | 2020-03-12 | Template reconstruction method based on self-attention mechanism |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111382845B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112861071B (en) * | 2021-02-05 | 2022-09-02 | 哈尔滨工程大学 | High-speed rail traction system anomaly detection method based on depth self-coding |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016014354A1 (en) * | 2014-07-25 | 2016-01-28 | Siemens Aktiengesellschaft | Iterative analysis-based non-convex prior for enhanced sparse recovery |
CN106709875A (en) * | 2016-12-30 | 2017-05-24 | 北京工业大学 | Compressed low-resolution image restoration method based on combined deep network |
CN110151181A (en) * | 2019-04-16 | 2019-08-23 | 杭州电子科技大学 | Rapid magnetic resonance imaging method based on the U-shaped network of recurrence residual error |
CN110223234A (en) * | 2019-06-12 | 2019-09-10 | 杨勇 | Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion |
CN110659727A (en) * | 2019-09-24 | 2020-01-07 | 中国科学技术大学 | Sketch-based image generation method |
-
2020
- 2020-03-12 CN CN202010171371.1A patent/CN111382845B/en not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016014354A1 (en) * | 2014-07-25 | 2016-01-28 | Siemens Aktiengesellschaft | Iterative analysis-based non-convex prior for enhanced sparse recovery |
CN106709875A (en) * | 2016-12-30 | 2017-05-24 | 北京工业大学 | Compressed low-resolution image restoration method based on combined deep network |
CN110151181A (en) * | 2019-04-16 | 2019-08-23 | 杭州电子科技大学 | Rapid magnetic resonance imaging method based on the U-shaped network of recurrence residual error |
CN110223234A (en) * | 2019-06-12 | 2019-09-10 | 杨勇 | Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion |
CN110659727A (en) * | 2019-09-24 | 2020-01-07 | 中国科学技术大学 | Sketch-based image generation method |
Non-Patent Citations (3)
Title |
---|
A Fast Surface Defect Detection Method Based on Background Reconstruction;Chengkan Lv等;《International Journal of Precision Engineering and Manufacturing》;20191109;363–375 * |
基于稀疏表示的图像超分辨率重建研究;李欣;《中国优秀博硕士学位论文全文数据库(博士)信息科技辑》;20170515(第05期);I138-11 * |
自注意力与域适应对抗模板重建方法;贾可等;《现代信息科技》;20200925;第4卷(第18期);1-6 * |
Also Published As
Publication number | Publication date |
---|---|
CN111382845A (en) | 2020-07-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109509152B (en) | Image super-resolution reconstruction method for generating countermeasure network based on feature fusion | |
CN113139907B (en) | Generation method, system, device and storage medium for visual resolution enhancement | |
Lim et al. | DSLR: Deep stacked Laplacian restorer for low-light image enhancement | |
CN111062872B (en) | Image super-resolution reconstruction method and system based on edge detection | |
CN107154023B (en) | Based on the face super-resolution reconstruction method for generating confrontation network and sub-pix convolution | |
CN110136062B (en) | Super-resolution reconstruction method combining semantic segmentation | |
CN112419150B (en) | Image super-resolution reconstruction method of arbitrary multiple based on bilateral upsampling network | |
CN111861880B (en) | Image super-fusion method based on regional information enhancement and block self-attention | |
CN114549308B (en) | Image super-resolution reconstruction method and system with large receptive field and oriented to perception | |
CN110796622B (en) | Image bit enhancement method based on multi-layer characteristics of series neural network | |
CN110689483A (en) | Image super-resolution reconstruction method based on depth residual error network and storage medium | |
CN111242999A (en) | Parallax estimation optimization method based on up-sampling and accurate re-matching | |
CN113421186A (en) | Apparatus and method for unsupervised video super-resolution using a generation countermeasure network | |
CN115345791A (en) | Infrared image deblurring algorithm based on attention mechanism residual error network model | |
CN113724134A (en) | Aerial image blind super-resolution reconstruction method based on residual distillation network | |
Yang et al. | A survey of super-resolution based on deep learning | |
CN116563100A (en) | Blind super-resolution reconstruction method based on kernel guided network | |
CN111382845B (en) | Template reconstruction method based on self-attention mechanism | |
Sharma et al. | Different techniques of image SR using deep learning: a review | |
Chen et al. | Dynamic degradation intensity estimation for adaptive blind super-resolution: A novel approach and benchmark dataset | |
CN117593187A (en) | Remote sensing image super-resolution reconstruction method based on meta-learning and transducer | |
CN116188265A (en) | Space variable kernel perception blind super-division reconstruction method based on real degradation | |
CN110211059A (en) | A kind of image rebuilding method based on deep learning | |
CN115511733A (en) | Image degradation modeling method, neural network training method and device | |
CN115131414A (en) | Unmanned aerial vehicle image alignment method based on deep learning, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20220902 |
|
CF01 | Termination of patent right due to non-payment of annual fee |