CN110246084B - Super-resolution image reconstruction method, system and device thereof, and storage medium - Google Patents
Super-resolution image reconstruction method, system and device thereof, and storage medium Download PDFInfo
- Publication number
- CN110246084B CN110246084B CN201910409645.3A CN201910409645A CN110246084B CN 110246084 B CN110246084 B CN 110246084B CN 201910409645 A CN201910409645 A CN 201910409645A CN 110246084 B CN110246084 B CN 110246084B
- Authority
- CN
- China
- Prior art keywords
- image
- resolution
- reconstruction
- super
- convolution
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 43
- 238000013528 artificial neural network Methods 0.000 claims abstract description 25
- 238000011176 pooling Methods 0.000 claims abstract description 24
- 238000004364 calculation method Methods 0.000 claims abstract description 18
- 238000007781 pre-processing Methods 0.000 claims abstract description 4
- 230000006870 function Effects 0.000 claims description 12
- 238000007405 data analysis Methods 0.000 claims description 9
- 238000005070 sampling Methods 0.000 claims description 9
- 238000012544 monitoring process Methods 0.000 claims description 7
- 230000002708 enhancing effect Effects 0.000 claims description 4
- 238000010191 image analysis Methods 0.000 claims description 4
- 230000003321 amplification Effects 0.000 claims description 3
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 3
- 230000008707 rearrangement Effects 0.000 claims description 2
- 238000004422 calculation algorithm Methods 0.000 abstract description 6
- 238000003062 neural network model Methods 0.000 abstract description 3
- 238000013527 convolutional neural network Methods 0.000 description 8
- 230000008569 process Effects 0.000 description 8
- 230000000694 effects Effects 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 238000012545 processing Methods 0.000 description 5
- 230000004913 activation Effects 0.000 description 4
- 230000009466 transformation Effects 0.000 description 4
- 238000003384 imaging method Methods 0.000 description 3
- 239000011159 matrix material Substances 0.000 description 3
- 238000013138 pruning Methods 0.000 description 3
- 238000012549 training Methods 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 210000002569 neuron Anatomy 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000010187 selection method Methods 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 210000004027 cell Anatomy 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000008034 disappearance Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000009191 jumping Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000004321 preservation Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000009966 trimming Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4053—Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Processing (AREA)
Abstract
The invention discloses a super-resolution image reconstruction method, a system, a device and a storage medium thereof, wherein firstly, an original low-resolution image is input, and bicubic interpolation is carried out on the input low-resolution image through image preprocessing; and then, the preprocessed images are connected and arranged with each other through a sparse deep convolution neural network containing pooling manifold constraint, high-frequency detail information components missing in the images are learned, and reconstruction of high-resolution clear images is achieved. The invention can effectively reduce the calculation amount of network parameters and simultaneously reserve the important detail information of the image. And inputting the original low-resolution image into a sparse depth convolution neural network model trained by a set algorithm, and realizing super-resolution reconstruction into a high-resolution image.
Description
Technical Field
The invention relates to the technical field of image reconstruction, in particular to a depth sparse convolution neural network face super-resolution image reconstruction method based on pooling manifold constraint, a system, a device and a storage medium thereof.
Background
In recent years, the country has entered the period of rapid development of society, people have increased their attention to the protection of personal privacy, property and personal safety, and in order to maintain social security and citizen safety, video monitoring systems are widely used in the fields of traffic and safety. In case of reconnaissance and tracking, police need to obtain clear video monitoring image quality so as to improve the rate of solving a case. However, in actual monitoring, due to the limitation of the resolution of the imaging device of the video monitoring system and the interference of factors such as the distance between the target imaging face and the camera, blur, noise, weather environment and the like, only a video imaging effect with small resolution and poor quality can be obtained, so that the identification of the target face becomes difficult. Therefore, it is very critical to restore a low-resolution image to a high-resolution image by a super-resolution reconstruction technique.
In the current mainstream super-resolution reconstruction algorithm, the application of the deep convolutional neural network embodies great advantages. The calculation performance of the deep convolutional neural network is mainly determined by the number of convolution multiplication operations in the network, and the application effect is limited due to the huge calculation amount of the convolutional neural network; meanwhile, considering that image detail information loss in the network data transmission process is inevitably caused by a pooling layer (namely, sampling operation is performed on image parameter data) in the network, a deep neural network structure using a full convolution layer is mostly selected, so that the image detail information is effectively reserved, and the super-resolution reconstruction effect is closer to the reality. However, once the deep neural network structure has no pooling layer, the image data is trained and tested in the form of full convolution layer, although the input-output characteristic diagram between the network layers is kept unchanged, the calculation amount of the neural network is increased, and the time consumption is increased, thereby causing negative effects. Although the existing research proposes that the number of multiplication operations in the convolutional neural network is reduced by utilizing the linear transformation of the neurons and the convolutional kernel in the convolutional operation or by modifying the network weight to compress the neural network, once the linear transformation is carried out on the neurons and the convolutional kernel, the network sparsity disappears, and the network acceleration cannot be carried out by utilizing the sparsity; the improved activation function is used in the full convolution network to reduce the calculation complexity between the neural network layers, but the occupied consumption of the network memory and the network parameter quantity are not reduced.
Disclosure of Invention
In order to solve the above problems, the present invention provides a super-resolution image reconstruction method, a system, an apparatus, and a storage medium thereof, which can effectively reduce the amount of network parameter calculation while retaining important detail information of an image, and input an original low-resolution image into a sparse deep convolutional neural network model trained by a set algorithm, thereby realizing super-resolution reconstruction into a high-resolution image.
The technical scheme adopted by the invention for solving the problems is as follows:
in a first aspect, an embodiment of the present invention provides a super-resolution image reconstruction method, including:
acquiring an original image;
carrying out bicubic interpolation on the original image, and enhancing the number of image pixels to obtain a preprocessed image;
carrying out image data analysis clustering and reconstruction on the preprocessed image through a sparse depth convolution neural network containing pooling manifold constraint;
and outputting the reconstructed image.
Further, the acquiring the original image includes:
and for the monitoring video, performing screenshot operation on the video by frames by using a Matlab writing program to obtain an original image data set.
Further, the bicubic interpolation calculation involves 16 pixel points, and the calculation formula is as follows:
wherein, (i ', j') represents the fractional pixel coordinates in the original image, dx, dy
The coordinates are respectively in the directions of the x axis and the y axis, and F (i ', j') represents the original image passing through
Calculating the sum of weight convolution between 16 pixel points nearest to each pixel point coordinate
R (x) represents an interpolation expression, and the calculation formula is as follows:
the number of image pixels can be enhanced by bi-cubic interpolation of the original image.
Further, the image data analysis clustering and reconstruction of the preprocessed images through the sparse deep convolution neural network containing the pooling manifold constraint comprises the following steps:
the preprocessed image is subjected to manifold function constrained pooling sampling, feature map extraction in a convolutional layer, data jump connection and convolutional neural network sparsification through two pairs of image data of a mode I and a mode II to respectively obtain an image A and an image B, and the image A and the image B are subjected to combined convolution operation to obtain a reconstructed image.
Further, the mode one includes:
the method comprises the steps of enabling a preprocessed image to pass through a pooling layer with manifold constraint, connecting and rearranging down-sampling copied feature map data to obtain a new feature map, neglecting an interpolation feature map generated by bicubic interpolation, inputting the interpolation feature map into a convolution and jump connection architecture with 10 layers, then inputting generated feature map into a next layer of convolution layer to obtain a vector with the size of (H/2, W/2, 4), then carrying out sparsification on a convolution neural network, and finally carrying out sub-pixel amplification operation on the obtained vector to sample up to a target resolution ratio to generate an image A with the size of (H, W).
Further, the second mode includes:
the preprocessed image is firstly input into a convolution layer with 64 filters to generate a feature map with the output size of (H, W, 64), and then input into a 10-layer convolution and jump link architecture, the feature map with the output size of (H, W, 64) is thinned through a network, and then an image B is generated through the convolution layer with a single filter.
In a second aspect, an embodiment of the present invention further provides a super-resolution image reconstruction system, including:
an image acquisition unit for acquiring an original image;
the image data enhancement unit is used for carrying out bicubic interpolation on the original image, enhancing the number of image pixels and obtaining a preprocessed image;
the image analysis clustering and reconstruction unit is used for carrying out image data analysis clustering and reconstruction on the preprocessed images through a sparse depth convolution neural network containing pooling manifold constraints;
an image output unit for outputting the reconstructed image.
In a third aspect, an embodiment of the present invention further provides a super-resolution image reconstruction apparatus, including:
at least one processor; and (c) a second step of,
a memory communicatively coupled to the at least one processor; wherein,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of the first aspect of the invention.
In a fourth aspect, the embodiment of the present invention further provides a computer-readable storage medium, where computer-executable instructions are stored, and the computer-executable instructions are configured to enable a computer to execute the method described in the first aspect of the present invention.
One or more technical schemes provided in the embodiment of the invention have at least the following beneficial effects: the invention provides a super-resolution image reconstruction method, a system, a device and a storage medium thereof.A primary low-resolution image is input, and bicubic interpolation is carried out on the input low-resolution image through image preprocessing; then, the preprocessed images are connected and arranged with each other through a sparse depth convolution neural network containing pooling manifold constraint, high-frequency detail information components missing in the images are learned, and reconstruction of high-resolution clear images is achieved. The invention can effectively reduce the calculation amount of network parameters and simultaneously reserve the important detail information of the image. And inputting the original low-resolution image into a sparse deep convolution neural network model trained by a set algorithm, and realizing super-resolution reconstruction into a high-resolution image. The method can keep the network sparsity, relieve the disappearance of important detail information caused by the sampling of data information by a pooling layer, and relieve the negative influence caused by the occupation consumption of network memory and the increase of parameter calculation amount caused by a full convolution deep neural network.
Drawings
The invention is further illustrated with reference to the following figures and examples.
FIGS. 1a and 1b are flow charts of a super-resolution image reconstruction method according to a first embodiment of the present invention;
FIG. 2 is a flowchart of super resolution reconstruction of an image based on a deep convolutional neural network according to a first embodiment of the present invention;
FIG. 3 is a diagram illustrating a convolution and jump-join process performed on image data according to a first embodiment of the present invention;
FIG. 4 is a flowchart of the network sparsification operation of the convolution and jump connection architecture for image data according to the first embodiment of the present invention;
FIG. 5 is a schematic diagram of a super-resolution image reconstruction system according to a second embodiment of the present invention;
fig. 6 is a schematic configuration diagram of a super-resolution image reconstruction apparatus according to a third embodiment of the present invention.
Detailed Description
Reference will now be made in detail to the present preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout.
In the description of the present invention, it should be understood that the orientation or positional relationship referred to in the description of the orientation, such as the upper, lower, front, rear, left, right, etc., is based on the orientation or positional relationship shown in the drawings, and is only for convenience of description and simplification of description, and does not indicate or imply that the device or element referred to must have a specific orientation, be constructed and operated in a specific orientation, and thus, should not be construed as limiting the present invention.
In the description of the present invention, the meaning of a plurality of means is one or more, the meaning of a plurality of means is two or more, and larger, smaller, larger, etc. are understood as excluding the number, and larger, smaller, inner, etc. are understood as including the number. If there is a description of first and second for the purpose of distinguishing technical features only, this is not to be understood as indicating or implying a relative importance or implicitly indicating the number of technical features indicated or implicitly indicating the precedence of technical features indicated.
In the description of the present invention, unless otherwise explicitly limited, terms such as arrangement, installation, connection and the like should be understood in a broad sense, and those skilled in the art can reasonably determine the specific meanings of the above terms in the present invention in combination with the specific contents of the technical solutions.
The embodiments of the present invention will be further explained with reference to the drawings.
Referring to fig. 1a and 1b, a first embodiment of the present invention provides a super-resolution image reconstruction method, including but not limited to the following steps:
s100, acquiring an original image;
s200, carrying out bicubic interpolation on the original image to enhance the number of image pixels and obtain a preprocessed image;
s300, carrying out image data analysis clustering and reconstruction on the preprocessed image through a sparse depth convolution neural network containing pooling manifold constraint;
and S400, outputting the reconstructed image.
The original data acquisition is realized through S100, and S100 specifically includes: and for the monitoring video, a Matlab programming program is used for carrying out screenshot operation on the video by frames to obtain a low-resolution image data set.
The image data enhancement and preprocessing are realized through S200, and the purpose is to emphasize some interesting features of an original image, inhibit uninteresting features, enrich image information quantity and enhance image identification effect. The specific function method is as follows:
the bicubic interpolation calculation involves 16 pixel points, (i ', j') represents the pixel coordinates of a fractional part in an original image, dx and dy are the coordinates in the directions of an x axis and a y axis respectively, and F (i ', j') represents a new pixel value calculated by the sum of weight convolutions between 16 pixel points closest to the coordinates of each pixel point in the original input image. R (x) represents an interpolation expression, because the super-resolution reconstruction of the face image is researched by the invention, the processing of the face image has certain complexity, a convolution sampling interpolation calculation based on a polynomial is adopted, and the formula is as follows:
bicubic interpolation of the original image can enhance the number of image pixels.
The method comprises the steps of S300, analyzing, clustering and reconstructing image data, inputting an image enhanced by the image data to perform super-resolution reconstruction based on a deep convolutional neural network, and mainly comprising two aspects of a pooling sampling process with manifold function constraint, extraction of feature maps in convolutional layers, data jump connection and convolution neural network sparseness for reconstructing a high-resolution image. The network structure is roughly shown in fig. 2.
The main process of the super-resolution reconstruction of the image data designed by the invention is as follows: after an originally input low-resolution image is subjected to data enhancement through bicubic interpolation, convolution and connection processing are performed on image data through two different modes, as shown in fig. 2, a feature diagram result image A and an image B processed through the two modes are subjected to combined convolution operation, a final super-resolution reconstruction image is output, and fig. 3 is a visual description of connection of convolution and jumping.
In the first mode, an image after bicubic interpolation data enhancement processing passes through a pooling layer with manifold constraint, connection and rearrangement between downsampling copied feature map data are carried out to obtain a new feature map, an interpolation feature map generated by bicubic interpolation is ignored and is input into a framework of convolution (including conversion convolution) and jump connection by 10 layers, the generated feature map is input into a convolution layer of the next layer to obtain a vector size of (H/2, W/2 and 4), then the convolution neural network is thinned, and finally the obtained vector is subjected to sub-pixel amplification operation to be up-sampled to a target resolution ratio to generate a high-quality super-resolution image A with the size of (H, W).
In the second mode, the image of the original low-resolution image after the bicubic interpolation processing is directly input into a 10-layer convolution and jump link architecture, and more image detail information can be provided for reconstructing a high-quality super-resolution image. As shown in the dotted line portion of fig. 2, after the bicubic interpolation of the low-resolution image, the feature map with the output size (H, W, 64) is first input to a convolution layer with 64 filters to generate a feature map with the output size (H, W, 64), and then input to the convolution and jump connection structure (with the same weight as the mode one) described above, the feature map with the output size (H, W, 64) is thinned through a network, and then the high-quality super-resolution image B is generated by the convolution layer with a single filter.
And combining the high-quality super-resolution images A and B generated by the two modes, and obtaining a final super-resolution reconstruction image through a convolution layer of a single filter.
Finally, outputting the high resolution image is achieved through S400.
The following is a detailed description of the main process:
because the original input low-resolution image contains more image detail information, and simultaneously, in order to avoid network memory occupation consumption and parameter calculation amount increase caused by a full convolution layer network architecture, the invention provides a pooling process with popular constraint, and the input feature map is sampled to a plurality of resolution reduction versions (without image detail information loss) by readjusting pixel positions.
And a manifold learning idea of local Linear Embedding (localization Linear Embedding) is adopted between the pooling layers, so that the detail information of the data in the pooling sampling process is not lost, and the network calculation amount is reduced. For example, after the samples are mapped from the high-dimensional space to the low-dimensional space, the linear relationship between the samples in each domain is not changed.
The coordinate of the sample point xi to be calculated is found by using the K neighbor idea in its original high-dimensional neighborhood, and three samples xj, xl, xk nearest to it are assumed that xi can be linearly represented by xj, xl, xk, that is:
x i =w ij x j +w ik x k +w il x l
wherein wij, wjk, wil are weight coefficients. After dimension reduction is carried out by LLE, the projections x 'i, x' j, x 'k corresponding to the projections x' i and xj, xl, xk of xi in the low-dimensional space keep the same linear relation as much as possible. Namely:
x' i ≈w ij x' j +w ik x' k +w il x' l
the weight parameters are kept as consistent as possible in the low-dimensional space and the high-dimensional space.
The algorithm can be divided into two parts, the first step is to calculate the field reconstruction coefficients w of all samples according to the field relation, namely to find out the linear relation between each sample and the samples in the field, and the formula is as follows:
wherein order C jk =(x i -x j ) T (x i -x k ),
The second step is to solve the coordinate of each sample in the low-dimensional space according to the invariance of the neighborhood reconstruction coefficient, and the formula is as follows:
let C jk =(x i -x j ) T z=(z 1 ,z 2 ,...,z m )∈R d'×m ,(W) ij =w ij ,M=(I-W) T+1
In the convolutional layer, a self-adaptive sample selection method is added to realize the manifold analysis and clustering of the local structure of the face, and the self-adaptive sample selection method is used for recovering the missing high-frequency information in the low-frequency image block.
Extracting a characteristic image block pk (k =1, \8230;, n) at a position corresponding to the d-th layer convolution layer, constructing a set Pi = [ p1, p2, \8230;, pn ] in a column vector mode, and calculating an LPP transformation matrix Ai and a mapping data matrix Yi = AiTPi of Pi. For image blocks of an input low-resolution image (LR), a low-dimensional feature set of the image blocks is calculated, and then image block pairs corresponding to a set closest to the low-dimensional features are selected in a mapping data matrix Yi according to Euclidean distances to serve as a training set for recovering high-frequency information missing in the low-frequency image blocks.
The network sparsification operation following the convolution and jump join architecture leaves the sparsity of the weights and activation functions preserved and significantly reduces the computational effort before the multiplication operation is performed, as shown in figure 4. The method is carried out in three stages: intensive training, pruning and retraining. Intensive training: we train a dense p × p kernel directly in the transform domain. Inverse transform-elimination, which directly initializes and trains transformed kernels by back propagation, requiring preservation of the kernels in the spatial domain or transformation of the spatial kernels; trimming: pruning the transformed kernel by calculating a threshold t required for realizing the required pruning rate r and resetting the ownership with an absolute value less than t to zero; and (3) retraining: keeping the clipped weight zero, the sparse mask is calculated during the clipping step and constant retraining is maintained during this period. The formulation is as follows:
S=A T [[Prune(GgG T )]⊙[ReLU(B T dB)]]A
where GgGT L is the weight gradient and dL is the calculated gradient of the input activation.
And finally, defining a residual image as r = Y-x in order to enable the reconstructed high-resolution image to be close to the real image Y of the original face database to the maximum extent, wherein x represents a low-resolution image, and Y represents a corresponding high-resolution image in the face database. Inputting residual estimation and low-resolution images at a loss layer, corresponding to the images of the face database, and adopting a mean square error function as a judgment loss value, wherein the mean square error function is expressed as:
where n is the number of samples, θ = { W1, W2, \8230;, wd, B1, B2, \8230;, bd } are network parameters.
In the whole network architecture, a ReLU activation function is used, in order to determine the nonlinear characteristics of the function and the whole network, the neural network can be arbitrarily approximated to any nonlinear function, and all negative conversions are activated to return to zero in network sparseness, so that the multiplication times in a domain are reduced. The formula is as follows:
F(x)=max(0,x)
and reconstructing the low-resolution image through a depth convolution neural network to obtain a corresponding high-resolution image output.
In addition, referring to fig. 5, a second embodiment of the present invention provides a super-resolution image reconstruction system including:
an image acquisition unit 110 for acquiring an original image;
the image data enhancement unit 120 is configured to perform bicubic interpolation on the original image, enhance the number of image pixels, and obtain a preprocessed image;
the image analysis clustering and reconstruction unit 130 is used for performing image data analysis clustering and reconstruction on the preprocessed image through a sparse depth convolution neural network containing pooling manifold constraint;
an image output unit 140 for outputting the reconstructed image.
The super-resolution image reconstruction system in the present embodiment is based on the same inventive concept as the super-resolution image reconstruction method in the first embodiment, and therefore, the super-resolution image reconstruction system in the present embodiment has the same advantageous effects: the image acquisition unit 110 acquires an original image; the image data enhancement unit 120 performs bicubic interpolation on the original image to enhance the number of image pixels, so as to obtain a preprocessed image; the image analysis clustering and reconstruction unit 130 performs image data analysis clustering and reconstruction on the preprocessed images through a sparse depth convolution neural network containing pooling manifold constraint; the image output unit 140 outputs the reconstructed image. The super-resolution image reconstruction system can effectively reduce the calculated amount of network parameters and simultaneously reserve important detail information of images. And inputting the original low-resolution image into a sparse depth convolution neural network model trained by a set algorithm, and realizing super-resolution reconstruction into a high-resolution image.
Referring to fig. 6, the third embodiment of the present invention also provides a super-resolution image reconstruction apparatus including:
at least one processor;
and a memory communicatively coupled to the at least one processor;
wherein the memory stores instructions executable by the at least one processor to enable the at least one processor to perform a super-resolution image reconstruction method as in any one of the first embodiments above.
The apparatus 200 may be any type of smart terminal, such as a cell phone, a tablet, a personal computer, etc.
The processor and memory may be connected by a bus or other means, such as by a bus in FIG. 6.
The memory, which is a non-transitory computer-readable storage medium, may be used to store non-transitory software programs, non-transitory computer-executable programs, and modules, such as program instructions/modules corresponding to the face recognition model construction method in the embodiments of the present invention. The processor executes various functional applications and data processing of the apparatus 200 by executing non-transitory software programs, instructions and modules stored in the memory, that is, implements the super-resolution image reconstruction method of any one of the above-described method embodiments.
The memory may include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function; the storage data area may store data created according to the use of the device 200, and the like. Further, the memory may include high speed random access memory, and may also include non-transitory memory, such as at least one disk storage device, flash memory device, or other non-transitory solid state storage device. In some embodiments, the memory optionally includes memory located remotely from the processor, and such remote memory may be coupled to the device 200 via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The one or more modules are stored in the memory and when executed by the one or more processors perform the super-resolution image reconstruction method of any of the method embodiments described above, e.g., perform method steps S100 to S400 of the first embodiment described above.
The fourth embodiment of the present invention also provides a computer-readable storage medium storing computer-executable instructions, which are executed by one or more control processors, for example, by one of the processors in fig. 6, and can cause the one or more processors to execute a super-resolution image reconstruction method in the above-described method embodiment, for example, the method steps S100 to S400 in the first embodiment.
The above-described embodiments of the apparatus are merely illustrative, and the units described as separate parts may or may not be physically separate, may be located in one place, or may be distributed over a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment.
Through the above description of the embodiments, those skilled in the art will clearly understand that each embodiment can be implemented by software plus a general hardware platform, and certainly can also be implemented by hardware. It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by hardware related to instructions of a computer program, which can be stored in a computer readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. The storage medium may be a magnetic disk, an optical disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), or the like.
While the preferred embodiments of the present invention have been described, the present invention is not limited to the above embodiments, and those skilled in the art can make various equivalent modifications or substitutions without departing from the spirit of the present invention, and such equivalent modifications or substitutions are to be included within the scope of the present invention defined by the claims.
Claims (6)
1. A super-resolution image reconstruction method is characterized by comprising the following steps:
acquiring an original image;
carrying out bicubic interpolation on the original image, and enhancing the number of image pixels to obtain a preprocessed image;
carrying out image data analysis clustering and reconstruction on the preprocessed image through a sparse depth convolution neural network containing pooling manifold constraint;
outputting a reconstructed image;
the image data analysis clustering and reconstruction of the preprocessed image through the sparse depth convolution neural network containing the pooling manifold constraint comprise the following steps: performing manifold function constrained pooling sampling on preprocessed image data through a mode I and a mode II, extracting feature maps in a convolutional layer, performing data jump connection and convolution neural network sparsification to respectively obtain an image A and an image B, and performing combined convolution operation on the image A and the image B to obtain a reconstructed image;
wherein the first mode comprises:
preprocessing an image, passing through a pooling layer with manifold constraint, performing connection and rearrangement between downsampling copy feature map data to obtain a new feature map, neglecting an interpolation feature map generated by bicubic interpolation, inputting the interpolation feature map into a convolution and jump connection architecture with 10 layers, then inputting the generated feature map into a next layer of convolution layer to obtain a vector with the size of (H/2, W/2, 4), then performing sparsification on a convolution neural network, and finally performing sub-pixel amplification operation on the obtained vector to obtain an up-sampling target resolution to generate an image A with the size of (H, W), wherein H is height and W is width;
the second mode includes: the preprocessed image is firstly input into a convolution layer with 64 filters to generate a feature map with the output size of (H, W, 64), and then input into a 10-layer convolution and jump link architecture, the feature map with the output size of (H, W, 64) is thinned through a network, and then an image B is generated through the convolution layer with a single filter.
2. The super-resolution image reconstruction method according to claim 1, wherein the acquiring the original image comprises:
and for the monitoring video, performing screenshot operation on the video by frames by using a Matlab writing program to obtain an original image data set.
3. The super-resolution image reconstruction method according to claim 1, wherein the bicubic interpolation calculation involves 16 pixel points, and the calculation formula is as follows:
wherein, (i ', j') represents the fractional part pixel coordinates in the original image, dx and dy are the coordinates in the directions of the x axis and the y axis respectively, m and n are preset constants, F (i ', j') represents a new pixel value calculated by the sum of weight convolutions between 16 pixel points closest to the coordinates of each pixel point in the original image, R (x) represents an interpolation expression, and the calculation formula is as follows:
4. a super-resolution image reconstruction system applied to the super-resolution image reconstruction method according to any one of claims 1 to 3, comprising:
an image acquisition unit for acquiring an original image;
the image data enhancement unit is used for carrying out bicubic interpolation on the original image, enhancing the number of image pixels and obtaining a preprocessed image;
the image analysis clustering and reconstruction unit is used for carrying out image data analysis clustering and reconstruction on the preprocessed image through a sparse depth convolution neural network containing pooling manifold constraint;
an image output unit for outputting the reconstructed image.
5. A super-resolution image reconstruction apparatus, comprising:
at least one processor; and the number of the first and second groups,
a memory communicatively coupled to the at least one processor; wherein,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1-3.
6. A computer-readable storage medium having stored thereon computer-executable instructions for causing a computer to perform the method of any one of claims 1-3.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910409645.3A CN110246084B (en) | 2019-05-16 | 2019-05-16 | Super-resolution image reconstruction method, system and device thereof, and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910409645.3A CN110246084B (en) | 2019-05-16 | 2019-05-16 | Super-resolution image reconstruction method, system and device thereof, and storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110246084A CN110246084A (en) | 2019-09-17 |
CN110246084B true CN110246084B (en) | 2023-03-31 |
Family
ID=67884554
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910409645.3A Active CN110246084B (en) | 2019-05-16 | 2019-05-16 | Super-resolution image reconstruction method, system and device thereof, and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110246084B (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112581363A (en) * | 2019-09-29 | 2021-03-30 | 北京金山云网络技术有限公司 | Image super-resolution reconstruction method and device, electronic equipment and storage medium |
CN110942425A (en) * | 2019-11-26 | 2020-03-31 | 贵州师范学院 | Reconstruction method and reconstruction system of super-resolution image and electronic equipment |
CN111222446B (en) * | 2019-12-31 | 2023-05-16 | Oppo广东移动通信有限公司 | Face recognition method, face recognition device and mobile terminal |
CN113409192A (en) * | 2021-06-17 | 2021-09-17 | Oppo广东移动通信有限公司 | Super-resolution chip, super-resolution algorithm updating method and electronic equipment |
CN117643059A (en) * | 2021-07-01 | 2024-03-01 | 抖音视界有限公司 | Super resolution downsampling |
CN114170619B (en) * | 2021-10-18 | 2022-08-19 | 中标慧安信息技术股份有限公司 | Data checking method and system based on edge calculation |
CN113675850B (en) * | 2021-10-25 | 2022-02-08 | 山东大学 | Power grid information rapid and accurate sensing method based on nonlinear robust estimation |
CN115546858B (en) * | 2022-08-15 | 2023-08-25 | 荣耀终端有限公司 | Face image processing method and electronic equipment |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106709875A (en) * | 2016-12-30 | 2017-05-24 | 北京工业大学 | Compressed low-resolution image restoration method based on combined deep network |
CN107480726A (en) * | 2017-08-25 | 2017-12-15 | 电子科技大学 | A kind of Scene Semantics dividing method based on full convolution and shot and long term mnemon |
CN107507134A (en) * | 2017-09-21 | 2017-12-22 | 大连理工大学 | Super-resolution method based on convolutional neural networks |
CN109461157A (en) * | 2018-10-19 | 2019-03-12 | 苏州大学 | Image, semantic dividing method based on multi-stage characteristics fusion and Gauss conditions random field |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TWI624804B (en) * | 2016-11-07 | 2018-05-21 | 盾心科技股份有限公司 | A method and system for providing high resolution image through super-resolution reconstrucion |
FR3059804B1 (en) * | 2016-12-07 | 2019-08-02 | Idemia Identity And Security | IMAGE PROCESSING SYSTEM |
US10616482B2 (en) * | 2017-03-10 | 2020-04-07 | Gopro, Inc. | Image quality assessment |
US10979718B2 (en) * | 2017-09-01 | 2021-04-13 | Apple Inc. | Machine learning video processing systems and methods |
-
2019
- 2019-05-16 CN CN201910409645.3A patent/CN110246084B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106709875A (en) * | 2016-12-30 | 2017-05-24 | 北京工业大学 | Compressed low-resolution image restoration method based on combined deep network |
CN107480726A (en) * | 2017-08-25 | 2017-12-15 | 电子科技大学 | A kind of Scene Semantics dividing method based on full convolution and shot and long term mnemon |
CN107507134A (en) * | 2017-09-21 | 2017-12-22 | 大连理工大学 | Super-resolution method based on convolutional neural networks |
CN109461157A (en) * | 2018-10-19 | 2019-03-12 | 苏州大学 | Image, semantic dividing method based on multi-stage characteristics fusion and Gauss conditions random field |
Non-Patent Citations (1)
Title |
---|
基于中间层监督卷积神经网络的图像超分辨率重建;李现国等;《中国图象图形学报》;20180716(第07期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN110246084A (en) | 2019-09-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110246084B (en) | Super-resolution image reconstruction method, system and device thereof, and storage medium | |
US12008797B2 (en) | Image segmentation method and image processing apparatus | |
CN112446270B (en) | Training method of pedestrian re-recognition network, pedestrian re-recognition method and device | |
US12062158B2 (en) | Image denoising method and apparatus | |
CN111683269B (en) | Video processing method, video processing device, computer equipment and storage medium | |
CN112308200B (en) | Searching method and device for neural network | |
CN111914997B (en) | Method for training neural network, image processing method and device | |
CN110599401A (en) | Remote sensing image super-resolution reconstruction method, processing device and readable storage medium | |
CN113658040B (en) | Human face super-resolution method based on priori information and attention fusion mechanism | |
CN111476719A (en) | Image processing method, image processing device, computer equipment and storage medium | |
CN112446835B (en) | Image restoration method, image restoration network training method, device and storage medium | |
JP7357176B1 (en) | Night object detection, training method and device based on self-attention mechanism in frequency domain | |
CN111951195A (en) | Image enhancement method and device | |
CN109064402B (en) | Single image super-resolution reconstruction method based on enhanced non-local total variation model prior | |
CN111951165A (en) | Image processing method, image processing device, computer equipment and computer readable storage medium | |
WO2021057091A1 (en) | Viewpoint image processing method and related device | |
CN116452930A (en) | Multispectral image fusion method and multispectral image fusion system based on frequency domain enhancement in degradation environment | |
CN115409697A (en) | Image processing method and related device | |
CN117593187A (en) | Remote sensing image super-resolution reconstruction method based on meta-learning and transducer | |
Liu et al. | CNN-Enhanced graph attention network for hyperspectral image super-resolution using non-local self-similarity | |
CN111861877A (en) | Method and apparatus for video hyper-resolution | |
CN113191947B (en) | Image super-resolution method and system | |
Mun et al. | Universal super-resolution for face and non-face regions via a facial feature network | |
CN113256556A (en) | Image selection method and device | |
CN117314756B (en) | Verification and protection method and device based on remote sensing image, computer equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |