CN117408921A - Image complement method based on generation countermeasure network inversion reasoning and self-encoder - Google Patents
Image complement method based on generation countermeasure network inversion reasoning and self-encoder Download PDFInfo
- Publication number
- CN117408921A CN117408921A CN202311459854.1A CN202311459854A CN117408921A CN 117408921 A CN117408921 A CN 117408921A CN 202311459854 A CN202311459854 A CN 202311459854A CN 117408921 A CN117408921 A CN 117408921A
- Authority
- CN
- China
- Prior art keywords
- image
- encoder
- self
- network
- reasoning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 38
- 230000000295 complement effect Effects 0.000 title claims abstract description 37
- 238000012549 training Methods 0.000 claims description 16
- 238000012545 processing Methods 0.000 claims description 13
- 238000003062 neural network model Methods 0.000 claims description 12
- 230000006870 function Effects 0.000 claims description 9
- 230000004913 activation Effects 0.000 claims description 6
- 238000013507 mapping Methods 0.000 claims description 4
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 3
- 230000008485 antagonism Effects 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 230000009466 transformation Effects 0.000 claims description 3
- 238000005065 mining Methods 0.000 abstract description 4
- 238000013473 artificial intelligence Methods 0.000 abstract description 2
- 230000000694 effects Effects 0.000 description 4
- 238000002474 experimental method Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 238000009826 distribution Methods 0.000 description 2
- BASFCYQUMIYNBI-UHFFFAOYSA-N platinum Chemical compound [Pt] BASFCYQUMIYNBI-UHFFFAOYSA-N 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000008439 repair process Effects 0.000 description 2
- RTAQQCXQSZGOHL-UHFFFAOYSA-N Titanium Chemical compound [Ti] RTAQQCXQSZGOHL-UHFFFAOYSA-N 0.000 description 1
- 230000003042 antagnostic effect Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 238000012733 comparative method Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000000873 masking effect Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000008832 photodamage Effects 0.000 description 1
- 229910052697 platinum Inorganic materials 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0475—Generative networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/094—Adversarial learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/04—Inference or reasoning models
- G06N5/042—Backward inferencing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention discloses an image complement method based on generation countermeasure network inversion reasoning and a self-encoder, which belongs to the field of artificial intelligence. Second, the closest potential codes are inferred using a learning-based GAN inversion, and a complete image is reconstructed using the trained GAN. Finally, the quantitative comparison shows that the invention has higher complement quality when the image has large-area damage. The image complement method based on the generated countermeasure network inversion reasoning and the self-encoder is good at mining information from the inside of the image, and can generate more semantic results with the help of the generated countermeasure network model which is trained in advance, so that more vivid and reasonable high-resolution image results are generated.
Description
Technical Field
The invention relates to the technical field of artificial intelligence, in particular to an image complement method based on generation countermeasure network inversion reasoning and a self-encoder.
Background
In daily life, images play a vital information transfer role, and ensuring that they are intact is vital to the integrity of the information transfer. However, image files may be subject to breakage or masking for various reasons, which can seriously affect the implementation of computer vision tasks, such as photo damage or monitoring that the image content is occluded. To ensure complete transfer of image information, scientists have proposed a number of approaches in the field of image processing and computer vision, where image completion is a key research point. The image complement method mainly utilizes the computer vision technology to complement the missing information, so as to restore the content of the original image. In this way, subsequent image analysis and processing tasks can be more accurate and reliable.
The current image complement methods can be broadly divided into two main categories, namely, a traditional method and a deep learning-based method. The traditional method comprises the following steps: structural-based image complement, texture-based image complement, sparse representation-based image complement, and the like. Structural repair methods are implemented by using partial differential equations, however, the robustness of these algorithms is poor, and uncertainty problems such as unclear after repair may occur. The texture-based method uses existing range texture information to construct the missing information part, and effectively solves the problem of unclear full range. However, it should be noted that this method is limited in the ability to acquire high-level semantic information, and thus is poor in processing high-resolution images of complex textures and the like. Based on the sparse representation method, the method can effectively represent the known information of the image. However, when the range in which the completion is required is excessively large, limited by the limitation of the known information, this method still has an undesirable situation in terms of the completion effect.
Disclosure of Invention
It is an object of the present invention to provide an image complement method based on generation of an antagonistic network inversion reasoning and self-encoder to solve the above mentioned problems in the background art.
In order to achieve the above object, the present invention provides an image complement method based on generation of an antagonism network inversion reasoning and self-encoder, comprising the steps of:
s1, extracting an online public data set by using a Python code and performing unified image transformation processing to obtain a damaged image data set;
s2, constructing an inversion reasoning and self-encoder model based on a generated type countermeasure network, and training by using a damaged image data set to obtain a trained neural network model;
s3, obtaining a damaged image to be complemented, carrying out normalization processing and inputting the damaged image to a trained neural network model to obtain a complement result of the picture.
Preferably, the step S1 specifically includes:
s11, acquiring various public data sets through a Kaggle official network and a Github;
s12, writing a Python program, performing pre-data processing on the data set, and performing image processing to ensure that the data set can be used in the model training process.
Preferably, the step S2 specifically includes:
s21, designing a self-encoder-based generation type countermeasure network, learning a mapping from random noise to a low-dimensional feature map through a generator of a training generation type countermeasure network, and converting the feature map generated by the generator into a high-definition resolution image through an encoder;
s22, adopting a learning-based inversion strategy of the generated type countermeasure network, fixing the trained generated type countermeasure network, designing and training an encoder network, and predicting the closest potential encoding knowledge from a given damaged image.
Preferably, the generator and the arbiter in the step S21 are implemented following the widely used deep convolution generating type countermeasure network DC-GAN.
Preferably, the encoder for reducing the dimension in step S21 is composed of 3 downsampling layers, and all the encoder is composed of 4×4 convolution kernels, the number of the convolution kernels is 64, 128, 16, and the activation functions are LeakyReLU, leakyReLU, tanh; the decoder for the dimension increase consists of 3 upsampling layers, the convolution kernels are 4×4 in size, the number of convolution kernels is 128, 64, 3, and ReLU, reLU, tanh is adopted as the activation function.
Preferably, the encoder network structure in the step S22 has a total of 9 layers, the first 7 layers adopt convolution layers, and the second 2 layers are implemented by using full connection layers.
Preferably, a small-batch gradient descent algorithm is adopted to update model parameters during neural network model training optimization, the batch_size is set to be 4, any loss function is optimized by using an Adam optimizer, and the learning rate is set to be 0.0001 to finely adjust the network.
Therefore, the image complement method based on the generation countermeasure network inversion reasoning and the self-encoder has the following beneficial effects:
(1) Building and training a neural network model, and based on the trained neural network model, the experimental qualitative result shows that the generated image has more reasonable structure and more real texture detail;
(2) The invention is better at mining information from the inside of the image, and can generate more semantic results with the help of the generated type countermeasure network model which is trained in advance, thereby generating more vivid and more reasonable high-resolution image results;
(3) The invention has higher complement quality when the complement task involves a large area of absence.
The technical scheme of the invention is further described in detail through the drawings and the embodiments.
Drawings
FIG. 1 is a block flow diagram of a process performed by an embodiment of the present invention;
FIG. 2 is a schematic diagram of a network structure in an embodiment of the present invention;
FIG. 3 is a neural network training diagram in an embodiment of the invention;
FIG. 4 is a graph of the qualitative complementation results of the present invention during an experiment;
FIG. 5 is a graph of the quantitative complementation results of the present invention during an experiment.
Detailed Description
Examples
The following description of the embodiments of the present invention will be made clearly and fully with reference to the accompanying drawings, in which it is evident that the embodiments described are only some, but not all embodiments of the invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Referring to fig. 1-3, the invention discloses an image complement method based on generation of an antagonism network inversion reasoning and self-encoder, which comprises the following steps:
s1, extracting an online public data set by using a Python code and performing unified image transformation processing to obtain a damaged image data set;
s2, constructing an inversion reasoning and self-encoder model based on a generated type countermeasure network, and training by using a damaged image data set to obtain a trained neural network model;
s3, obtaining a damaged image to be complemented, carrying out normalization processing and inputting the damaged image to a trained neural network model to obtain a complement result of the picture.
The data set used in this example totals 130000 sheets. From this 10% was taken as test set and the rest as training set. The experimental environment of the invention: python version Python3.6, trained using the open source Python machine learning library pythorch1.10.0 in combination with cuda 11.3. The model parameters during training and optimizing the neural network model are updated by adopting a small-batch gradient descent algorithm in all experiments, the batch_size is set to be 4, all loss functions are optimized by using an Adam optimizer, and the learning rate is set to be 0.0001 for fine tuning the network until the network converges.
Training hardware conditions: the GPU model is NVIDIA TITAN Xp GPU, and the video memory is 12GB; the CPU model is 12 cores Intel (R) Xeon (R) Platinum 8255.
Then constructing a neural network model, wherein the model is formed by two stages, firstly designing a self-encoder-based generation type countermeasure network, the network learns the mapping from noise to a low-dimensional feature map through training a generator of the generation type countermeasure network, and then converting the feature map generated by the generator into a high-definition resolution image, so that noise distribution is mapped to a high-resolution real image, the learning difficulty of a mapping relation is reduced, and a generator and a discriminator of the generation type countermeasure network are realized by following a widely used deep convolution generation type countermeasure network DC-GAN. The encoder ER for dimension reduction consists of 3 downsampling layers, all adopt 4×4 convolution kernels, the number of the convolution kernels is 64, 128, 16 respectively, and the activation functions adopt LeakyReLU, leakyReLU, tanh respectively. The decoder for increasing the dimension consists of 3 upsampling layers, the convolution kernels are 4 multiplied by 4, the number of the convolution kernels is 128, 64 and 3 respectively, and ReLU, reLU, tanh is adopted as an activation function respectively; secondly, we use a learning-based generated-type countermeasure network inversion strategy to design an encoder to infer the closest potential encoding knowledge from the corrupted image, and then reconstruct the complete image using the trained generated-type countermeasure network. The encoder model architecture in the inversion strategy is shown in table 1:
table 1 encoder structure
To qualitatively evaluate the effect of each aspect of the model, the present model was compared to several models of Pconv, gated Conv, and comdgan. Fig. 4 is a qualitative visual comparison of the various models, and it can be seen that the results produced by both Pconv, gated Conv methods include partially distorted content, certain artifact effects and color differences, and the results and performance of the comdgan model are similar to those of the present method, but it generally produces discordant content and unmasked regions. The completion method is more good at mining information from the inside of the image, and can generate more semantic results with the help of the generated type countermeasure network model which is trained in advance, so that different mask areas can be better processed, and more vivid and more reasonable high-resolution image results are generated.
To integrate the effects of the various aspects of the comparative model, structural Similarity (SSIM), peak signal-to-noise ratio (PSNR), frechet Inception Distance (FID value) were used as experimental evaluation indexes.
Structural similarity the similarity degree of two images is evaluated according to the brightness, structure and contrast of the images, the value range is [0,1], and the SSIM value is more prone to 1, which shows that the higher the similarity of the two images is, the better the image complement quality is.
Peak signal-to-noise ratio is that scientific researchers calculate the difference of pixels of two images by adopting the widest measuring method to evaluate the image complement quality, and the larger the value is, the more true the sample of the generated image is, so the better the complement quality is.
Frechet Inception Distance is an index for measuring distances between multidimensional variable distributions, and is particularly suitable for evaluating the diversity and quality of a generated image. The smaller the Frechet Inception Distance value, the more rich the diversity of the generated image, and the higher the quality of the image.
The method is superior to the comparative method in terms of 3 indexes of Structural Similarity (SSIM), peak signal-to-noise ratio (PSNR) and Frechet Inception Distance (FID), and the method shows extremely excellent performance, which proves that the method has higher complement quality when the complement task involves a large area of missing.
The quantitative results of the experiment are shown in FIG. 5.
It should be noted that, parts not described in detail in this application are all prior art.
Therefore, the image complement method based on the generated countermeasure network inversion reasoning and the self-encoder is good at mining information from the inside of the image, and can generate more semantic results with the help of the generated countermeasure network model which is trained in advance, so that more realistic and reasonable high-resolution image results are generated, and the image complement method has higher complement quality when the complement task relates to large-area missing.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present invention and not for limiting it, and although the present invention has been described in detail with reference to the preferred embodiments, it will be understood by those skilled in the art that: the technical scheme of the invention can be modified or replaced by the same, and the modified technical scheme cannot deviate from the spirit and scope of the technical scheme of the invention.
Claims (7)
1. An image complement method based on generation of an antagonism network inversion reasoning and self-encoder is characterized by comprising the following steps:
s1, extracting an online public data set by using a Python code and performing unified image transformation processing to obtain a damaged image data set;
s2, constructing an inversion reasoning and self-encoder model based on a generated type countermeasure network, and training by using a damaged image data set to obtain a trained neural network model;
s3, obtaining a damaged image to be complemented, carrying out normalization processing and inputting the damaged image to a trained neural network model to obtain a complement result of the picture.
2. The method for generating image complement for countering network inversion reasoning and self-encoder according to claim 1, wherein said step S1 specifically comprises:
s11, acquiring various public data sets through a Kaggle official network and a Github;
s12, writing a Python program, performing pre-data processing on the data set, and performing image processing.
3. The method for generating image complement for countering network inversion reasoning and self-encoder according to claim 1, wherein said step S2 specifically comprises:
s21, designing a self-encoder-based generation type countermeasure network, learning a mapping from random noise to a low-dimensional feature map through a generator of a training generation type countermeasure network, and converting the feature map generated by the generator into a high-definition resolution image through an encoder;
s22, adopting a learning-based inversion strategy of the generated type countermeasure network, fixing the trained generated type countermeasure network, designing and training an encoder network, and predicting approaching potential encoding knowledge from a given damaged image.
4. The image complement method based on generating countermeasure network inversion reasoning and self-encoder as recited in claim 3, wherein: the generator and the arbiter in said step S21 follow the deep convolution generating type countermeasure network DC-GAN.
5. The method of generating image complement for countering network inversion reasoning and self-encoder as recited in claim 4, wherein: the encoder for reducing the dimension in the step S21 consists of 3 downsampling layers, 4×4 convolution kernels are adopted, the number of the convolution kernels is 64, 128 and 16, and the activation functions are LeakyReLU, leakyReLU, tanh; the decoder for the dimension increase consists of 3 upsampling layers, the convolution kernels are 4×4 in size, the number of convolution kernels is 128, 64, 3, and ReLU, reLU, tanh is adopted as the activation function.
6. The image complement method based on generating countermeasure network inversion reasoning and self-encoder as recited in claim 3, wherein: the encoder network structure in the step S22 has a 9-layer structure, the first 7 layers adopt convolution layers, and the second 2 layers are realized by using full connection layers.
7. The image complement method based on generating countermeasure network inversion reasoning and self-encoder of claim 1, wherein: and updating model parameters during training and optimizing the neural network model by adopting a small-batch gradient descent algorithm, setting batch_size to be 4, optimizing any loss function by using an Adam optimizer, and setting learning rate to be 0.0001 to finely adjust the network.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311459854.1A CN117408921A (en) | 2023-11-03 | 2023-11-03 | Image complement method based on generation countermeasure network inversion reasoning and self-encoder |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311459854.1A CN117408921A (en) | 2023-11-03 | 2023-11-03 | Image complement method based on generation countermeasure network inversion reasoning and self-encoder |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117408921A true CN117408921A (en) | 2024-01-16 |
Family
ID=89495959
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311459854.1A Pending CN117408921A (en) | 2023-11-03 | 2023-11-03 | Image complement method based on generation countermeasure network inversion reasoning and self-encoder |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117408921A (en) |
-
2023
- 2023-11-03 CN CN202311459854.1A patent/CN117408921A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109919204B (en) | Noise image-oriented deep learning clustering method | |
CN108875935B (en) | Natural image target material visual characteristic mapping method based on generation countermeasure network | |
CN109035142B (en) | Satellite image super-resolution method combining countermeasure network with aerial image prior | |
CN109671022A (en) | A kind of picture texture enhancing super-resolution method based on depth characteristic translation network | |
CN112598602A (en) | Mask-based method for removing Moire of deep learning video | |
CN114898159B (en) | SAR image interpretability feature extraction method for generating countermeasure network based on decoupling characterization | |
CN113920210B (en) | Image low-rank reconstruction method based on adaptive graph learning principal component analysis method | |
CN114926883A (en) | Face image processing method meeting various degradation models | |
CN112965968B (en) | Heterogeneous data pattern matching method based on attention mechanism | |
CN112560719B (en) | High-resolution image water body extraction method based on multi-scale convolution-multi-core pooling | |
CN116884067B (en) | Micro-expression recognition method based on improved implicit semantic data enhancement | |
CN111695507B (en) | Static gesture recognition method based on improved VGGNet network and PCA | |
CN116934613A (en) | Branch convolution channel attention module for character repair | |
CN117408921A (en) | Image complement method based on generation countermeasure network inversion reasoning and self-encoder | |
CN117196963A (en) | Point cloud denoising method based on noise reduction self-encoder | |
Jiang et al. | AGP-Net: Adaptive graph prior network for image denoising | |
CN115358952A (en) | Image enhancement method, system, equipment and storage medium based on meta-learning | |
CN111681156B (en) | Deep compressed sensing image reconstruction method applied to wireless sensor network | |
CN114092579A (en) | Point cloud compression method based on implicit neural network | |
CN113269282A (en) | Unsupervised image classification method based on automatic encoder | |
CN113269702A (en) | Low-exposure vein image enhancement method based on cross-scale feature fusion | |
CN112907456A (en) | Deep neural network image denoising method based on global smooth constraint prior model | |
CN112686807A (en) | Image super-resolution reconstruction method and system | |
CN114627196B (en) | Latent variable space decoupling method based on variational automatic encoder | |
CN113313153B (en) | Low-rank NMF image clustering method and system based on self-adaptive graph regularization |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |