CN115546338A - Image coloring method based on Transformer and generation countermeasure network - Google Patents

Image coloring method based on Transformer and generation countermeasure network Download PDF

Info

Publication number
CN115546338A
CN115546338A CN202211247125.5A CN202211247125A CN115546338A CN 115546338 A CN115546338 A CN 115546338A CN 202211247125 A CN202211247125 A CN 202211247125A CN 115546338 A CN115546338 A CN 115546338A
Authority
CN
China
Prior art keywords
color image
image
network
vector
transformer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211247125.5A
Other languages
Chinese (zh)
Inventor
薛涛
马鹏森
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Polytechnic University
Original Assignee
Xian Polytechnic University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Polytechnic University filed Critical Xian Polytechnic University
Priority to CN202211247125.5A priority Critical patent/CN115546338A/en
Publication of CN115546338A publication Critical patent/CN115546338A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/001Texturing; Colouring; Generation of texture or colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Multimedia (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Computer Graphics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method for coloring images based on a Transformer and a generation countermeasure network, which solves the problem of coloring images by using the generation countermeasure network GAN and the Transformer instead of simply using a convolutional neural network CNN. The local enhanced forward propagation network and the hopping connection ensure that shallow features can be efficiently transmitted and utilized in the network so that the transform-GAN can efficiently capture the correlation between global and local information. An optimal training process is also explored through data enhancement and objective function selection, and a color image generator and a discriminator are formed to enable the Transformer-GAN to perform well in the aspect of image colorization. The best visual effect is achieved.

Description

Image coloring method based on Transformer and generation countermeasure network
Technical Field
The invention belongs to the technical field of image processing, and relates to an image coloring method based on a Transformer and a generation countermeasure network.
Background
In the image rendering task, our goal is to generate a color image from an input grayscale image. By category, from earlier conventional non-hopping connection algorithms based on CNN structures, networks of user-specified image colors (which require the user to enter color values in a particular layer) have emerged in the future. And end-to-end feed-forward architecture using animated image colorization that generates a countermeasure network (GAN), in addition to infrared colorization, radar image colorization, etc. for specific areas, and multi-modal coloring models (text-based coloring networks) that emerge later. Diversified colorization networks compensate for the lack of diversity by generating different color images, network architectures that incorporate multi-path networks that learn different characteristics in different network paths or levels, and users give reference images as input samples for the colored network. All the above models have in common that they are networks based on convolutional neural networks CNN, however unlike previous work i constructed image colorization networks using transformers and generation of confrontation networks GAN, which, to our knowledge, was the first study of image colorization using transformers as the main network.
Disclosure of Invention
The invention aims to provide an image coloring method based on a Transformer and a generation countermeasure network, which solves the problems of poor coloring effect and poor coloring diversity of the conventional image coloring network.
The technical scheme adopted by the invention is that the image coloring method based on the Transformer and the generation countermeasure network is implemented according to the following steps:
step 1, constructing an image coloring model based on a generation countermeasure network, wherein the image coloring model comprises a color image generator and a discriminator; the color image generator is used for generating a color image, and the discriminator is used for judging whether the input image is a real color image or a false color image;
step 2, inputting the gray image into a color image generator of the image coloring model to generate a pseudo color image;
and 3, respectively updating parameters of the discriminator and the color image generator:
step 3.1: firstly, fixing parameters of a color image generator, inputting the false color image and the real color image corresponding to the gray image into an identifier in sequence, then calculating the loss between the real color image corresponding to the gray image and a label value of 1 according to a loss function, calculating the loss between the false color image generated by the gray image and a label value of 0 according to the loss function, and finally updating the parameters of the identifier by using a back propagation algorithm; wherein a label value of 1 represents a real image and a label value of 0 represents a generated pseudo color image;
step 3.2: and fixing parameters of the discriminator, calculating the loss between the generated pseudo color image and the label value of 1 according to a loss function, and finally updating the parameters of the color image generator by utilizing a back propagation algorithm.
Step 3.3: continuously circulating the process of updating parameters of the discriminator and the color image generator in the steps 3.1 and 3.2 until the loss value is converged, and generating a false color image with good effect by the color image generator, namely obtaining an optimized image coloring model;
and 4, directly coloring the gray image by using the optimized image coloring model.
The present invention is also characterized in that,
in step 1, the color image generator includes a plurality of MWin-transformer modules, where the MWin-transformer modules are used to extract and reconstruct features of an image, and output a 3-channel effective color image:
the Mwin-transformer module consists of three core parts: a window-based multi-head self-attention mechanism, a layer normalization operation LN, and a local enhanced forward propagation network LeFF.
The flow of the color image generator to generate a pseudo color image is as follows:
X′=Embedded Tokens(X in )
X″=W-MSA(LN(X))+X′
X out =LeFF(LN(X″))+X″
wherein, X in Representing an input as a gray image or a pseudo-color image;
embedding Tokens represents dividing X in Converting into a vector;
x' represents X in Inputting the vector obtained by Embedding Tokens and outputting the vector;
then inputting a result LN (X ') obtained by carrying out layer normalization on the vector X' into a multi-head self-attention mechanism W-MSA based on a window to obtain a vector with extracted characteristic information, and adding the vector with X 'to obtain a vector X' with more collected characteristic information; x 'represents the input of X' into the output of the window-based multi-headed autofocusing mechanism and layer normalization operation;
continuously carrying out layer normalization on the vector X ', inputting the normalized LN (X ') into a local enhanced forward propagation network to obtain a vector with more extracted local characteristic information, and adding the vector with X ' to obtain a vector X with more gathered local characteristic information out ,X out Represents the input of X "into the locally enhanced forward propagation network LeFF and the resulting output of the layer normalization operation.
The layer normalization LN operation is to solve the problem of internal covariate offset, and the calculation process of the layer normalization operation is as follows:
Figure BDA0003887136350000031
wherein, the action object of LN layer is
Figure BDA0003887136350000032
X represents a vector, μ and δ represent the mean and variance of each sample,
Figure BDA0003887136350000033
and
Figure BDA0003887136350000034
as affine learning parameters, d k Is the dimension of the concealment that is in place,
Figure BDA0003887136350000035
indicating that the number is a k-dimensional vector.
The window-based multi-head self-attention mechanism is as follows:
the false color image is divided into a plurality of windows, then self-attention calculation is carried out in the different windows, and as the patch number in one window is far less than all small blocks in one image and the number of the windows is kept unchanged, the calculation complexity of the multi-head self-attention mechanism based on the windows and the image size are changed into a linear relation from a square relation, and the calculation complexity of the model is greatly reduced.
And adding convolution to a forward propagation network in the Mwin-transformer module so as to form a local enhanced forward propagation network LeFF.
The loss function is:
Figure BDA0003887136350000041
wherein,
Figure BDA0003887136350000042
Figure BDA0003887136350000043
wherein G is * The sum of the loss functions is represented,
Figure BDA0003887136350000044
the presentation condition is generated against the network loss,
Figure BDA0003887136350000045
denotes the Charbonier loss, λ denotes the weighting coefficient of the Charbonier loss;
x represents an input gray image;
y represents a real color image corresponding to the input gray image;
log represents a base 2 logarithmic function;
Figure BDA0003887136350000046
representing the independent variables as x, y;
Figure BDA0003887136350000047
represents an independent variable of x;
ε represents a value of 10 -3 Constant coefficient of (d);
and | | represents solving an absolute value.
The invention has the beneficial effects that: the invention relates to a method for image coloring based on a Transformer and a generation countermeasure network, which can be used for coloring gray images. The image coloring method has good coloring effect on the gray image in detail or the whole, is suitable for the gray image with any size, and has high universality.
Drawings
FIG. 1 is a block diagram of an image rendering model of the present invention;
fig. 2 is a structural diagram of the color image generator G;
FIG. 3 (a) is a block diagram of an discriminator D according to the invention;
FIG. 3 (b) is a structural diagram of the MWin-transporter of the present invention;
fig. 3 (c) is a structural diagram of the local enhanced forward propagation network LeFF of the present invention.
Detailed Description
The invention discloses an image coloring method based on a Transformer and a generation countermeasure network, which is implemented according to the following steps:
step 1, constructing an image coloring model based on a generation countermeasure network, wherein the image coloring model comprises a color image generator and a discriminator; the color image generator is used for generating a color image, and the discriminator is used for judging whether the input image is a real color image or a false color image;
step 2, inputting the gray image into a color image generator of the image coloring model to generate a pseudo color image;
and 3, respectively updating parameters of the discriminator and the color image generator:
step 3.1: firstly, fixing parameters of a color image generator, sequentially and alternately inputting the pseudo color image and a real color image corresponding to the gray image into a discriminator, then calculating the loss between the real color image corresponding to the gray image and a label value of 1 according to a loss function, calculating the loss between the pseudo color image generated by the gray image and a label value of 0 according to the loss function, and finally updating the parameters of the discriminator by utilizing a back propagation algorithm; wherein a label value of 1 represents a real image and a label value of 0 represents a generated pseudo color image;
step 3.2: and fixing parameters of the discriminator, calculating the loss between the generated pseudo color image and the label value of 1 according to a loss function, and finally updating the parameters of the color image generator by using a back propagation algorithm.
Step 3.3: continuously circulating the process of updating parameters of the discriminator and the color image generator in the steps 3.1 and 3.2 until the loss value is converged, and generating a false color image with good effect by the color image generator, namely obtaining an optimized image coloring model;
and 4, directly coloring the gray image by using the optimized image coloring model.
As shown in FIG. 1, an image coloring model based on Transformer and generation countermeasure network GAN is constructed, G and D represent a color image generator and a discriminator respectively, and specifically, a gray image x ∈ R 3×H×W As input to the color image generator G, a pseudo-color image G (x) is generated, and then the pseudo-color image G (x) and the true-color image y are alternately input to the discriminator.
Firstly, fixing parameters of a color image generator, sequentially and alternately inputting the pseudo color image and a real color image corresponding to the gray image into a discriminator, then calculating the loss between the real color image corresponding to the gray image and a label value of 1 according to a loss function, calculating the loss between the pseudo color image generated by the gray image and a label value of 0 according to the loss function, and finally updating the parameters of the discriminator by utilizing a back propagation algorithm; where a label value of 1 represents a real image and a label value of 0 represents a generated pseudo color image.
And fixing parameters of the discriminator, calculating the loss between the generated pseudo color image and the label value of 1 according to a loss function, and finally updating the parameters of the color image generator by utilizing a back propagation algorithm. And (3) continuously circulating the process of updating the parameters of the discriminator and the color image generator in the steps 3.1 and 3.2 until the loss value is converged, and generating a false color image with good effect by the color image generator, namely obtaining an optimized image coloring model. And directly coloring the gray image by using the optimized image coloring model.
The design of the method mainly comprises two main points, namely the design of the color image generator and the discriminator and the design of the components thereof, and the detailed construction of the color image generator and the discriminator is described one by one below.
(1) Design of color image generator G:
the input and output of image colorization is a mapping relationship, and the depth conversion process of these components should be a symmetric relationship, based on which, as shown in fig. 2, the whole color image generator is designed into a U shape; in general, at the encoder stage, the gray image x is first subjected to input dimension adjustment, which is a convolution layer of convolution kernel 3 × 3, and the activation function LeakyReLU is used to adjust the input dimensions and extract low-level features. Then, after the designed window-based transformer module MWin-transformer, a downsampling layer consisting of convolution layers with a convolution kernel of step 2 by 4 is reached, and this step is repeated 2 times. The image is then passed through the MWin-transformer module as the bottleneck stage. The decoder and the encoder correspond to each other, and the design of the decoder and the encoder is completely symmetrical and consistent: first, it passes through the window MWin-transformer module and performs upsampling, where the upsampling operation is a transposed convolution with 2 x 2 of the convolution kernel with step size of 2. This step is repeated 2 times as the encoder in order to maintain the symmetry of the network. Finally, the output dimensions are adjusted using an output dimension adjustment consisting of a 3 x 3 convolution to ensure that the output is a 3-channel valid color image.
(2) Design of MWin-transform module
In FIG. 3 (b), the MWin-transformer module was constructed, consisting of three core parts: a W-MHSA mechanism, a layer normalization operation LN and a local enhanced forward propagation network LeFF network;
wherein the layer normalization operation LN is as follows:
Figure BDA0003887136350000071
the LN layer is an important guarantee for fast training and stable convergence of the image coloring model, and the LN layer is used as an action object
Figure BDA0003887136350000081
Figure BDA0003887136350000082
And
Figure BDA0003887136350000083
the mean and variance of each sample are represented separately,
Figure BDA0003887136350000084
and
Figure BDA0003887136350000085
as affine learning parameters, d k Is the hidden dimension. The calculation flow of the MWin-transformer module is as follows:
X′=Embedded Tokens(X in )
X″=W-MSA(LN(X))+X′
X out =LeFF(LN(X″))+X
wherein, X in Indicating inputIn, it is gray image or false color image;
embedding token Embedding Tokens denotes to assign X to in Converting into a vector;
x' represents X in Inputting vector output obtained by Embedding a token Embedding token;
then inputting a result LN (X ') obtained by carrying out layer normalization on the vector X' into a multi-head self-attention mechanism W-MSA based on a window to obtain a vector with extracted characteristic information, and adding the vector with X 'to obtain a vector X' with more collected characteristic information; x 'represents the input of X' into the window-based multi-headed autofocusing mechanism and the output resulting from the layer normalization operation;
continuously carrying out layer normalization on the vector X ', inputting the normalized LN (X ') into a local enhanced forward propagation network to obtain a vector with more extracted local characteristic information, and then adding the vector with X ' to obtain a vector X with more gathered local characteristic information out ,X out Represents the input of X "into the local enhanced forward propagation network and the resulting output of the layer normalization operation.
(3) Local enhanced forward propagation network (LeFF)
To enhance the ability of the image rendering model to capture local features, we add convolution to the forward propagation network, forming a local enhanced forward propagation network (LeFF), the specific design is shown in fig. 3 (c): firstly, an input sequence is changed into an image through a sequence changing module, the sequence is changed into the image, the image is convoluted by a convolution kernel of 1 × 1, then the image is activated through an activation function, then the image is convoluted by a convolution kernel of 3 × 3 and the convolution kernel of 1 × 1, then the image is activated through the activation function, and finally the image is changed into the sequence again to complete the local enhancement forward propagation network.
(4) Design of discriminator D
The essence of the discriminator is to determine whether the given data is "real", i.e. to determine whether the given data is real training data or false data generated by the color image generator G, as shown in fig. 3 (a), we first input real color images or false color images, and unwind them into small blocks (patch) by linear flattening, linearly flatten the blocks composed of convolution layers, then stack 4 MWin Transformer blocks having the same structure as G, and finally output true or false through the linear layers, thereby realizing the function of discrimination.
Example 1
In order to prove the effectiveness of the image coloring model, experiments are respectively carried out on an animal face image and a landscape image, and the image coloring model is compared with other current popular models: yoo et al (Yoo, S., bahng, H., chung, S., lee, J., chang, J., choo, J.: colouring with limited data: raw-shot colour assessed networks. In: proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.11283-11292 (2019)), su et al (Su, J. -W., chu, H., K., huang, J.: instance-aware image orientation. In: proceedings of the IEEE/F Conference and Pattern Recognition, pp.7968-7968)), and we compare the three indices: the Frechet increment distance score, the peak signal-to-noise ratio and the structural similarity are improved on the test of the two data sets, and the three indexes are improved by 0.003, 0.263 and 0.014 at the lowest. We find the more complex scenes the more realistic the details of model coloring, and we assume this is because a transform learns more about the distribution of large amounts of data than a CNN. In addition, we observed that the overall coloring effect of our model was smoother and more uniform than other methods, and without excessive color discontinuities, which demonstrates that the transformer can better capture the global information of the image.

Claims (7)

1. Image rendering method based on Transformer and generative confrontation network, characterized in that it is implemented according to the following steps,
step 1, constructing an image coloring model based on a generation countermeasure network, wherein the image coloring model comprises a color image generator and a discriminator; the color image generator is used for generating a color image, and the discriminator is used for judging whether the input image is a real color image or a false color image;
step 2, inputting the gray image into a color image generator of the image coloring model to generate a pseudo color image;
and 3, respectively updating parameters of the discriminator and the color image generator:
step 3.1: firstly, fixing parameters of a color image generator, inputting the false color image and the real color image corresponding to the gray image into an identifier in sequence, then calculating the loss between the real color image corresponding to the gray image and a label value of 1 according to a loss function, calculating the loss between the false color image generated by the gray image and a label value of 0 according to the loss function, and finally updating the parameters of the identifier by using a back propagation algorithm; wherein a label value of 1 represents a real image and a label value of 0 represents a generated pseudo-color image;
step 3.2: fixing parameters of the discriminator, calculating the loss between the generated pseudo color image and the label value of 1 according to a loss function, and finally updating the parameters of the color image generator by utilizing a back propagation algorithm;
step 3.3: continuously circulating the process of updating parameters of the discriminator and the color image generator in the steps 3.1 and 3.2 until the loss value is converged, and generating a false color image with good effect by the color image generator, namely obtaining an optimized image coloring model;
and 4, directly coloring the gray image by using the optimized image coloring model.
2. The method for rendering images based on Transformer and generation countermeasure network as claimed in claim 1, wherein in step 1, the color image generator comprises multiple MWin-Transformer modules, and the function of the MWin-Transformer module is to extract and reconstruct the features of the images, and output a 3-channel effective color image: the Mwin-transformer module consists of three core parts: a window-based multi-head self-attention mechanism, a layer normalization operation LN, and a local enhanced forward propagation network LeFF.
3. The Transformer-based and confrontational network-generating image rendering method of claim 2, wherein the color image generator generates a pseudo-color image as follows:
X′=Embedded Tokens(X in )
X″=W-MSA(LN(X))+X′
X out =LeFF(LN(X″))+X″
wherein X in Representing an input as a gray image or a false color image;
embedding Tokens denotes to assign X to in Converting into a vector;
x' represents X in Inputting the vector obtained by Embedding Tokens and outputting the vector;
then inputting a result LN (X ') obtained by carrying out layer normalization on the vector X' into a multi-head self-attention mechanism W-MSA based on a window to obtain a vector with extracted characteristic information, and adding the vector with X 'to obtain a vector X' with more collected characteristic information; x 'represents the input of X' into the output of the window-based multi-headed autofocusing mechanism and layer normalization operation;
continuously carrying out layer normalization on the vector X ', inputting the normalized LN (X ') into a local enhanced forward propagation network to obtain a vector with more extracted local characteristic information, and adding the vector with X ' to obtain a vector X with more gathered local characteristic information out ,X out Represents the input of X "into the locally enhanced forward propagation network LeFF and the resulting output of the layer normalization operation.
4. The transform-and-generative-confrontation-network-based image rendering method according to claim 3, wherein the layer normalization operation is calculated by:
Figure FDA0003887136340000031
wherein, the action object of LN layer is
Figure FDA0003887136340000032
X represents a vector, μ and δ represent the mean sum of each sampleThe variance of the measured values is calculated,
Figure FDA0003887136340000033
and
Figure FDA0003887136340000034
as affine learning parameters, d k Is the dimension of the concealment that is to be hidden,
Figure FDA0003887136340000035
indicating that the number is a k-dimensional vector.
5. The transform-based and generative countermeasure network-based image rendering method of claim 3, wherein the window-based multi-headed self-attention mechanism is as follows:
the method comprises the steps of dividing a pseudo-color image into a plurality of windows, and then performing self-attention calculation in the different windows, wherein the patch number in one window is far less than all small blocks in one picture, and the number of the windows is kept unchanged, so that the calculation complexity of the multi-head self-attention mechanism based on the windows and the size of the image are changed into a linear relation from a square relation, and the calculation complexity of a model is greatly reduced.
6. The transform-and-generate countermeasure-network-based image shading method of claim 2, wherein a convolution is added to the forward propagation network in the Mwin-transform module, thereby forming a local enhanced forward propagation network LeFF.
7. The Transformer-based and generative countermeasure network-based image rendering method of claim 1, wherein the loss function is:
Figure FDA0003887136340000036
wherein,
Figure FDA0003887136340000037
Figure FDA0003887136340000038
wherein, G * The sum of the loss functions is represented,
Figure FDA0003887136340000039
the presentation conditions are generated to counter network loss,
Figure FDA00038871363400000310
a weighting coefficient representing a Chardonnier loss, and λ represents the Chardonnier loss;
x represents an input gray image;
y represents a real color image corresponding to the input gray image;
log represents a base 2 logarithmic function;
Figure FDA0003887136340000041
representing the independent variables as x, y;
Figure FDA0003887136340000042
represents an independent variable of x;
ε represents a value of 10 -3 Constant coefficient of (d);
the absolute value is obtained.
CN202211247125.5A 2022-10-12 2022-10-12 Image coloring method based on Transformer and generation countermeasure network Pending CN115546338A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211247125.5A CN115546338A (en) 2022-10-12 2022-10-12 Image coloring method based on Transformer and generation countermeasure network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211247125.5A CN115546338A (en) 2022-10-12 2022-10-12 Image coloring method based on Transformer and generation countermeasure network

Publications (1)

Publication Number Publication Date
CN115546338A true CN115546338A (en) 2022-12-30

Family

ID=84733928

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211247125.5A Pending CN115546338A (en) 2022-10-12 2022-10-12 Image coloring method based on Transformer and generation countermeasure network

Country Status (1)

Country Link
CN (1) CN115546338A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115908617A (en) * 2023-01-09 2023-04-04 长春理工大学 Infrared image colorizing method and system
CN116137043A (en) * 2023-02-21 2023-05-19 长春理工大学 Infrared image colorization method based on convolution and transfomer
CN116433788A (en) * 2023-02-24 2023-07-14 北京科技大学 Gray image coloring method and device based on self-attention and generation countermeasure network

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115908617A (en) * 2023-01-09 2023-04-04 长春理工大学 Infrared image colorizing method and system
CN115908617B (en) * 2023-01-09 2024-06-07 长春理工大学 Infrared image colorization method and system
CN116137043A (en) * 2023-02-21 2023-05-19 长春理工大学 Infrared image colorization method based on convolution and transfomer
CN116433788A (en) * 2023-02-24 2023-07-14 北京科技大学 Gray image coloring method and device based on self-attention and generation countermeasure network

Similar Documents

Publication Publication Date Title
CN111145116B (en) Sea surface rainy day image sample augmentation method based on generation of countermeasure network
CN111798369B (en) Face aging image synthesis method for generating confrontation network based on circulation condition
CN115546338A (en) Image coloring method based on Transformer and generation countermeasure network
Wang et al. A reinforcement learning paradigm of configuring visual enhancement for object detection in underwater scenes
CN108961350B (en) Wind painting migration method based on saliency matching
CN109815893A (en) The normalized method in colorized face images illumination domain of confrontation network is generated based on circulation
CN112507793A (en) Ultra-short-term photovoltaic power prediction method
CN112489164B (en) Image coloring method based on improved depth separable convolutional neural network
CN113283444B (en) Heterogeneous image migration method based on generation countermeasure network
CN112002009B (en) Unsupervised three-dimensional face reconstruction method based on generation of confrontation network
Xiao et al. Progressive data augmentation method for remote sensing ship image classification based on imaging simulation system and neural style transfer
CN113989100B (en) Infrared texture sample expansion method based on style generation countermeasure network
CN111914617B (en) Face attribute editing method based on balanced stack type generation type countermeasure network
CN111401132A (en) Pedestrian attribute identification method guided by high-level semantics under monitoring scene
CN112270300A (en) Method for converting human face sketch image into RGB image based on generating type confrontation network
CN114399533B (en) Single-target tracking method based on multi-level attention mechanism
CN114581552A (en) Gray level image colorizing method based on generation countermeasure network
Zhang et al. MSGAN: generative adversarial networks for image seasonal style transfer
CN118230176B (en) Wetland classification method and system based on multisource remote sensing and space-time depth integrated learning
CN118172499B (en) Building height inversion method based on resource third-order remote sensing image
CN113888399B (en) Face age synthesis method based on style fusion and domain selection structure
CN116935043A (en) Typical object remote sensing image generation method based on multitasking countermeasure network
KumarSingh et al. An Enhanced Image Colorization using Modified Generative Adversarial Networks with Pix2Pix Method
Lee et al. Design of CycleGAN model for SAR image colorization
CN116137043A (en) Infrared image colorization method based on convolution and transfomer

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination