CN112802160B - U-GAT-IT-based improved method for migrating cartoon style of Qin cavity character - Google Patents

U-GAT-IT-based improved method for migrating cartoon style of Qin cavity character Download PDF

Info

Publication number
CN112802160B
CN112802160B CN202110037379.3A CN202110037379A CN112802160B CN 112802160 B CN112802160 B CN 112802160B CN 202110037379 A CN202110037379 A CN 202110037379A CN 112802160 B CN112802160 B CN 112802160B
Authority
CN
China
Prior art keywords
image
gat
cartoon
qin
qin cavity
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110037379.3A
Other languages
Chinese (zh)
Other versions
CN112802160A (en
Inventor
耿国华
马星锐
刘喆
周明全
冯龙
李启航
刘晓宁
刘阳洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NORTHWEST UNIVERSITY
Original Assignee
NORTHWEST UNIVERSITY
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NORTHWEST UNIVERSITY filed Critical NORTHWEST UNIVERSITY
Priority to CN202110037379.3A priority Critical patent/CN112802160B/en
Publication of CN112802160A publication Critical patent/CN112802160A/en
Application granted granted Critical
Publication of CN112802160B publication Critical patent/CN112802160B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • G06T13/403D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a U-GAT-IT-based improved method for transferring cartoon style of a Qin cavity role, which comprises the steps of firstly, manufacturing a unique cartoon data set of the Qin cavity role, carrying out preprocessing operations such as modifying the size of a collected face picture, and secondly, adding two stacked up-down sampling convolution modules before a coder and after a decoder on the basis of a U-GAT-IT network, so as to gradually improve the characteristic abstraction and reconstruction capability of a model. In addition, a face loss function is added into the U-GAT-IT network, and the newly generated cartoon diagram of the Qin cavity role is restrained, so that the identity information of the cartoon diagram of the Qin cavity role is kept consistent with the input face information as much as possible. And finally, constructing a network model through a pytorch, and transmitting the processed face image into the network to obtain the cartoon graph with the characteristic of the Qin cavity role. The method solves the problem that the identity information of the original image is easy to lose when the cartoon image is migrated in style, improves the defect that a large number of data sets are needed by U-GAT-IT, and can more accurately outline the detailed information of the cartoon image.

Description

U-GAT-IT-based improved method for migrating cartoon style of Qin cavity character
Technical Field
The invention belongs to the technical field of computer graphics processing, and particularly relates to a U-GAT-IT-based improved method for migrating a cartoon style of a Qin cavity character.
Background
The Qin cavity is a quite ancient theatrical species, and is called as a nasal ancestor of China's drama. The role of the Qin cavity basically inherits the scale system of the Yuan-omnipotent drama and the Ming dynasty legend 'river and lake twelve roles'. In the general view, four major lines are raw, denier, net and ugly. With the rapid development of modern economy, entertainment modes of the masses of the people are changed. The original entertainment function of the Qin cavity is weakened, and the audience group is aged and faces the crisis of survival and development. The image of the face of the Qin cavity is transferred into the image of the cartoon style through the image cartoon style on the basis of U-GAT-IT, and the computer is used for refreshing the cultural heritage of the non-matter of the Qin cavity art, thereby realizing cultural innovation of the Qin cavity art and enabling people to regain interest in traditional drama in a more modern way.
Where U-GAT-IT refers to an unsupervised generation attention network with adaptive layer instance normalization for image-to-image translation.
The migration of the cartoon style of the human face is to keep the identity information of the original image unchanged, keep as many texture details as possible and convert the real human face image into a non-realistic image of the cartoon style. The cartoon image is a simplified line, the line has clear texture and smooth color blocks, but the technology is easy to have the following technical problems of losing identity information and image details of a human face, having great acquisition difficulty of cartoon image data, having high manufacturing cost of the cartoon image and the like in the process of cartoon image processing.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention aims to provide the improved method for the cartoon style migration of the Qin cavity character based on the U-GAT-IT, which solves the problem that the identity information of the original image is easy to lose during the style migration of the cartoon image, improves the defect that the U-GAT-IT needs a large number of data sets, and can more accurately outline the detailed information of the cartoon image.
In order to achieve the above purpose, the invention adopts the following technical scheme:
a U-GAT-IT-based improved method for migrating the cartoon style of a Qin cavity character comprises the following steps of;
step 1: acquiring a cartoon data set of a Qin cavity role and a face picture set;
step 2: constructing an original U-GAT-IT network frame by using a pyrach, wherein the U-GAT-IT network frame comprises a Generator and a Discriminator, and one face image acquired in the step 1 is input into the U-GAT-IT network frame as source domain images X-X S Simultaneously inputting the cartoon image of the Qin cavity character obtained in the step 1 as the target domain images X-X y
Step 3: the source domain image and the target domain image obtained in the step 2 are processed through two stacked up-down sampling convolution modules, the characteristics of the source domain image and the target domain image are extracted under the condition that the semantic information position is unchanged, and then the characteristics of the source domain image and the target domain image are processed through an encoder E in a Generator S Coding to obtain a coded characteristic diagram
Step 4: the feature map obtained in the step 3 is passed through an auxiliary classifier eta of two classifications in a Generator s Classifying to distinguish whether the image comes from the source domain or the target domain to obtain the weight information of each feature map
Step 5: multiplying the weight information of each feature map obtained in the step 4 with the encoded feature map to obtain a feature map with attention;
step 6: obtaining a class activation diagram from the attention characteristic diagram obtained in the step 5 through a convolution and activation function layer of 1*1
Wherein a represents each class activation graph, w represents weight information of each feature graph, and E represents the encoded feature graph;
step 7: the class activation diagram obtained in the step 6 is processedThe decoder G in the Generator is obtained by the full connection layer fc t 2c parameters gamma and beta used for carrying out normalization treatment on AdaLIN in the Adaptive Layer-Instance Normalization Layer;
step 8: the gamma and beta obtained in the step 7 are further generated into an image through an adaptive residual block Adaptive Residual Blocks;
step 9: the image obtained in the step 8 is further generated into a target domain image from the embedded feature through two newly added stacked up-down sampling convolution modules;
step 10: judging the authenticity of the image by the CAM activation diagram in the Discriminator;
step 11: and (3) calculating the result judged in the step (10) through GAN Loss, cycle Loss, identity Loss, CAM Loss and Face Loss functions in the U-GAT-IT network framework to obtain the cartoon image of the character characteristics of the Qin cavity.
Further, the normalization formulas of Instance Normalization and Layer Normalization in AdaLIN in step 7 are:
wherein mu I And delta I Mean and standard deviation, mu, respectively, of Instance Normalization example normalization mode L And delta L The mean value and standard deviation are respectively calculated in a Layer Normalization layer normalization mode;
the normalized weighted sum formula of Instance Normalization and Layer Normalization in AdaLIN is:
wherein λ is a feature map with attention, γ is a scaling factor, and β is an offset; θ is a learning weight, updated by back propagation.
Further, the θ interval clipping formula is:
θ←clip[0,1](θ-τΔθ)
where τ is the learning rate.
Further, the two stacking up-down sampling convolution modules in the step 9 include four pairs of downsampling blocks and upsampling blocks which are corresponding to each other from end to end, so as to extract the surface features of the face in the picture before entering the main network, abstract the input image into a form easy to encode, and ensure the unchanged position of semantic information while extracting the features.
Further, the function formula of the Loss function Face Loss in step 11 is:
wherein P is source P is a source domain image target Is a target domain image; l (L) face The cosine distance is used to constrain the features of the source domain image and the features of the target domain image.
Further, the cartoon data set of the Qin cavity role in the step 1 is a flower denier cartoon data set, and the face picture set is a female face picture set.
Compared with the prior art, the invention has the beneficial effects that:
the invention discloses a method for transferring cartoon style of a Qin cavity character based on U-GAT-IT improvement, which comprises the steps of firstly obtaining a cartoon data set and a face picture set of the Qin cavity character, and secondly adding two stacked up-down sampling convolution modules before an encoder and after a decoder on the basis of a U-GAT-IT network so as to gradually improve the characteristic abstract and reconstruction capability of a model. In addition, a face loss function is added into the U-GAT-IT network, and the newly generated cartoon diagram of the Qin cavity role is restrained, so that the identity information of the cartoon diagram of the Qin cavity role is kept consistent with the input face information as much as possible. And finally, constructing a network model through a pytorch, and transmitting the processed face image into the network to obtain the cartoon graph with the characteristic of the Qin cavity role. The method solves the problem that the identity information of the original image is easy to lose when the cartoon image is migrated in style, improves the defect that a large number of data sets are needed by U-GAT-IT, and can more accurately outline the detailed information of the cartoon image.
Drawings
FIG. 1 is a cartoon diagram of a Qin cavity flower denier made by the invention;
FIG. 2 is a network block diagram of the present invention;
FIG. 3 is a block diagram of a stacked up-down sample convolution module of the present invention;
FIG. 4 is a network architecture diagram within the Generator of the present invention;
FIG. 5 is a network structure diagram of the inside of the Discriminator of the present invention;
FIG. 6 is a block diagram of an AdaLIN example normalization of the present invention;
fig. 7 is a graph of the results of a cavity role cartoon style migration through a modified U-GAT-IT network.
Detailed Description
The present invention will be described in further detail with reference to specific examples, but is not limited thereto.
The invention provides a U-GAT-IT-based improved method for transferring the cartoon style of a Qin cavity character, which comprises the following steps of;
step 1: the Qin cavity flower denier cartoon data set is manufactured by collecting flower denier pictures on the Qin cavity website, and the picture size is 512 x 512mm. Collecting female face pictures, wherein the size is set as follows: 256 x 256mm (see fig. 1).
Step 2: constructing an original U-GAT-IT network framework (see FIG. 2) using a pytorch, the U-GAT-IT network framework including a generator GeneThe Discriminator (see fig. 4) and the Discriminator (see fig. 5), one face image acquired in step 1 is input as source domain images X to X in the U-GAT-IT network frame S Simultaneously inputting the cartoon image of the Qin cavity character obtained in the step 1 as the target domain images X-X y
Step 3: the source domain image and the target domain image obtained in the step 2 are processed through two stacked up-down sampling convolution modules (refer to figure 3), the characteristics of the source domain image and the target domain image are extracted under the condition that the semantic information positions are unchanged, and then the characteristics of the source domain image and the target domain image are processed through an encoder E in a Generator S Coding to obtain a coded characteristic diagram
Step 4: the feature map in the step 3 is passed through an auxiliary classifier eta of two classifications in a Generator s Classifying to distinguish whether the image comes from the source domain or the target domain to obtain the weight information of each feature map
Step 5: and (3) multiplying the weight information of each feature map obtained in the step (4) with the encoded feature map to obtain the feature map with attention.
Step 6: obtaining a class activation diagram from the attention characteristic diagram obtained in the step 5 through a convolution and activation function layer of 1*1
Where a represents each class activation map, w represents weight information of each feature map, and E represents the encoded feature map.
Step 7: in step 6The resulting class activation diagramThe decoder G in the Generator is obtained by the full connection layer fc t 2c parameters gamma and beta used for normalization processing of AdaLIN (refer to figure 6) in the Adaptive Layer-Instance Normalization Layer;
the normalized formulas of Instance Normalization and Layer Normalization in AdaLIN are:
wherein mu I And delta I Mean and standard deviation, mu, calculated for the Instance Normalization example normalization L And delta L Mean and standard deviation were determined for the Layer Normalization layer normalization approach.
The normalized weighted sum formula for Instance Normalization and Layer Normalization in AdaLIN is:
where λ is the feature map with attention, λ is the scaling factor, and β is the offset. θ is a learning weight, updated by back propagation.
In order to prevent theta from exceeding the range of [0,1], the theta is subjected to interval clipping with the clipping formula of
θ←clip[0,1](θ-τΔθ)
Where τ is the learning rate.
Step 8: the gamma and beta obtained in the step 7 are further generated into an image through an adaptive residual block Adaptive Residual Blocks;
step 9: the image obtained in the step 8 is further generated into a target domain image from the embedded feature through two newly added stacked up-down sampling convolution modules;
step 10: judging the authenticity of the image by the CAM activation diagram in the Discriminator;
step 11: and (3) calculating the result judged in the step (10) through GAN Loss, cycle Loss, identity Loss, CAM Loss and Face Loss functions in the U-GAT-IT network framework to obtain a cartoon image with the Qin cavity flower denier characteristic (refer to figure 7).
The function formula of the Face Loss function is as follows:
wherein P is source P is a source domain image target Is a target domain image. L (L) face The cosine distance is used to constrain the features of the source domain image and the features of the target domain image. The Face Loss function is added to restrict the newly generated flower denier cartoon graph of the Qin cavity, so that the identity information of the flower denier cartoon graph is kept consistent with the input Face information as much as possible.
In summary, the embodiment realizes style migration from the face to the Qin cavity flower denier cartoon on the basis of ensuring the consistency of the face information and the identity information of the Qin cavity flower denier cartoon, solves the problems of high difficulty in acquiring traditional cartoon image data and high manufacturing cost of the cartoon image, simultaneously solves the problem that the identity information of the original image is easy to lose during style migration, improves the defect that a large number of data sets are needed by U-GAT-IT, and can more accurately outline the detail information of the cartoon image.

Claims (6)

1. A U-GAT-IT-based improved method for migrating the cartoon style of a Qin cavity character is characterized in that: comprises the following steps of;
step 1: acquiring a cartoon data set of a Qin cavity role and a face picture set;
step 2: constructing an original U-GAT-IT network frame by using a pyrach, wherein the U-GAT-IT network frame comprises a Generator and a Discriminator, and inputting the person acquired in the step 1 into the U-GAT-IT network frameFace image is used as source domain image X-X S Simultaneously inputting the cartoon image of the Qin cavity character obtained in the step 1 as the target domain images X-X y
Step 3: the source domain image and the target domain image obtained in the step 2 are processed through two stacked up-down sampling convolution modules, the characteristics of the source domain image and the target domain image are extracted under the condition that the semantic information position is unchanged, and then the characteristics of the source domain image and the target domain image are processed through an encoder E in a Generator S Coding to obtain a coded characteristic diagram
Step 4: the feature map obtained in the step 3 is passed through an auxiliary classifier eta of two classifications in a Generator s Classifying to distinguish whether the image comes from the source domain or the target domain to obtain the weight information of each feature map
Step 5: multiplying the weight information of each feature map obtained in the step 4 with the encoded feature map to obtain a feature map with attention;
step 6: obtaining a class activation diagram from the attention characteristic diagram obtained in the step 5 through a convolution and activation function layer of 1*1
Wherein a represents each class activation graph, w represents weight information of each feature graph, and E represents the encoded feature graph;
step 7: the class activation diagram obtained in the step 6 is processedObtaining a green body through the full connection layer fcDecoder G in the Generator t 2c parameters gamma and beta used for carrying out normalization treatment on AdaLIN in the Adaptive Layer-Instance Normalization Layer;
step 8: the gamma and beta obtained in the step 7 are further generated into an image through an adaptive residual block Adaptive Residual Blocks;
step 9: the image obtained in the step 8 is further generated into a target domain image from the embedded feature through two newly added stacked up-down sampling convolution modules;
step 10: judging the authenticity of the image by the CAM activation diagram in the Discriminator;
step 11: and (3) calculating the result judged in the step (10) through GAN Loss, cycle Loss, identity Loss, CAM Loss and Face Loss functions in the U-GAT-IT network framework to obtain the cartoon image of the character characteristics of the Qin cavity.
2. The method for improving the cartoon style migration of the Qin cavity roles based on U-GAT-IT according to claim 1, wherein the method comprises the following steps of: the normalized formulas of Instance Normalization and Layer Normalization in AdaLIN in step 7 are:
wherein mu I And delta I Mean and standard deviation, mu, respectively, of Instance Normalization example normalization mode L And delta L The mean value and standard deviation are respectively calculated in a Layer Normalization layer normalization mode;
the normalized weighted sum formula of Instance Normalization and Layer Normalization in AdaLIN is:
wherein λ is a feature map with attention, γ is a scaling factor, and β is an offset; θ is a learning weight, updated by back propagation.
3. The method for improving the cartoon style migration of the Qin cavity roles based on U-GAT-IT according to claim 2, wherein the method comprises the following steps of: the theta interval clipping formula is as follows:
θ←clip[0,1](θ-τΔθ)
where τ is the learning rate.
4. A method for improving the cartoonization style migration of a Qin cavity character based on U-GAT-IT according to any one of claims 1 to 3, wherein: the two stacking up-down sampling convolution modules in the step 9 comprise four pairs of downsampling blocks and upsampling blocks which are mutually corresponding from beginning to end, and the purpose is to extract the surface features of the face in the picture before entering the main network, abstract the input image into a form easy to encode, and ensure the invariable semantic information position while extracting the features.
5. A method for improving the cartoonization style migration of a Qin cavity character based on U-GAT-IT according to any one of claims 1 to 3, wherein: the function formula of the Loss function Face Loss in step 11 is:
wherein P is source P is a source domain image target Is a target domain image; l (L) face The cosine distance is used to constrain the features of the source domain image and the features of the target domain image.
6. A method for improving the cartoonization style migration of a Qin cavity character based on U-GAT-IT according to any one of claims 1 to 3, wherein: the cartoon data set of the Qin cavity role in the step 1 is a flower denier cartoon data set, and the face picture set is a female face picture set.
CN202110037379.3A 2021-01-12 2021-01-12 U-GAT-IT-based improved method for migrating cartoon style of Qin cavity character Active CN112802160B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110037379.3A CN112802160B (en) 2021-01-12 2021-01-12 U-GAT-IT-based improved method for migrating cartoon style of Qin cavity character

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110037379.3A CN112802160B (en) 2021-01-12 2021-01-12 U-GAT-IT-based improved method for migrating cartoon style of Qin cavity character

Publications (2)

Publication Number Publication Date
CN112802160A CN112802160A (en) 2021-05-14
CN112802160B true CN112802160B (en) 2023-10-17

Family

ID=75810121

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110037379.3A Active CN112802160B (en) 2021-01-12 2021-01-12 U-GAT-IT-based improved method for migrating cartoon style of Qin cavity character

Country Status (1)

Country Link
CN (1) CN112802160B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019042139A1 (en) * 2017-08-29 2019-03-07 京东方科技集团股份有限公司 Image processing method, image processing apparatus, and a neural network training method
CN109886881A (en) * 2019-01-10 2019-06-14 中国科学院自动化研究所 Face dressing minimizing technology
CN110310221A (en) * 2019-06-14 2019-10-08 大连理工大学 A kind of multiple domain image Style Transfer method based on generation confrontation network
CN111243066A (en) * 2020-01-09 2020-06-05 浙江大学 Facial expression migration method based on self-supervision learning and confrontation generation mechanism
WO2020168731A1 (en) * 2019-02-19 2020-08-27 华南理工大学 Generative adversarial mechanism and attention mechanism-based standard face generation method
CN111667401A (en) * 2020-06-08 2020-09-15 武汉理工大学 Multi-level gradient image style migration method and system
CN111754446A (en) * 2020-06-22 2020-10-09 怀光智能科技(武汉)有限公司 Image fusion method, system and storage medium based on generation countermeasure network
CN112132208A (en) * 2020-09-18 2020-12-25 北京奇艺世纪科技有限公司 Image conversion model generation method and device, electronic equipment and storage medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019042139A1 (en) * 2017-08-29 2019-03-07 京东方科技集团股份有限公司 Image processing method, image processing apparatus, and a neural network training method
CN109886881A (en) * 2019-01-10 2019-06-14 中国科学院自动化研究所 Face dressing minimizing technology
WO2020168731A1 (en) * 2019-02-19 2020-08-27 华南理工大学 Generative adversarial mechanism and attention mechanism-based standard face generation method
CN110310221A (en) * 2019-06-14 2019-10-08 大连理工大学 A kind of multiple domain image Style Transfer method based on generation confrontation network
CN111243066A (en) * 2020-01-09 2020-06-05 浙江大学 Facial expression migration method based on self-supervision learning and confrontation generation mechanism
CN111667401A (en) * 2020-06-08 2020-09-15 武汉理工大学 Multi-level gradient image style migration method and system
CN111754446A (en) * 2020-06-22 2020-10-09 怀光智能科技(武汉)有限公司 Image fusion method, system and storage medium based on generation countermeasure network
CN112132208A (en) * 2020-09-18 2020-12-25 北京奇艺世纪科技有限公司 Image conversion model generation method and device, electronic equipment and storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
U-gat-it: Unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation;Junho Kim等;《arXiv》;20200408;第1-19页 *
关键人脸轮廓区域卡通风格化生成算法;范林龙等;《图学学报》;20201225;第42卷(第1期);第1-8页 *

Also Published As

Publication number Publication date
CN112802160A (en) 2021-05-14

Similar Documents

Publication Publication Date Title
CN110599409B (en) Convolutional neural network image denoising method based on multi-scale convolutional groups and parallel
CN110348330B (en) Face pose virtual view generation method based on VAE-ACGAN
CN108921786B (en) Image super-resolution reconstruction method based on residual convolutional neural network
CN110020989B (en) Depth image super-resolution reconstruction method based on deep learning
CN109087258B (en) Deep learning-based image rain removing method and device
CN102156875B (en) Image super-resolution reconstruction method based on multitask KSVD (K singular value decomposition) dictionary learning
CN112967178B (en) Image conversion method, device, equipment and storage medium
CN110060286B (en) Monocular depth estimation method
CN111986108A (en) Complex sea-air scene image defogging method based on generation countermeasure network
CN107273908A (en) A kind of target identification method based on compressed sensing
CN105335929A (en) Depth map super-resolution method
Zhao et al. CREAM: CNN-REgularized ADMM framework for compressive-sensed image reconstruction
CN114821100B (en) Image compressed sensing reconstruction method based on structural group sparse network
CN117274059A (en) Low-resolution image reconstruction method and system based on image coding-decoding
CN116797461A (en) Binocular image super-resolution reconstruction method based on multistage attention-strengthening mechanism
CN110647891A (en) CNN (convolutional neural network) -based automatic extraction method and system for time sequence data characteristics of self-encoder
CN112802160B (en) U-GAT-IT-based improved method for migrating cartoon style of Qin cavity character
CN108090460B (en) Weber multidirectional descriptor-based facial expression recognition feature extraction method
CN111950454B (en) Finger vein recognition method based on bidirectional feature extraction
CN117237188A (en) Multi-scale attention network saliency target detection method based on remote sensing image
CN116630964A (en) Food image segmentation method based on discrete wavelet attention network
CN116703719A (en) Face super-resolution reconstruction device and method based on face 3D priori information
CN113706407B (en) Infrared and visible light image fusion method based on separation characterization
CN115936990A (en) Synchronous processing method and system for multi-scale super-resolution and denoising of seismic data
Zou et al. EDCNN: a novel network for image denoising

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant