CN112069769B - Intelligent word effect migration method and system for special effect words - Google Patents

Intelligent word effect migration method and system for special effect words Download PDF

Info

Publication number
CN112069769B
CN112069769B CN201910440039.8A CN201910440039A CN112069769B CN 112069769 B CN112069769 B CN 112069769B CN 201910440039 A CN201910440039 A CN 201910440039A CN 112069769 B CN112069769 B CN 112069769B
Authority
CN
China
Prior art keywords
special effect
network
picture
mask
migration
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910440039.8A
Other languages
Chinese (zh)
Other versions
CN112069769A (en
Inventor
刘家瑛
胡煜章
汪文靖
杨帅
郭宗明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Peking University
Original Assignee
Peking University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Peking University filed Critical Peking University
Priority to CN201910440039.8A priority Critical patent/CN112069769B/en
Publication of CN112069769A publication Critical patent/CN112069769A/en
Application granted granted Critical
Publication of CN112069769B publication Critical patent/CN112069769B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Processing (AREA)

Abstract

Hair brushThe invention provides an intelligent character effect migration method and system aiming at special effect characters, wherein the method comprises the following steps: training a mask extraction subnetwork to extract a decoration element mask by using a training data set, and training a basic special effect migration subnetwork to perform basic character special effect migration; special effect word D with decorative elements y And its paired font pictures C y Inputting the data into a trained mask extraction sub-network to obtain a decorative element mask M y (ii) a Will D y And its paired character pattern picture C y And a target font picture C x Inputting the result into a trained basic special effect migration sub-network to obtain a result S of basic special effect migration and decoration element elimination x (ii) a By means of M y ,C y And C x Carrying out element recombination to fuse decorative elements into S x Obtaining the moved special effect character D with the decorative elements corresponding to the target character pattern x . The method can transfer the decorative elements while transferring the special effect of the characters, and can not cause the loss and distortion of the decorative elements.

Description

Intelligent word effect migration method and system for special effect words
Technical Field
The invention belongs to the field of character special effect style migration, and particularly relates to an intelligent character effect migration method and system for special effect characters.
Background
The special effect word style migration aims to migrate the style of a given special effect word to a given font, and batch generation of the special effect word is achieved. In recent years, special effect word style migration has become an increasingly academic concern.
The special effect word style migration method can be divided into three categories. Conventional methods based on block fusion implement style migration by finding similar texture blocks on a given special effect word that block match the current word pattern. To better preserve the legibility of glyphs, this type of approach typically computes priors for the glyph structure and considers the similarity on the glyphs when finding similar blocks. The method based on the global statistic generally extracts the features by means of a deep neural network pre-trained on a classification task, then calculates the global statistic of the features, and enables a migration result to be consistent with a special effect of a character to be migrated on the global statistic in the modes of iteration, training a forward network and the like. The method based on the deep neural network firstly collects a pair training set, then builds a deep neural network framework, trains the network on the training set, and finally applies the trained network to the special effect of the character to be migrated to realize character effect migration.
However, the existing method for transferring the style of special effect words cannot process the decorative elements on the special effect of words, which can cause the loss and distortion of the decorative elements.
Disclosure of Invention
Aiming at the technical problems, the invention provides an intelligent character effect transferring method and system for special effect characters, which can transfer the special effects of characters and simultaneously transfer decorative elements of the characters without losing and distorting the decorative elements.
The technical scheme adopted by the invention is as follows:
an intelligent word effect migration method for special effect words comprises the following steps:
training a mask extraction sub-network to extract a decoration element mask by utilizing a training data set, and training a basic special effect migration sub-network to migrate a basic character special effect;
special effect word D with decorative elements y And its paired character pattern picture C y Inputting the data into the trained mask extraction sub-network to obtain a decorative element mask M y
Will D y And its paired character pattern picture C y And a target font picture C x Inputting the result into a trained basic special effect migration sub-network to obtain a result S of basic special effect migration and decoration element elimination x
Using M y ,C y And C x Carrying out element recombination to fuse decorative elements into S x Obtaining the moved special effect character D with the decorative elements corresponding to the target character pattern x
Further, the training data set comprises a synthesized special effect word picture D with decoration and a matched character pattern picture C thereof, and a collected special effect word picture D with decoration and manually marked character pattern w And the paired character pattern picture C w (ii) a Wherein the picture D is obtained by randomly adding the decoration element picture to a collected or synthesized special effect word picture obtained by randomly combining from a collected texture picture to a glyph picture not containing a special effect.
Furthermore, the mask extraction sub-network adopts a U-net network structure, and a discriminator network netSegD for judging a picture domain where the input picture is located is built, wherein the netSegD comprises four convolution modules, and each convolution module comprises a convolution layer and a linear rectification function.
Further, the training method for mask extraction sub-network comprises the following steps:
combining the D and the C in the channel dimension and inputting the mask to extract the sub-network to obtain the decoration element mask
Figure GDA0003521086410000021
Simultaneously extracting the feature P of the penultimate layer;
will D w And C w Merging the input masks in the channel dimension to extract sub-networks to obtain the decorative element mask
Figure GDA0003521086410000022
Simultaneous extraction of features P of the penultimate layer w
According to
Figure GDA0003521086410000023
P、P w And setting a loss function to obtain a trained mask extraction sub-network.
Further, the basic special effect migration sub-network adopts a multi-scale framework structure and is trained sequentially under three resolutions of 64 × 64, 128 × 128 and 256 × 256.
Further, the training method of the basic special effect migration sub-network comprises the following steps:
will D y 、C y 、C x Inputting the data into a basic special effect migration sub-network, merging the data in a channel domain to obtain a migration result S x
According to S x And setting a loss function to obtain a trained basic special effect migration sub-network.
Further, the method for recombining the elements comprises the following steps:
using DenseCrF for M y Optimizing;
will M y Scaling to a certain resolution, clustering by using DBSCAN, and obtaining different decorative elements and masks thereof by searching for connected shapes;
for in C y Each decorative element belonging to the area E is searched to obtain the corresponding decorative element C x Thereby the decorative element is moved from D to E y Is copied to S x Adjusting the size of the decorative element to obtain D x
Further, for the case of C y Each decorative element belonging to the area E is searched according to the following formula to obtain the corresponding decorative element C x At the suitable position E':
Figure GDA0003521086410000031
wherein, M guide (·),
Figure GDA0003521086410000032
And M Exi (. The) denotes the region at M guide
Figure GDA0003521086410000033
And M Exi The sum of the values of (a) and (b);
M guide the obtaining method comprises the following steps: at C y Upper computation level prior M Hor And vertical prior M Ver Normalized to [0,1 ]]Interval, fuzzy with Gauss fuzzy kernel; at C y Upper computation distribution prior M Dis (ii) a Will obscure M Hor And M Ver And M Dis Combining in channel dimensions to yield M guide
Figure GDA0003521086410000034
Is at C x The same as M guide The method for obtaining.
An intelligent word effect migration system for special effect words, comprising:
the mask extraction sub-network module is used for inputting special effect words with decorative elements and the matching font pictures thereof and obtaining a decorative element mask from the special effect words and the matching font pictures;
a basic special effect migration sub-network module, which is used for inputting special effect words with decorative elements, the character pattern matching pictures thereof and the target character pattern pictures, migrating the basic special effects of the special effect words to the target character patterns, eliminating the decorative elements of the special effect words and outputting results;
and the element recombination module is used for fusing the decoration elements on the result output by the basic special effect migration sub-network module by utilizing the shade obtained by the shade extraction sub-network module to obtain the migrated special effect characters with the decoration elements corresponding to the target character patterns.
The method comprises the steps of firstly extracting a sub-network through a shade to obtain the shade of the decoration element for separating the character special effect and the decoration element, then carrying out basic character special effect migration through a basic special effect migration sub-network, and finally recombining the decoration element and the special effect character after the basic character special effect migration according to the distribution of the decoration element to obtain a special effect migration result. The method can transfer the decorative elements of the character special effect while transferring the special effect of the character, and cannot cause the loss and distortion of the decorative elements.
Drawings
Fig. 1 is a block diagram of a mask extraction sub-network used in the present invention.
Fig. 2 is a block diagram of a basic effect migration sub-network used in the present invention.
Fig. 3 is a block diagram of an intelligent word effect migration framework used in the present invention.
Detailed Description
In order to make the aforementioned and other features and advantages of the invention more comprehensible, embodiments accompanied with figures are described in detail below.
The embodiment discloses an intelligent word effect migration method for special effect words, which is specifically described as follows:
step 1: a set of character special effect picture pair data sets is collected, and one group of picture pairs consists of a special effect character picture and a corresponding character pattern picture without special effect. The special effect words can be generated in batches through a Photoshop special effect word synthesis action. The data set contains 60 special effect character styles, each style containing 19 fonts of English capital and lowercase letters, and corresponding special effect character picture pairs of 988 fonts. The resolution of the special effect word picture and the font picture is 320 multiplied by 320. 85% of the picture pairs are divided as a training set, and 15% of the picture pairs are divided as a test set. 300 cartoon or geometric texture pictures are collected, and synthesized special effect words are obtained through random combination and used as supplement of training. Thereafter, 4000 sheets of vector graphics were collected as decorative elements. In the network training process, the decorative element pictures are randomly added to the collected or synthesized special effect word pictures to synthesize special effect words with decorations as a synthesized special effect word domain. 1000 special effect character pictures are collected from a network and other channels, and characters are manually marked to serve as a real special effect character field.
Step 2: a mask extraction subnetwork is constructed.
The network structure is shown in fig. 1. The mask extraction subnetwork adopts a U-net network structure. In the special effect word synthesizing field, firstly, the synthesized special effect word D with decoration and the character pattern picture C matched with the special effect word D are merged and input into a network in a channel dimension, input data passes through a mask extraction sub-network to obtain a mask of decoration elements
Figure GDA0003521086410000041
And simultaneously extracting the feature P of the penultimate layer. In the real special effect word field, firstly, the special effect word D with decoration is put w Character pattern picture C matched with it w Merging input networks in channel dimension, extracting sub-networks from input data through masks to obtain masks of decorative elements
Figure GDA0003521086410000042
Simultaneously extracting features P of the penultimate layer w . Meanwhile, a discriminator network netSegD is established. The netSegD consists of four convolution modules,each convolution module includes convolution layers and a linear rectification function. netSegD outputs a determination of the picture domain in which the input picture is located.
And 3, step 3: the training mask extracts the sub-network.
Total loss function L mask The device consists of two parts:
L mask =λ seg L segadv L adv
parameter lambda seg Is set to 1, lambda adv Set to 0.01. L is seg Extract the loss function for the mask:
Figure GDA0003521086410000043
parameter lambda L1 Is set to 1, lambda Per Set to 1, M is the labeled decoration element mask of the special effect word picture, VGG (·) is a feature extracted at five layers ReLU1, ReLU2, ReLU3, ReLU4, ReLU5 using a pre-trained classification network VGG.
L adv Extract the penalty-fighting function for the mask:
L adv =-log(netSegD(P w )),
netSegD(P w ) Inputting P for arbiter netSegD w The resulting output. netSegD is trained on the picture domain from which the input features are judged to come, with a loss function of:
Figure GDA0003521086410000044
and 4, step 4: and building a basic special effect migration network.
The network structure is as shown in fig. 2, a multi-scale frame structure is adopted, and the network is difficult to directly train on high resolution, so that a training strategy with gradually improved resolution is adopted, the training difficulty is gradually increased from easy to difficult, and specifically, the network is trained sequentially under three resolutions of 64 × 64, 128 × 128 and 256 × 256. Inputting the special effect word D with synthesized decoration elements y Match the characterFigure C y And target font picture C x . Merging the three input pictures in a channel domain, and obtaining a migration result S after passing through a basic special effect migration network x
And 5: and training a basic special effect migration network. Total loss function L transfer Consists of two parts, using WGAN-GP:
Figure GDA0003521086410000051
parameter lambda L1 Is set to 1, lambda adv Set to 0.01. L is a radical of an alcohol adv To combat the loss function:
Figure GDA0003521086410000052
in the above formula, the first and second carbon atoms are,
Figure GDA0003521086410000053
meaning that the mean is calculated for distribution Y with respect to X. D represents a discriminator network, and D (a, b, c, D) represents the discrimination result of a after the input of a, b, c, D by the discriminator network. The arbiter network adopts the network structure of PatchGAN.
Is a real target special effect word in the data set,
Figure GDA0003521086410000054
in distribution of
Figure GDA0003521086410000055
And S x To be uniform.
Step 6: and building an intelligent word effect migration system framework. The trained mask extraction sub-network and the base special effects migration sub-network are combined as in fig. 3. Within the system framework, a given special effect word D to be migrated with a decorative element is first of all assigned y And its paired font pictures C y Inputting the obtained data to the trained mask extraction sub-network to obtain a decorative element mask M y . Then the special effect word D with the decorative elements is put into y Matching character pattern picture C y And target font picture C x Inputting the result into a basic special effect migration sub-network to obtain a result S of basic special effect migration and decoration element elimination x . Then carrying out element recombination to obtain the special effect word D with the decoration element after the final style migration x
The element recombination part is firstly passed through C y Calculating the level prior M Hor . Definition of x y,min Is C y The pixel position, x, of the foreground font at the leftmost end of each upper line y,max The rightmost end thereof. Then generate
Figure GDA0003521086410000056
Figure GDA0003521086410000057
Wherein, K w =0.06(x y,max -x y,min )。
Thereafter, M is generated by recursion Hor
M Hor (x y,center ,y)=0,
Figure GDA0003521086410000058
x y,center Is represented by C y The center of the foreground font of each upper line is located at the pixel position.
Then M is added Hor Normalized to [0,1 ]]Interval, and fuzzy with Gaussian fuzzy core to obtain final M Hor
Similarly, by C y Calculating vertical prior M Ver . Definition of y x,min Is C y The pixel position of the foreground font at the top of each column, y x,max For its lowest end, an estimate of the vertical variation of the glyph may then be generated
Figure GDA0003521086410000061
Figure GDA0003521086410000062
Wherein, K' w =0.06(y x,max -y x,min )。
Thereafter, M is generated by recursion Ver
M Ver (x,y x,center )=0,
Figure GDA0003521086410000063
y x,center Is represented by C y The center of the foreground font of each column is at the pixel position.
Then M is added Ver Normalized to [0,1 ]]Interval, and fuzzy with Gaussian fuzzy core to obtain final M Ver
At C y Upper computation distribution prior M Dis
M Dis =(1-Dis(x,y)) 5
Dis (x, y) denotes the glyph distribution prior calculated by Yang et al (cf. Shuai Yang, Jianying Liu, Zhouhui Lian, and Zongming Guo. "Awessome type graphics: Statistics-Based Text Effects Transfer", Proc. of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, Hawaii, Jul. 2017.).
Combining M in channel dimension Hor 、M Ver 、M Dis To obtain the total prior M guide . Similarly, at C x The total prior is obtained by the upper calculation
Figure GDA0003521086410000064
Setting a superposed recording matrix M Exi When the element position has a decorative element, M Exi (x, y) is 0, otherwise M Exi (x,y)=1。
The element recombination was performed by first using DenseCrF on M y And (6) optimizing. Then, at M y To get M y Scaled to 5 × 5 resolution and clustered with DBSCAN. Different decorative elements and their shades are obtained by finding connected shapes.For each decorative element, note it at C y Upper part belongs to the area E and is searched to obtain the corresponding area C x At the suitable position E':
Figure GDA0003521086410000065
M guide (·),
Figure GDA0003521086410000066
and M Exi (. The) denotes the region at M guide
Figure GDA0003521086410000067
And M Exi The sum of the values of (a) and (b).
Decorative elements from D y Is copied to S x If the superposition area of the decoration element and the font is reduced after copying, the decoration element is slightly reduced and is closer to the font, otherwise, the decoration element is slightly enlarged and is further away from the font, and the special effect character D with the decoration element after final style migration is obtained x
And 7: special effect word D to be transferred with decorative elements y And its corresponding font picture C y Target font picture C x Inputting the system frame to obtain the special effect character D with the decoration elements after the final style migration x
The above embodiments are only intended to illustrate the technical solution of the present invention and not to limit the same, and a person skilled in the art can modify the technical solution of the present invention or substitute the same without departing from the spirit and scope of the present invention, and the scope of the present invention should be determined by the claims.

Claims (10)

1. An intelligent word effect migration method for special effect words comprises the following steps:
training a mask extraction sub-network to extract a decoration element mask by utilizing a training data set, and training a basic special effect migration sub-network to migrate a basic character special effect;
special effect word D with decorative elements y And its paired font pictures C y Inputting the data into a trained mask extraction sub-network to obtain a decorative element mask M y
Will D y And its paired character pattern picture C y Target font picture C x Inputting the result into a trained basic special effect migration sub-network to obtain a result S of basic special effect migration and decoration element elimination x
Using M y ,C y And C x Carrying out element recombination to fuse decorative elements into S x Obtaining the moved special effect character D with the decorative elements corresponding to the target character pattern x
2. The method of claim 1, wherein the mask extraction sub-network is a U-net network structure, and a discriminator network netSegD is constructed for determining a picture domain in which the input picture is located, the netSegD comprising four convolution modules, each convolution module comprising a convolution layer and a linear rectification function.
3. The method of claim 2, wherein the training data set comprises a composite decorated special effect word picture D and its paired glyph picture C, and a collected and artificially annotated glyph-decorated special effect word picture D w And the paired character pattern picture C w (ii) a Wherein the picture D is obtained by randomly adding the decoration element picture to a collected or synthesized special effect word picture obtained by randomly combining from a collected texture picture to a glyph picture not containing a special effect.
4. The method of claim 3, wherein the training method of the mask extraction subnetwork comprises the steps of:
combining the D and the C in the channel dimension and inputting the shade to extract a sub-network to obtain a decoration element shade
Figure FDA0003521086400000011
Simultaneously extracting the feature P of the penultimate layer;
will D w And C w Merging the input masks in the channel dimension to extract sub-networks to obtain the decorative element mask
Figure FDA0003521086400000012
Simultaneous extraction of features P of the penultimate layer w
According to
Figure FDA0003521086400000013
P、P w And setting a loss function to obtain a trained mask extraction sub-network.
5. The method of claim 4, wherein the method is based on
Figure FDA0003521086400000014
P、P w The method of setting the loss function is as follows:
the loss function comprises a mask extraction loss function L seg Sum mask extraction resisting loss function L adv The total loss function L is composed of the two loss functions mask
L mask =λ seg L segadv L adv
Wherein λ is seg Is set to 1, lambda adv Set to 0.01;
Figure FDA0003521086400000015
wherein λ is L1 Is set to 1, lambda Per Setting as 1, M is a labeled decoration element mask of a special effect word picture, and VGG (-) is a feature extracted by a pre-trained classification network VGG on five layers of ReLU1, ReLU2, ReLU3, ReLU4 and ReLU 5;
L adv =-log(netSegD(P w )),
among them, netSegD (P) w ) Inputting P for arbiter netSegD w The resulting output, netSegD, is trained in the picture domain from which the input features are judged to come from, with a loss function of:
Figure FDA0003521086400000021
6. the method of claim 1, wherein the base special effects migration subnetwork employs a multi-scale framework structure, and the network is sequentially trained with a resolution-escalating training strategy comprising sequential training at three resolutions of 64 x 64, 128 x 128, and 256 x 256.
7. Method according to claim 1 or 2, wherein the training of the base effect migration subnetwork comprises the following steps:
will D y 、C y 、C x Inputting the data into a basic special effect migration sub-network, merging the data in a channel domain to obtain a migration result S x
According to S x And setting a loss function to obtain a trained basic special effect migration sub-network.
8. The method of claim 1, wherein the recombination of elements comprises the steps of:
using DenseCrF for M y Optimizing;
will M y Scaling to a certain resolution, clustering by using DBSCAN, and obtaining different decorative elements and masks thereof by searching for connected shapes;
for in C y Each decorative element belonging to the area E is searched to obtain the corresponding decorative element C x Thereby the decorative element is moved from D to E y Is copied to S x Adjusting the size of the decorative element to obtain D x
9. The method of claim 8Of a process in which y Each decorative element belonging to the area E is searched according to the following formula to obtain the corresponding decorative element C x At the suitable position E':
Figure FDA0003521086400000022
wherein M is gujde (·),
Figure FDA0003521086400000023
And M Exi (. respectively) indicates the regions in M guide
Figure FDA0003521086400000024
And M Exi The sum of the values of (a) and (b);
M guide the obtaining method comprises the following steps: at C y Up-computation level prior M Hor And vertical prior M Ver Normalized to [0,1 ]]Interval, fuzzy with Gauss fuzzy kernel; at C y Upper computation distribution prior M Dis (ii) a Will be blurred M Hor And M Ver And M Dis Combining in channel dimensions to yield M guide
Figure FDA0003521086400000031
Is at C x The same as M guide The method of obtaining.
10. A smart word effect migration system for special effect words, comprising:
the mask extraction sub-network module is used for inputting special effect characters with decorative elements and paired character patterns of the special effect characters and obtaining a decorative element mask from the special effect characters;
a basic special effect migration sub-network module, which is used for inputting special effect words with decorative elements, the character pattern matching pictures thereof and the target character pattern pictures, migrating the basic special effects of the special effect words to the target character patterns, eliminating the decorative elements of the special effect words and outputting results;
and the element recombination module is used for fusing the decoration elements on the result output by the basic special effect migration sub-network module by utilizing the mask obtained by the mask extraction sub-network module to obtain the migrated special effect characters with the decoration elements corresponding to the target character patterns.
CN201910440039.8A 2019-05-24 2019-05-24 Intelligent word effect migration method and system for special effect words Active CN112069769B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910440039.8A CN112069769B (en) 2019-05-24 2019-05-24 Intelligent word effect migration method and system for special effect words

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910440039.8A CN112069769B (en) 2019-05-24 2019-05-24 Intelligent word effect migration method and system for special effect words

Publications (2)

Publication Number Publication Date
CN112069769A CN112069769A (en) 2020-12-11
CN112069769B true CN112069769B (en) 2022-07-26

Family

ID=73658095

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910440039.8A Active CN112069769B (en) 2019-05-24 2019-05-24 Intelligent word effect migration method and system for special effect words

Country Status (1)

Country Link
CN (1) CN112069769B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107577651A (en) * 2017-08-25 2018-01-12 上海交通大学 Chinese character style migratory system based on confrontation network
CN109146989A (en) * 2018-07-10 2019-01-04 华南理工大学 A method of birds and flowers characters in a fancy style image is generated by building neural network

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107577651A (en) * 2017-08-25 2018-01-12 上海交通大学 Chinese character style migratory system based on confrontation network
CN109146989A (en) * 2018-07-10 2019-01-04 华南理工大学 A method of birds and flowers characters in a fancy style image is generated by building neural network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
《Adaptive Batch Normalization for practical domain adaptation》;Li, YH 等;《PATTERN RECOGNITION》;20181228;全文 *
《基于神经风格迁移的字体特效渲染技术》;蔡湧达 等;《电脑知识与技术》;20190228;全文 *

Also Published As

Publication number Publication date
CN112069769A (en) 2020-12-11

Similar Documents

Publication Publication Date Title
CN108846358B (en) Target tracking method for feature fusion based on twin network
Azadi et al. Multi-content gan for few-shot font style transfer
CN104732506B (en) A kind of portrait photographs' Color Style conversion method based on face semantic analysis
CN107844795B (en) Convolutional neural networks feature extracting method based on principal component analysis
CN110032942B (en) Action identification method based on time domain segmentation and feature difference
CN111126256B (en) Hyperspectral image classification method based on self-adaptive space-spectrum multi-scale network
Zhang et al. Content-adaptive sketch portrait generation by decompositional representation learning
CN111723585A (en) Style-controllable image text real-time translation and conversion method
US20190385346A1 (en) Synthesizing new font glyphs from partial observations
CN109285162A (en) A kind of image, semantic dividing method based on regional area conditional random field models
Zheng et al. Attention-based spatial-temporal multi-scale network for face anti-spoofing
CN109064522A (en) The Chinese character style generation method of confrontation network is generated based on condition
CN109359527A (en) Hair zones extracting method and system neural network based
CN105550712B (en) Aurora image classification method based on optimization convolution autocoding network
CN109086777A (en) A kind of notable figure fining method based on global pixel characteristic
CN112967218B (en) Multi-scale image restoration system based on wire frame and edge structure
CN111881716A (en) Pedestrian re-identification method based on multi-view-angle generation countermeasure network
CN110826534B (en) Face key point detection method and system based on local principal component analysis
CN110751271B (en) Image traceability feature characterization method based on deep neural network
CN110472591B (en) Shielded pedestrian re-identification method based on depth feature reconstruction
CN110738715B (en) Automatic migration method of dynamic text special effect based on sample
CN114387610A (en) Method for detecting optional-shape scene text based on enhanced feature pyramid network
CN112069769B (en) Intelligent word effect migration method and system for special effect words
CN113393546A (en) Fashion clothing image generation method based on clothing category and texture pattern control
CN108537855A (en) A kind of ceramic marble paper method for generating pattern and device that sketch is consistent

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant