CN109285111A - A kind of method, apparatus, equipment and the computer readable storage medium of font conversion - Google Patents

A kind of method, apparatus, equipment and the computer readable storage medium of font conversion Download PDF

Info

Publication number
CN109285111A
CN109285111A CN201811101699.5A CN201811101699A CN109285111A CN 109285111 A CN109285111 A CN 109285111A CN 201811101699 A CN201811101699 A CN 201811101699A CN 109285111 A CN109285111 A CN 109285111A
Authority
CN
China
Prior art keywords
font
training
target
picture
standard letter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811101699.5A
Other languages
Chinese (zh)
Other versions
CN109285111B (en
Inventor
刘怡俊
杨培超
叶武剑
翁韶伟
张子文
李学易
王峰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong University of Technology
Original Assignee
Guangdong University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong University of Technology filed Critical Guangdong University of Technology
Priority to CN201811101699.5A priority Critical patent/CN109285111B/en
Publication of CN109285111A publication Critical patent/CN109285111A/en
Application granted granted Critical
Publication of CN109285111B publication Critical patent/CN109285111B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • G06T3/04
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention discloses a kind of methods of font conversion, deep learning model is obtained previously according to standard letter pictures and the training of target font pictures, and in the training process be respectively each target font target character setting and the unique corresponding style embedded block of each target font, i.e. each target character is other than possessing identical character embedded block, respectively also there is unique style embedded block, thus the deep learning model that training obtains is when carrying out font conversion, different target fonts can be exported based on different style embedded blocks.Standard letter picture is further converted to by target font file picture, the text picture of available plurality of target font by the deep learning model that training obtains.The method provided through the invention improves the efficiency of the model of trained multiple fonts and the efficiency of deep learning model conversion new font.The present invention also provides device, equipment and the computer readable storage mediums of a kind of conversion of font, have above-mentioned beneficial effect.

Description

A kind of method, apparatus, equipment and the computer readable storage medium of font conversion
Technical field
The present invention relates to image identifying and processing field, more particularly to a kind of method, apparatus of font conversion, equipment and Computer readable storage medium.
Background technique
With the increase of computer font type, hand-written script is gradually added into computer library, the font of some famous persons It is pursued by many users.It is every in computer library character library to increase a kind of new font, require the picture according to the font Carry out learning training transformation model.Learn a kind of new font by the way of training deep learning model in the prior art, and will This new font increases storage.But in practical applications, it is often inadequate for only generating a kind of font of style, and trains It practises generation model to take a long time, a kind of every font for generating new style requires to remake training set and re -training, then More times can be expended.
Therefore, the training transfer efficiency for generating new font how is improved, is the technology that those skilled in the art need to solve Problem.
Summary of the invention
The object of the present invention is to provide method, apparatus, equipment and the computer readable storage mediums of a kind of conversion of font, use In the training transfer efficiency for improving generation new font.
In order to solve the above technical problems, the present invention provides a kind of method of font conversion, comprising:
Deep learning model is obtained previously according to standard letter pictures and the training of multiple target font pictures, and is being instructed Respectively uniquely corresponding style is embedded in for the target character insertion of each target font and each target font during practicing Block;Wherein, the character embedded block that the target character of each target font includes is identical;
Obtain standard letter text picture;
The standard letter text picture is inputted into the deep learning model, obtains target font file picture.
Optionally, the deep learning model is specially cGAN network model.
Optionally, described to obtain deep learning model according to standard letter pictures and the training of target font pictures, and Respectively uniquely corresponding style is embedded in for the target character insertion of each target font and each target font in the training process Block specifically includes:
The standard letter pictures and the target font pictures are processed to obtain tagged binary system instruction Practice data;
Determine generator, arbiter, the first encoder for being embedded in the genre category block and for calculating penalty values Second encoder, construct original cGAN network model;
Using the binary system training data training original cGAN network model, and according to transformation result and the damage Mistake value adjusting parameter determines the cGAN network model until the penalty values reach the first preset condition.
Optionally, the acquisition standard letter text picture, specifically includes:
Training obtains the convolutional neural networks model of initial text for identification in advance;
The initial text picture is inputted the convolutional neural networks model and obtained by the initial text picture for receiving input The standard letter text picture.
Optionally, the preparatory training obtains the convolutional neural networks model of initial text for identification, specifically includes:
It determines Chinese characters in common use pictures and the corresponding number of each Chinese character, generates the data set for training;
Determine input layer, convolutional layer, down-sampling layer and the output layer of convolutional neural networks;
It is determined by the data set training convolutional neural networks until training parameter reaches the second preset condition The convolutional neural networks model.
Optionally, described that the initial text picture input convolutional neural networks model is obtained into the standard letter Text picture, specifically includes:
The initial text picture is inputted into the convolutional neural networks model, obtains text number;
According to preset corresponding relationship, obtains and export standard letter text picture corresponding with text number.
In order to solve the above technical problems, the present invention also provides a kind of devices of font conversion, comprising:
First training unit, for obtaining depth previously according to standard letter pictures and the training of target font pictures Practise model, and in the training process be respectively each target font target character insertion it is uniquely right with each target font The style embedded block answered;Wherein, the character embedded block that the target character of each target font includes is identical;
First converting unit will be described in standard letter text picture input for obtaining standard letter text picture Deep learning model obtains target font file picture.
Optionally, further includes:
Second training unit obtains the convolutional neural networks model of initial text for identification for training in advance;
The initial text picture is inputted the volume by the second converting unit, initial text picture for receiving input Product neural network model obtains the standard letter text picture.
In order to solve the above technical problems, the present invention also provides a kind of equipment of font conversion, comprising:
Memory, for storing instruction, described instruction include the steps that the method for the conversion of font described in above-mentioned any one;
Processor, for executing described instruction.
In order to solve the above technical problems, being stored thereon with calculating the present invention also provides a kind of computer readable storage medium Machine program realizes the step for the method that the font as described in above-mentioned any one is converted when the computer program is executed by processor Suddenly.
The method of font conversion provided by the present invention, is instructed previously according to standard letter pictures and target font pictures Deep learning model is got, and the target character setting for being in the training process respectively each target font and each target font are only One corresponding style embedded block, even if each target character other than possessing identical character embedded block, respectively also has unique Style embedded block, the deep learning model that thus training obtains can be embedded in when carrying out font conversion by identical character The same character is mapped to the same vector by block, and can export different target fonts based on different style embedded blocks. Standard letter picture is further converted to by target font file picture by the deep learning model that training obtains, it is available The text picture of plurality of target font.The method provided through the invention, can obtain in a model training can convert The deep learning model of the font of a variety of styles improves the efficiency of the model of trained multiple fonts, and passes through the deep learning Standard letter text picture is disposably converted to the target font file picture of multiple fonts by model, improves deep learning mould The efficiency of type conversion new font.The present invention also provides device, equipment and the computer readable storage medium of a kind of conversion of font, tools There is above-mentioned beneficial effect, details are not described herein.
Detailed description of the invention
It, below will be to embodiment or existing for the clearer technical solution for illustrating the embodiment of the present invention or the prior art Attached drawing needed in technical description is briefly described, it should be apparent that, the accompanying drawings in the following description is only this hair Bright some embodiments for those of ordinary skill in the art without creative efforts, can be with root Other attached drawings are obtained according to these attached drawings.
Fig. 1 is a kind of flow chart of the method for font conversion provided in an embodiment of the present invention;
Fig. 2 is a kind of flow chart of the specific embodiment of step S10 provided in an embodiment of the present invention;
Fig. 3 is the flow chart of the method for another font conversion provided in an embodiment of the present invention;
Fig. 4 is a kind of flow chart of the specific embodiment of step S30 provided in an embodiment of the present invention;
Fig. 5 is the schematic diagram of convolutional neural networks model provided in an embodiment of the present invention;
Fig. 6 is a kind of flow chart of the specific embodiment of step S32 provided in an embodiment of the present invention;
Fig. 7 is a kind of schematic diagram of the device of font conversion provided in an embodiment of the present invention;
Fig. 8 is the schematic diagram of the device of another font conversion provided in an embodiment of the present invention;
Fig. 9 is a kind of structural schematic diagram of the equipment of font conversion provided in an embodiment of the present invention.
Specific embodiment
Core of the invention is to provide method, apparatus, equipment and the computer readable storage medium of a kind of font conversion, uses In the training transfer efficiency for improving generation new font.
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete Site preparation description, it is clear that described embodiments are only a part of the embodiments of the present invention, instead of all the embodiments.It is based on Embodiment in the present invention, it is obtained by those of ordinary skill in the art without making creative efforts every other Embodiment shall fall within the protection scope of the present invention.
Fig. 1 is a kind of flow chart of the method for font conversion provided in an embodiment of the present invention.Fig. 2 mentions for the embodiment of the present invention A kind of flow chart of the specific embodiment of the step S10 supplied.
As shown in Figure 1, the method for font conversion includes:
S10: obtaining deep learning model previously according to standard letter pictures and the training of multiple target font pictures, and It is in the training process respectively target character insertion and the unique corresponding style embedded block of each target font of each target font.
Wherein, the character embedded block that the target character of each target font includes is identical.
In specific implementation, deep learning model can specifically use cGAN network model.Through experiments, it was found that cGAN network The effect that model learns font and converts is better than other models.CGAN, that is, conditional GAN, can learn the image x from With generation model G: i.e. { x, z } → y of random noise vector z to y mapping.As shown in figure 3, G (x) is the mould in training process Type, y are object module, and generator G is that cannot be differentiated by the arbiter D Jing Guo antagonistic training by training, the output of generation The image of " true " out, and arbiter D detects the image of the "false" of generator by training as much as possible.That is generator The target of G be generate the unresolvable image of arbiter D, and the target of arbiter D be tell the received picture of institute whether be The process of true picture, their games is exactly the process of e-learning, when two sides reach balance, generator G energy Generate the picture of high quality.
It is in the training process respectively target character insertion and the unique corresponding style of each target font of each target font Embedded block, and since the character embedded block that the target character of each target font includes is identical, deep learning model is in training process The same vector can be mapped to according to identical character in conversion process, further according to respective style embedded block, output is not Same target character.
As shown in Fig. 2, step S10 can specifically include:
S20: standard letter pictures and target font pictures are processed to obtain tagged binary system training number According to.
Standard letter can be regular script, and standard letter pictures can preparatory self manufacture.
S21: generator, arbiter, the first encoder for being embedded in genre category block and for calculating penalty values are determined Second encoder, construct original cGAN network model.
Building generator is U-Net network structure, the cGAN network model that arbiter is convolutional neural networks model, simultaneously The first encoder for being embedded in a variety of genre categories and the second encoder for calculating loss is added.
Wherein, the first encoder (classification encoder) is used to carry out style using convolutional neural networks model to random noise The extraction of feature encodes.Multiple fonts this one are converted to by a kind of font since common U-Net and GAN model can not be handled Uncertainty in many-many relationship, the embodiment of the present invention introduce the first coding of a variety of genre categories of insertion in generator Device links together before by decoder using a non-training Gaussian noise as style embedded block and character embedded block. By the method, the same character still can be mapped to the same vector by encoder, and but then, decoder will will use The content of two kinds of characters and style is embedded in generate target character.
Second encoder (auxiliary coder) principle is similar to the first encoder, focus on to generator generate picture into Row feature extraction, carries out the subsequent calculating for generating picture loss, this loss is equally used for the learning process of constraint cGAN.
Penalty values herein are calculated with L1 Regularization function:
LL1(G)=Ex,y,z[||y-G(x,z)||1] (1)
The loss majorized function of last entire cGAN network model are as follows:
Wherein, x indicates that source picture, y indicate that the picture that generator generates, z indicate Gaussian noise, and G indicates the letter of generator Number, D indicate the function of arbiter, LcGANIt is the loss function of cGAN, Ex,y,zIt is expectation value function, G*It is objective optimization function, λ It is scale factor.
S22: it using the original cGAN network model of binary system training data training, and is adjusted according to transformation result and penalty values Parameter determines cGAN network model until penalty values reach the first preset condition.
The repetition training of cGAN network model, observation conversion are carried out using binary system training data obtained in step S20 The effect of obtained text picture, and the text in text picture is calculated compared to target word according to formula (1) and formula (2) The penalty values of the corresponding text of body picture look after and guide network settings parameter repeatedly, test tuning.When the numerical value symbol that loss function obtains When closing expected, compares true picture and generate picture, determine the final cGAN network model for font conversion.
S11: standard letter text picture is obtained.
Receive the standard letter text picture of other units transmission.
S12: standard letter text picture is inputted into deep learning model, obtains target font file picture.
By trained deep learning model, the text for being once converted to multiple fonts is realized.
Optionally, target font file picture is exported.
In specific implementation, target font file picture is shown in human-computer interaction interface or is sent to designated position.
It can also include: to receive the user setting selection font to be exported, be exported by the first encoder corresponding selection Font.
The method of font conversion provided in an embodiment of the present invention, previously according to standard letter pictures and target font picture Training gets deep learning model, and is in the training process respectively the target character setting and each target word of each target font The unique corresponding style embedded block of body, even if each target character other than possessing identical character embedded block, respectively also has only One style embedded block, the deep learning model that thus training obtains can pass through identical character when carrying out font conversion The same character is mapped to the same vector by embedded block, and can export different targets based on different style embedded blocks Font.Standard letter picture is further converted to by target font file picture by the deep learning model that training obtains, it can To obtain the text picture of plurality of target font.The method provided through the embodiment of the present invention, can be in a model training The deep learning model that can convert the font of a variety of styles is obtained, improves the efficiency of the model of trained multiple fonts, and lead to The target font file picture that standard letter text picture is disposably converted to multiple fonts by the deep learning model is crossed, is improved The efficiency of deep learning model conversion new font.
Fig. 3 is the flow chart of the method for another font conversion provided in an embodiment of the present invention.Fig. 4 is the embodiment of the present invention A kind of flow chart of the specific embodiment of the step S30 provided.Fig. 5 is convolutional neural networks mould provided in an embodiment of the present invention The schematic diagram of type.Fig. 6 is a kind of flow chart of the specific embodiment of step S32 provided in an embodiment of the present invention.
As shown in figure 3, on the basis of the above embodiments, in another embodiment, the step S11 of the method for font conversion It specifically includes:
S30: training obtains the convolutional neural networks model of initial text for identification in advance.
It should be noted that step S30 and step S10 out-of-order relationship.
In order to further user-friendly, it may provide the user with handwriting input text and be then converted into another text Mode.This needs first to identify the text of user's handwriting input, i.e., to the image of the initial text of user's input carry out processing and Identification.In traditional character image process field, need to carry out the pretreatment such as String localization, Character segmentation of Chinese character.Character Partitioning algorithm mainly includes projection and connected domain analysis.And mainstream Chinese character recognition system generally use extract feature of Chinese characters structure and The two ways of statistical nature.It, can be accurate but since the personal writing style of hand-written script is different, it is totally different to write environment The feature of Chinese characters structure for extracting various handwritten forms is nearly impossible.So being all based on the method that statistical nature is classified It is identified, there are several types of: elastic mesh feature, directional element features, Gabor are special for currently used Chinese character statistical nature Sign, moment characteristics.
In order to improve the accuracy of the identification to initial text, as shown in figure 4, step S30 can specifically include:
S40: determining Chinese characters in common use pictures and the corresponding number of each Chinese character, generates the data set for training.
In specific implementation, the hand-written pictures of Chinese characters in common use CSAIA-HWDB1.1 and each Chinese characters in common use pair can be applied Data set of the number label production answered for training, is divided into training set and test set for data set with the ratio of 10:1.
S41: input layer, convolutional layer, down-sampling layer and the output layer of convolutional neural networks are determined.
As shown in figure 5, convolutional neural networks model may include input layer (input), 5 convolutional layers (conv1~5), 4 A down-sampling layer (pool1~4) and the output layer (output) connected entirely, wherein the activation primitive of full linking layer uses ReLU letter Number, other convolution kernel sizes are using 3 × 3.
S42: convolution is determined until training parameter reaches the second preset condition by data set training convolutional neural networks Neural network model.
It specifically can be by hand-written script data set Dropout (i.e. in the training process of deep learning network, for mind Through network unit, it is temporarily abandoned from network according to certain probability) method carry out network training, repetition training volume Product neural network model shows training parameter after the accuracy that frequency of training reaches preset threshold or test reaches target value The second preset condition is had reached, determines convolutional neural networks model at this time for the convolutional neural networks model of final application.
S31: the initial text picture of input is received.
It is specifically as follows the initial text picture for receiving user's handwriting input.
S32: initial text picture input convolutional neural networks model is obtained into standard letter text picture.
In specific implementation, as shown in fig. 6, step S32 may include:
S60: inputting convolutional neural networks model for initial text picture, obtains text number.
S61: it according to preset corresponding relationship, obtains and exports standard letter text picture corresponding with text number.
Due to trained convolutional neural networks model output be text number, and the input of cGAN network model and Output must be same data, therefore can correspond the number of text and UNICODE code, in addition UNICODE code is raw At the functional block of corresponding standard letter text picture.
The method of font conversion provided in an embodiment of the present invention provides a kind of hand-written to another kind by a kind of hand-written script The method of font conversion, extends the function of font conversion.
The corresponding each embodiment of method of font conversion as detailed above, on this basis, the invention also discloses with The device of the corresponding font conversion of the above method.
Fig. 7 is a kind of schematic diagram of the device of font conversion provided in an embodiment of the present invention.As shown in fig. 7, the present invention is real Applying the device that the font that example provides is converted includes:
First training unit 701, for obtaining depth previously according to standard letter pictures and the training of target font pictures Spend learning model, and in the training process be respectively each target font target character insertion it is uniquely right with each target font The style embedded block answered;Wherein, the character embedded block that the target character of each target font includes is identical;
Standard letter text picture is inputted depth for obtaining standard letter text picture by the first converting unit 702 Model is practised, target font file picture is obtained.
Fig. 8 is the schematic diagram of the device of another font conversion provided in an embodiment of the present invention.As shown in figure 8, above-mentioned On the basis of embodiment, in another embodiment, the device of font conversion further include:
Second training unit 801 obtains the convolutional neural networks model of initial text for identification for training in advance;
Initial text picture is inputted convolutional Neural by the second converting unit 802, initial text picture for receiving input Network model obtains standard letter text picture.
Since the embodiment of device part is corresponded to each other with the embodiment of method part, the embodiment of device part is asked Referring to the description of the embodiment of method part, wouldn't repeat here.
Fig. 9 is a kind of structural schematic diagram of the equipment of font conversion provided in an embodiment of the present invention.As shown in figure 9, the word The equipment of body conversion can generate bigger difference because configuration or performance are different, may include one or more processors (central processing units, CPU) 910 (for example, one or more processors) and memory 920, one Or (such as one or more mass memories are set the storage medium 930 of more than one storage application program 933 or data 932 It is standby).Wherein, memory 920 and storage medium 930 can be of short duration storage or persistent storage.It is stored in the journey of storage medium 930 Sequence may include one or more modules (diagram does not mark), and each module may include to a series of in computing device Instruction operation.Further, processor 910 can be set to communicate with storage medium 930, in the equipment 900 of font conversion The upper series of instructions operation executed in storage medium 930.
Font conversion equipment 900 can also include one or more power supplys 940, one or more it is wired or Radio network interface 950, one or more input/output interfaces 990, and/or, one or more operating systems 931, such as Windows ServerTM, Mac OS XTM, UnixTM,LinuxTM, FreeBSDTMEtc..
Step in the method for the conversion of font described in above-mentioned Fig. 1 to Fig. 6 is based on Fig. 9 institute by the equipment that font is converted The structure shown is realized.
It is apparent to those skilled in the art that for convenience and simplicity of description, the font of foregoing description The equipment of conversion and the specific work process of computer readable storage medium, can be with reference to corresponding in preceding method embodiment Journey, details are not described herein.
In several embodiments provided herein, it should be understood that disclosed method, apparatus, equipment and calculating Machine readable storage medium storing program for executing, may be implemented in other ways.For example, Installation practice described above is only schematic , for example, the division of module, only a kind of logical function partition, there may be another division manner in actual implementation, such as Multiple module or components can be combined or can be integrated into another system, or some features can be ignored or not executed.Separately A bit, shown or discussed mutual coupling, direct-coupling or communication connection can be through some interfaces, device Or the indirect coupling or communication connection of module, it can be electrical property, mechanical or other forms.Module as illustrated by the separation member It may or may not be physically separated, the component shown as module may or may not be physics mould Block, it can it is in one place, or may be distributed on multiple network modules.It can be selected according to the actual needs In some or all of the modules achieve the purpose of the solution of this embodiment.
It, can also be in addition, can integrate in a processing module in each functional module in each embodiment of the application It is that modules physically exist alone, can also be integrated in two or more modules in a module.Above-mentioned integrated mould Block both can take the form of hardware realization, can also be realized in the form of software function module.
If integrated module is realized and when sold or used as an independent product in the form of software function module, can To be stored in a computer readable storage medium.Based on this understanding, the technical solution of the application substantially or Say that all or part of the part that contributes to existing technology or the technical solution can embody in the form of software products Out, which is stored in a storage medium, including some instructions are used so that a computer equipment The whole of (can be personal computer, funcall device or the network equipment etc.) execution each embodiment method of the application Or part steps.And storage medium above-mentioned include: USB flash disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic or disk etc. are various can store program The medium of code.
Above to a kind of method, apparatus, equipment and the computer readable storage medium of font conversion provided by the present invention It is described in detail.Each embodiment is described in a progressive manner in specification, the highlights of each of the examples are With the difference of other embodiments, the same or similar parts in each embodiment may refer to each other.Embodiment is disclosed Device for, since it is corresponded to the methods disclosed in the examples, so be described relatively simple, related place is referring to method Part illustrates.It should be pointed out that for those skilled in the art, before not departing from the principle of the invention It puts, can be with several improvements and modifications are made to the present invention, these improvement and modification also fall into the guarantor of the claims in the present invention It protects in range.
It should also be noted that, in the present specification, relational terms such as first and second and the like be used merely to by One entity or operation are distinguished with another entity or operation, without necessarily requiring or implying these entities or operation Between there are any actual relationship or orders.Moreover, the terms "include", "comprise" or its any other variant meaning Covering non-exclusive inclusion, so that the process, method, article or equipment for including a series of elements not only includes that A little elements, but also including other elements that are not explicitly listed, or further include for this process, method, article or The intrinsic element of equipment.In the absence of more restrictions, the element limited by sentence "including a ...", is not arranged Except there is also other identical elements in the process, method, article or apparatus that includes the element.

Claims (10)

1. a kind of method of font conversion characterized by comprising
Deep learning model is obtained previously according to standard letter pictures and the training of multiple target font pictures, and was being trained It is respectively target character insertion and the unique corresponding style embedded block of each target font of each target font in journey;Its In, the character embedded block that the target character of each target font includes is identical;
Obtain standard letter text picture;
The standard letter text picture is inputted into the deep learning model, obtains target font file picture.
2. the method according to claim 1, wherein the deep learning model is specially cGAN network model.
3. according to the method described in claim 2, it is characterized in that, described according to standard letter pictures and target font picture Training gets deep learning model, and respectively the target character of each target font is embedded in and each mesh in the training process The unique corresponding style embedded block of marking-up body, specifically includes:
The standard letter pictures and the target font pictures are processed to obtain tagged binary system training number According to;
Determine generator, arbiter, the first encoder for being embedded in the genre category block and for calculating penalty values Two encoders construct original cGAN network model;
Using the binary system training data training original cGAN network model, and according to transformation result and the penalty values Adjusting parameter determines the cGAN network model until the penalty values reach the first preset condition.
4. the method according to claim 1, wherein the acquisition standard letter text picture, specifically includes:
Training obtains the convolutional neural networks model of initial text for identification in advance;
The initial text picture for receiving input obtains the initial text picture input convolutional neural networks model described Standard letter text picture.
5. according to the method described in claim 4, it is characterized in that, the preparatory training obtains the volume of initial text for identification Product neural network model, specifically includes:
It determines Chinese characters in common use pictures and the corresponding number of each Chinese character, generates the data set for training;
Determine input layer, convolutional layer, down-sampling layer and the output layer of convolutional neural networks;
By the data set training convolutional neural networks, until training parameter reaches the second preset condition, determine described in Convolutional neural networks model.
6. according to the method described in claim 4, it is characterized in that, described input the convolution mind for the initial text picture The standard letter text picture is obtained through network model, is specifically included:
The initial text picture is inputted into the convolutional neural networks model, obtains text number;
According to preset corresponding relationship, obtains and export standard letter text picture corresponding with text number.
7. a kind of device of font conversion characterized by comprising
First training unit, for obtaining deep learning mould previously according to standard letter pictures and the training of target font pictures Type, and in the training process be respectively each target font target character insertion it is uniquely corresponding with each target font Style embedded block;Wherein, the character embedded block that the target character of each target font includes is identical;
The standard letter text picture is inputted the depth for obtaining standard letter text picture by the first converting unit Learning model obtains target font file picture.
8. device according to claim 7, which is characterized in that further include:
Second training unit obtains the convolutional neural networks model of initial text for identification for training in advance;
The initial text picture is inputted the convolution mind by the second converting unit, initial text picture for receiving input The standard letter text picture is obtained through network model.
9. a kind of equipment of font conversion characterized by comprising
Memory, for storing instruction, described instruction include the method for the conversion of font described in claim 1 to 6 any one Step;
Processor, for executing described instruction.
10. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the computer program The step of method that the font as described in claim 1 to 6 any one is converted is realized when being executed by processor.
CN201811101699.5A 2018-09-20 2018-09-20 Font conversion method, device, equipment and computer readable storage medium Active CN109285111B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811101699.5A CN109285111B (en) 2018-09-20 2018-09-20 Font conversion method, device, equipment and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811101699.5A CN109285111B (en) 2018-09-20 2018-09-20 Font conversion method, device, equipment and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN109285111A true CN109285111A (en) 2019-01-29
CN109285111B CN109285111B (en) 2023-05-09

Family

ID=65181771

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811101699.5A Active CN109285111B (en) 2018-09-20 2018-09-20 Font conversion method, device, equipment and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN109285111B (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110211203A (en) * 2019-06-10 2019-09-06 大连民族大学 The method of the Chinese character style of confrontation network is generated based on condition
CN110211032A (en) * 2019-06-06 2019-09-06 北大方正集团有限公司 Generation method, device and the readable storage medium storing program for executing of chinese character
CN110852042A (en) * 2019-12-13 2020-02-28 北京华宇信息技术有限公司 Character type conversion method and device
CN110866543A (en) * 2019-10-18 2020-03-06 支付宝(杭州)信息技术有限公司 Picture detection and picture classification model training method and device
CN111242114A (en) * 2020-01-08 2020-06-05 腾讯科技(深圳)有限公司 Character recognition method and device
CN111402367A (en) * 2020-03-27 2020-07-10 维沃移动通信有限公司 Image processing method and electronic equipment
CN112417959A (en) * 2020-10-19 2021-02-26 上海臣星软件技术有限公司 Picture generation method and device, electronic equipment and computer storage medium
CN112861471A (en) * 2021-02-10 2021-05-28 上海臣星软件技术有限公司 Object display method, device, equipment and storage medium
CN112966470A (en) * 2021-02-23 2021-06-15 北京三快在线科技有限公司 Character generation method and device, storage medium and electronic equipment
CN113807430A (en) * 2021-09-15 2021-12-17 网易(杭州)网络有限公司 Model training method and device, computer equipment and storage medium
WO2023125379A1 (en) * 2021-12-29 2023-07-06 北京字跳网络技术有限公司 Character generation method and apparatus, electronic device, and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2000271317A1 (en) * 2000-02-09 2001-11-01 Amtrol Licensing Inc. Full jacket gas cylinder
CN107506736A (en) * 2017-08-29 2017-12-22 北京大生在线科技有限公司 Online education video fineness picture intercept method based on deep learning
WO2018081089A1 (en) * 2016-10-26 2018-05-03 Deepmind Technologies Limited Processing text sequences using neural networks
CN108170649A (en) * 2018-01-26 2018-06-15 广东工业大学 A kind of Hanzi font library generation method and device based on DCGAN depth networks
US20180261071A1 (en) * 2017-03-10 2018-09-13 Turing Video, Inc. Surveillance method and system based on human behavior recognition

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
ATE366389T1 (en) * 2000-02-09 2007-07-15 Amtrol Inc FULL LINING FOR GAS CYLINDERS

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2000271317A1 (en) * 2000-02-09 2001-11-01 Amtrol Licensing Inc. Full jacket gas cylinder
WO2018081089A1 (en) * 2016-10-26 2018-05-03 Deepmind Technologies Limited Processing text sequences using neural networks
US20180261071A1 (en) * 2017-03-10 2018-09-13 Turing Video, Inc. Surveillance method and system based on human behavior recognition
CN107506736A (en) * 2017-08-29 2017-12-22 北京大生在线科技有限公司 Online education video fineness picture intercept method based on deep learning
CN108170649A (en) * 2018-01-26 2018-06-15 广东工业大学 A kind of Hanzi font library generation method and device based on DCGAN depth networks

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
古险峰等: "基于深度无监督学习的图像分类算法", 《平顶山学院学报》 *

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110211032B (en) * 2019-06-06 2021-02-09 北大方正集团有限公司 Chinese character generating method and device and readable storage medium
CN110211032A (en) * 2019-06-06 2019-09-06 北大方正集团有限公司 Generation method, device and the readable storage medium storing program for executing of chinese character
CN110211203A (en) * 2019-06-10 2019-09-06 大连民族大学 The method of the Chinese character style of confrontation network is generated based on condition
CN110866543A (en) * 2019-10-18 2020-03-06 支付宝(杭州)信息技术有限公司 Picture detection and picture classification model training method and device
CN110866543B (en) * 2019-10-18 2022-07-15 支付宝(杭州)信息技术有限公司 Picture detection and picture classification model training method and device
CN110852042A (en) * 2019-12-13 2020-02-28 北京华宇信息技术有限公司 Character type conversion method and device
CN111242114B (en) * 2020-01-08 2023-04-07 腾讯科技(深圳)有限公司 Character recognition method and device
CN111242114A (en) * 2020-01-08 2020-06-05 腾讯科技(深圳)有限公司 Character recognition method and device
CN111402367A (en) * 2020-03-27 2020-07-10 维沃移动通信有限公司 Image processing method and electronic equipment
CN111402367B (en) * 2020-03-27 2023-09-26 维沃移动通信有限公司 Image processing method and electronic equipment
CN112417959A (en) * 2020-10-19 2021-02-26 上海臣星软件技术有限公司 Picture generation method and device, electronic equipment and computer storage medium
CN112861471A (en) * 2021-02-10 2021-05-28 上海臣星软件技术有限公司 Object display method, device, equipment and storage medium
CN112966470A (en) * 2021-02-23 2021-06-15 北京三快在线科技有限公司 Character generation method and device, storage medium and electronic equipment
CN113807430A (en) * 2021-09-15 2021-12-17 网易(杭州)网络有限公司 Model training method and device, computer equipment and storage medium
CN113807430B (en) * 2021-09-15 2023-08-08 网易(杭州)网络有限公司 Model training method, device, computer equipment and storage medium
WO2023125379A1 (en) * 2021-12-29 2023-07-06 北京字跳网络技术有限公司 Character generation method and apparatus, electronic device, and storage medium

Also Published As

Publication number Publication date
CN109285111B (en) 2023-05-09

Similar Documents

Publication Publication Date Title
CN109285111A (en) A kind of method, apparatus, equipment and the computer readable storage medium of font conversion
CN107122375B (en) Image subject identification method based on image features
CN112464993B (en) Multi-mode model training method, device, equipment and storage medium
CN110232439B (en) Intention identification method based on deep learning network
CN107480688B (en) Fine-grained image identification method based on zero sample learning
CN107391495B (en) Sentence alignment method of bilingual parallel corpus
CN109993164A (en) A kind of natural scene character recognition method based on RCRNN neural network
CN104463101A (en) Answer recognition method and system for textual test question
CN114022882B (en) Text recognition model training method, text recognition device, text recognition equipment and medium
CN113298151A (en) Remote sensing image semantic description method based on multi-level feature fusion
CN112819686A (en) Image style processing method and device based on artificial intelligence and electronic equipment
CN109145946B (en) Intelligent image recognition and description method
CN115700515A (en) Text multi-label classification method and device
CN114528835A (en) Semi-supervised specialized term extraction method, medium and equipment based on interval discrimination
CN114299512A (en) Zero-sample small seal character recognition method based on Chinese character etymon structure
CN113656547A (en) Text matching method, device, equipment and storage medium
CN113806645A (en) Label classification system and training system of label classification model
CN113688955B (en) Text recognition method, device, equipment and medium
CN113609819B (en) Punctuation mark determination model and determination method
CN114153942A (en) Event time sequence relation extraction method based on dynamic attention mechanism
CN113761188A (en) Text label determination method and device, computer equipment and storage medium
CN109034279A (en) Handwriting model training method, hand-written character recognizing method, device, equipment and medium
WO1996008787A1 (en) System and method for automatic subcharacter unit and lexicon generation for handwriting recognition
CN111523325A (en) Chinese named entity recognition method based on strokes
CN113553847A (en) Method, device, system and storage medium for parsing address text

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant