CN112364838B - Method for improving handwriting OCR performance by utilizing synthesized online text image - Google Patents

Method for improving handwriting OCR performance by utilizing synthesized online text image Download PDF

Info

Publication number
CN112364838B
CN112364838B CN202011429519.3A CN202011429519A CN112364838B CN 112364838 B CN112364838 B CN 112364838B CN 202011429519 A CN202011429519 A CN 202011429519A CN 112364838 B CN112364838 B CN 112364838B
Authority
CN
China
Prior art keywords
content
image
style
data set
encoder
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011429519.3A
Other languages
Chinese (zh)
Other versions
CN112364838A (en
Inventor
陈俊杰
杨淑爱
黄坤山
谢克庆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Foshan Nanhai Guangdong Technology University CNC Equipment Cooperative Innovation Institute
Foshan Guangdong University CNC Equipment Technology Development Co. Ltd
Original Assignee
Foshan Nanhai Guangdong Technology University CNC Equipment Cooperative Innovation Institute
Foshan Guangdong University CNC Equipment Technology Development Co. Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Foshan Nanhai Guangdong Technology University CNC Equipment Cooperative Innovation Institute, Foshan Guangdong University CNC Equipment Technology Development Co. Ltd filed Critical Foshan Nanhai Guangdong Technology University CNC Equipment Cooperative Innovation Institute
Priority to CN202011429519.3A priority Critical patent/CN112364838B/en
Publication of CN112364838A publication Critical patent/CN112364838A/en
Application granted granted Critical
Publication of CN112364838B publication Critical patent/CN112364838B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/22Character recognition characterised by the type of writing
    • G06V30/226Character recognition characterised by the type of writing of cursive writing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/04Context-preserving transformations, e.g. by using an importance map
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Multimedia (AREA)
  • Character Discrimination (AREA)
  • Image Processing (AREA)

Abstract

The invention provides a method for improving the performance of handwritten OCR (optical character recognition) by utilizing a synthesized online text image, which comprises the following steps: s1, selecting and dividing a data set, and adopting an IAM data set, wherein the IAM data set comprises an IAM handwriting data set and an IAM online handwriting data set; s2, constructing a generator of the style GAN network, wherein the generator comprises three parts, namely a content encoder, a content decoder and a style encoder; s3, training a generator in the network; and S4, synthesizing the text image in the online data set through the trained generator network model. The handwriting image generated by the framework can effectively improve the OCR recognition accuracy, and a feasible alternative scheme is provided for collecting and constructing a large-scale handwriting data set.

Description

Method for improving handwriting OCR performance by utilizing synthesized online text image
Technical Field
The invention relates to the technical field of image processing, in particular to a method for improving handwriting OCR performance by utilizing a synthesized online text image.
Background
The generation of confrontational networks (GAN) has become a popular research direction in the deep learning field. The GAN network actually contains 2 networks, one being a Generator network (Generator) and the other being a Discriminator network (Discriminator). The two networks may be neural networks, from convolutional neural networks, recurrent neural networks to autoencoders. In this configuration, the two networks participate in a competitive game and attempt to overtake each other while helping them complete their own tasks. After thousands of iterations, if everything is successful, the generator network, i.e., the generator, can perfectly generate a realistic false image, and the discriminator network, i.e., the discriminator, can well judge whether the image is real or false. The core of GAN is actually the generator, and as for why there is a discriminator, mainly to introduce countertraining by which the generator network can generate high quality pictures. The generator aims to generate pictures which are as vivid as possible in the training process to enable the discriminator to judge whether the picture is a real picture or a generated false picture, and the discriminator aims to distinguish the real picture and the false picture as far as possible in the training process, so that the generator hopes that the error rate of the discriminator is maximized, the discriminator hopes that the error rate of the discriminator is minimized, and the discriminator are mutually confronted and progress in competition.
In the field of handwriting OCR, an ideal training set should be able to cover various writing styles, background and light and shadow variations, all possible occurring vocabularies, etc. for a deep learning-based handwriting OCR engine. However, acquiring such a training set is time consuming and labor intensive, and in some cases severely limits the accuracy of handwriting OCR recognition. The importance of the composition of the handwritten image is thus apparent. Owing to the recent development of antagonistic generation networks, many scholars have proposed methods for generating handwritten-style text lines from text or printed-style text lines. However, the handwriting style generated by the method is still not rich enough, and the style of the handwritten characters is lacked. Another idea of handwritten text line generation is to convert online handwritten data to offline images. The online data can be conveniently acquired by devices such as a mobile phone and a writing board, the data volume is large, the style is variable, and if the online data can be converted into a vivid offline handwritten image, the training of handwriting OCR can be well assisted.
Disclosure of Invention
The invention aims to provide a method for improving handwriting performance by utilizing a synthesized online text image, so as to ensure the recognition accuracy of a handwriting OCR (optical character recognition), and convert the handwriting image into a vivid offline handwriting image for assisting OCR training.
To achieve the above object, the present invention provides a method for improving the performance of handwritten OCR by using a synthesized online text image, comprising the steps of:
s1, selecting and dividing a data set: adopting an IAM data set, wherein the IAM data set comprises an IAM handwriting data set and an IAM online handwriting data set; the real image I gt Obtaining a real handwritten image I after data processing sty, And storing the data in an IAM handwriting data set; s2, constructing a generator of the style GAN network, wherein the generator comprises a content encoder, a content decoder and a style encoder;
step S3, training a generator of the GAN network: for real hand-written image I sty Performing skeletonization operation to obtain a skeleton diagram I ske Then the skeleton diagram I is drawn ske The content features are extracted by inputting the content into the content encoder, a feature map is generated and output to a content decoder; will be a true hand-written image I sty Inputting the style features into a style encoder to extract style features, and outputting a 512-dimensional style vector s obtained through global pooling; the content decoder receives the characteristic diagram from the content encoder, receives the component obtained by affine transformation of the style vector s from the style encoder, changes the distribution of the characteristic diagram on the level of the characteristic diagram by adopting the operation of AdaIN, blends style information into the characteristic diagram, and outputs a composite diagram I syn
S4, synthesizing the text images in the online data set through the trained generator: and converting the IAM online handwritten data into a skeleton diagram, selecting a test set from the IAM handwritten data set as a style diagram, inputting the test set into a generator together, and generating an offline handwritten image.
Further, in the step S2, the content encoder, by taking advantage of the network structure of VGG-19, is composed of a plurality of pooling layers and convolutional layers for down-sampling the input; the content encoder includes five convolutional layer modules and three full-link layers, respectively.
Further, in the step S3, the AdaIN changes the data distribution on the aspect diagram level, and may implement the effect of style migration by controlling and changing the affine parameters in the AdaIN layer; the AdaIN layer changes the distribution of the characteristic diagram inside the network, hands over the task of the format migration to AdaIN, and realizes other tasks on the network structure, wherein AdaIN is operated as follows:
Figure 101159DEST_PATH_IMAGE001
the feature map after the content picture coding represented by X is represented by a style vector s through affine sigma and mu, and is a mean value and a standard deviation.
Further, in step S2, the structure of the content decoder is similar to that of the decoder in U-Net, the structure is symmetrical to that of the content encoder, and a composite handwritten text picture with the same input size is generated by using a multilayer convolution and a bilinear upsampling layer; the content decoder includes five convolutional layer modules.
Figure 947892DEST_PATH_IMAGE002
Further, in the step S3, for the skeleton diagram I with the height H and the width W ske The content encoder converts the feature map into a feature map, and the input of the style encoder is a handwritten image I from the true sty Outputting a 512-dimensional style vector s obtained through global pooling; the content encoder generates an image I with a height H and a width W using multiple layers of convolution and bilinear upsampling layers syn The style vector s output by the style encoder adjusts the feature map of the intermediate layer of the content decoder in the mode of AdaIN so as to integrate the style into the final output I of the content decoder syn
Further, in the step S3, by adopting three loss functions including a content loss function, a perceptual loss function and an antagonistic loss function, the synthetic picture generated by the training generator is as realistic as possible and has a style corresponding to a real picture;
the content loss function accepts the difference from the content feature pixel level between the synthesized text image and the real image to optimize the best parameters, and the generator generates I the synthesized image syn The true hand-written image is I gt
Figure 751900DEST_PATH_IMAGE003
Figure 684084DEST_PATH_IMAGE004
The perceptual loss function solves the problem that the content loss function causes the generated image to be excessively smooth. Using a pre-trained VGG-19 network on ImageNet, in the feature space after relu activation function after the first convolution layer of five convolution modules:
calculating a composite map I syn And a real graph I gt The characteristic difference is as follows:
Figure 966161DEST_PATH_IMAGE005
MSE represents a mean square error function, and the values of alpha are respectively 1/32,1/16, 1/8, 1/4 and 1;
the countermeasure loss function adopts the method of patchGAN, and the arbiter network of former GAN network has been changed into the full convolution network, and patchGAN's arbiter network differentiates a small region of input image receptive field and outputs, makes the model more can focus on image details, and the convolution that superposes one by one finally outputs N × N's matrix, and each element wherein actually represents a bigger receptive field in the original image, corresponds to the patch in the original image, specifically as follows:
Figure 932980DEST_PATH_IMAGE006
d represents a network of discriminators for determining the probability of the comparison between the two being true or false.
Compared with the prior art, the invention has the beneficial effects that: a GAN frame is provided, which adopts the structure of an encoder-decoder, and uses a style encoder to extract style characteristics from a real handwritten image as conditional input of the decoder; the method has good advantages for maintaining high-resolution and high-detail images, and can improve the fidelity of local areas of the generated images; the generated synthetic image is used as a real handwritten text image to train an OCR model, so that the precision is greatly improved; after the generator network structure is trained, the obtained generator network model can convert the online handwriting data into a vivid offline handwriting image for assisting OCR training and being more beneficial to improving the recognition accuracy of the OCR model; the handwriting image generated by the framework can effectively improve the OCR recognition accuracy, and a feasible alternative scheme is provided for collecting and constructing a large-scale handwriting data set.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a block diagram of a stylistic GAN network of the present invention;
FIG. 3 is a diagram of the training process of the GAN network of the present invention;
fig. 4 is a structural diagram of a generator in the present invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and examples.
1-4, the present invention provides a method for improving handwritten OCR performance using synthesized online text images, comprising the steps of:
s1, selecting and dividing a data set, and adopting an IAM data set which comprises a plurality of types of data sets, wherein the data sets comprise an IAM handwriting data set and an IAM online handwriting data set; in the IAM handwritten data set, there are 6,161 offline handwritten text lines in the training set, the height is normalized to H =96, the width is set to W =3H, 1840 text lines in the verification set, 1861 lines in the test set, and 13049 text lines in the IAM online handwritten data set. Wherein the real imageI gt Obtaining a real handwritten image I after data processing sty, Storing in an IAM handwriting dataset;
s2, constructing a generator of the style GAN network, wherein the generator comprises three parts, namely a content encoder, a content decoder and a style encoder; the content encoder comprises five convolutional layer modules and three fully-connected layers, wherein the first convolutional layer module conv1 comprises two convolutional layers, the convolutional core of the convolutional layer is (3 x 64), a ReLU activator is arranged behind each convolutional layer, and a pooling layer is arranged behind two adjacent convolutional modules and used for down-sampling; the second convolutional layer module conv2 contains two convolutional layers (3 × 128), each followed by a ReLU activator; the third convolution module comprises four convolution layers (3 x 256), each convolution layer is followed by a ReLU activation function, and the activation function of the last convolution layer is followed by a pooling layer for down-sampling; the fourth convolution module contains four convolution layers (3 x 512), each convolution layer is followed by a ReLU activation function, and the activation function of the last convolution layer is followed by a pooling layer for down-sampling; the fifth convolution module comprises four convolution layers, the convolution kernel of each convolution layer is (3 × 512), a pooling layer is connected behind the last convolution layer, a full connection layer FC4096 is sequentially connected with a ReLU activation function, a full connection layer FC4096, a ReLU activation function, a full connection layer FC1000 and a softmax classifier, the result is output, the skeletonized skeleton image is used as input, an image space is converted into a feature space through a content encoder, the content features of the handwritten character image are extracted, a feature image is formed, and then the feature image is input into a content decoder;
for the style encoder, it is identical to the structure of the content encoder except that a real handwritten image I is received sty Outputting the style vector s as a 512-dimensional style vector s obtained through global pooling, and adjusting the feature map of the middle layer of the content decoder by the style vector s in an AdaIN operation mode, so that the style signs extracted by the style encoder can be combined with the content features of the style signs, and the style signs extracted by the style encoder can be combined with the content features of the style signs; adaIN is changing data at the feature map levelDistribution, the effect of style migration can be achieved by controlling the affine parameters in the changing AdaIN layer. Because the AdaIN layer changes the distribution of feature maps within the network, the task of format migration can be handed over to AdaIN, implementing other tasks on the network fabric, where AdaIN operates:
Figure 907889DEST_PATH_IMAGE007
wherein X represents the characteristic diagram of the content picture after coding, V σ ,V μ The method is obtained by affine transformation of a style vector s, parameters of affine transformation are different for different layers of a content decoder, and mean values and standard deviations are represented by sigma and mu;
for a content decoder, the structure of the content decoder is similar to that of a decoder part in U-Net, the structure of the content decoder is symmetrical to that of a content encoder, and a composite handwritten text picture with the same input size is generated by utilizing a multilayer convolution and a bilinear upsampling layer; the first convolutional layer module conv1 comprises four convolutional layers, the convolutional kernel of which is (3 × 512), each convolutional layer is followed by a ReLU activator, and two adjacent convolutional modules have an upsampled layer to amplify the image; the second convolutional layer module conv2 contains four convolutional layers (3 × 512), each of which is followed by a ReLU activator; the third convolution module comprises four convolution layers (3 × 256), each convolution layer being followed by a ReLU activation function; the fourth convolution module has two convolution layers, the convolution kernel of which is (3 × 128); the fifth convolution block has two convolution layers (3 x 64). An activation function is arranged behind each convolution layer, an up-sampling layer is arranged between two adjacent convolution modules to amplify the size of an image, wherein the up-sampling layer adopts a bilinear method to perform up-sampling, a feature diagram of each size is output, a component obtained by a style encoder after affine transformation of a style vector s is received, the distribution of the feature diagram is changed on the feature diagram level by adopting AdaIN operation, and style features in the style encoder are blended into a content feature diagram;
step (ii) ofAnd S3, training a generator of the GAN network, and converting the skeleton diagram into a handwritten image which has the style of a real diagram and has the same and vivid content through the generator. For training to generate
Figure 327369DEST_PATH_IMAGE008
The device performs skeletonization operation on a real image to obtain a skeleton drawing I of the real image ske Then drawing the skeleton I ske Inputting into content encoder for content feature extraction, and extracting real image I gt The input to the genre encoder extracts the genre features therein, and the output of the content encoder is then fed to the content decoder. For the skeleton diagram I with the height of H and the width of W ske The input of the style encoder is from the real hand-written image I gt Outputting a 512-dimensional style vector s obtained through global pooling; the content encoder generates an image I of height H and width W using multiple 3 × 3 convolution and bilinear upsampling layers syn The style vector s output by the style encoder adjusts the feature map of the middle layer of the content decoder in the mode of AdaIN so as to blend the style into the final output I of the content decoder syn
Wherein the skeleton diagram I ske With said true hand-written image I sty Form pairs of training data I ske -I sty And the sample pair is used for extracting a verification set and a test set from the IAM handwriting data set, and detecting the effect of the synthetic picture of the generator network model so as to train the generator.
Three loss functions are adopted in the training experiment of the invention, including a content loss function, a perception loss function and an antagonistic loss function; the content loss function accepts the difference from the content feature pixel level between the synthesized text image and the real image to optimize the best parameters, and the generator generates the synthesized image as I syn The real image is I syn
Figure 944295DEST_PATH_IMAGE009
The perceptual loss function solves the problem that the content loss function causes the generated image to be excessively smooth. Feature space after ReLU activation function after the first convolution layer of five convolution modules using pre-trained VGG-19 network on ImageNet:
Figure 500042DEST_PATH_IMAGE010
calculating a composite map I syn And a real image I gt The characteristic difference is as follows:
Figure 177011DEST_PATH_IMAGE011
MSE represents the mean square error function, and alpha takes values of 1/32,1/16,1/18,1/4 and 1 respectively.
The method of PatchGAN is adopted for the resistance loss function, the original GAN discriminator network is replaced by a full convolution network, a general GAN network only needs to output a true or false vector, and the PatchGAN discriminator network discriminates and outputs a small area of an input image receptive field, so that the model can pay more attention to image details through training; because the convolution layer that superposes one by one finally outputs a matrix of N × N, each element therein actually represents a relatively large receptive field in the original image, and corresponds to the patch in the original image, so that the method has a good advantage for maintaining the high-resolution and high-detail image, and can improve the fidelity of the generated image local area, specifically as follows:
Figure 83787DEST_PATH_IMAGE012
d represents a discriminator network for judging the probability of the comparison between the two being true or not;
and S4, synthesizing the text image in the online data set through the trained generator. And converting the IAM online handwriting data into a skeleton diagram, selecting a test set from the IAM handwriting data set as a style diagram, inputting the style diagram into a generator together, and generating an offline handwriting image, namely generating a very vivid offline handwriting image. Experiments show that the generated synthetic image is used as a real handwritten text image to train an OCR model, and compared with an image synthesized by adopting other algorithms to synthesize online data, the accuracy is greatly improved; and the OCR model is trained in a mode of training the generated image and the real image together, so that the recognition accuracy of the OCR model is improved.
While the invention has been described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (6)

1. A method for improving handwritten OCR performance using a composite online text image, comprising the steps of:
s1, selecting and dividing a data set: adopting an IAM data set, wherein the IAM data set comprises an IAM handwriting data set and an IAM online handwriting data set; real image I gt Obtaining a real hand-written image I after data processing sty And, storing in an IAM handwriting data set;
step S2, constructing a generator of the style GAN network, wherein the generator comprises a content encoder, a content decoder and a style encoder; the content encoder is composed of a plurality of pooling layers and convolution layers for sampling input down by taking advantage of the network structure of VGG-19; the content encoder comprises five convolution layer modules and three full-connection layers respectively;
step S3, training a generator of the GAN network: for real hand-written image I sty Performing skeletonization operation to obtain a skeleton diagram I ske Then drawing the skeleton I ske The content features are extracted by inputting the content into the content encoder, a feature map is generated and output to a content decoder; will be a true hand-written image I sty Input to styleThe encoder extracts style features in the video data, and outputs a 512-dimensional style vector s obtained through global pooling; the content decoder receives the characteristic diagram from the content encoder, receives the component obtained by affine transformation of the style vector s from the style encoder, changes the distribution of the characteristic diagram on the level of the characteristic diagram by adopting the operation of AdaIN, blends style information into the characteristic diagram, and outputs a composite diagram I syn
S4, synthesizing the text images in the online data set through the trained generator: and converting the IAM online handwritten data into a skeleton diagram, selecting a test set from the IAM handwritten data set as a style diagram, inputting the style diagram and the test set into a generator together, and generating an offline handwritten image.
2. The method of claim 1, wherein in step S3, the skeleton map I is used to improve OCR performance ske With said true handwritten image I sty Form paired training data I ske -I sty And the sample pair is used for extracting a verification set and a test set from the IAM handwriting data set, and detecting the effect of the synthetic picture of the generator network model so as to train the generator.
3. The method of claim 1, wherein in step S3, adaIN changes data distribution at feature map level, and the effect of style migration can be achieved by controlling affine parameters in AdaIN layer; the AdaIN layer changes the distribution of the characteristic diagram inside the network, hands over the task of the format migration to AdaIN, and realizes other tasks on the network structure, wherein AdaIN is operated as follows:
Figure FDA0003767725770000021
the characteristic diagram of the content picture after being coded, which is expressed by X, is obtained by affine transformation of a style vector s, and parameters of the affine transformation are different for different layers of a content decoder; σ and μ represent mean and standard deviation.
4. The method for improving handwriting OCR performance by using synthesized online text image according to claim 1, wherein in step S2, said content decoder has a structure similar to the decoder portion in U-Net and a structure symmetrical to the structure of the content encoder, and generates a synthesized handwritten text picture with the same input size by using multi-layer convolution and bilinear upsampling layer; the content decoder includes five convolutional layer modules.
5. A method for improving handwritten OCR performance using synthesized online text images as recited in claim 1, wherein in step S3, for H height and W width
Figure FDA0003767725770000022
Skeleton drawing I ske The input of the style encoder is from a real hand-written image I sty Outputting a 512-dimensional style vector s obtained through global pooling; the content encoder generates an image I of height H and width W using multiple layers of 3 x 3 convolution and bilinear upsampling layers syn The style vector s output by the style encoder adjusts the feature map of the middle layer of the content decoder in an AdaIN mode so as to blend the style of the feature map into the final output I of the content decoder syn
6. A method for improving OCR performance by using synthesized online text image according to claim 1, wherein in step S3, by adopting three loss functions including content loss function, perceptual loss function and anti-loss function, the synthesized picture generated by the training generator is as realistic as possible and has a style corresponding to the real picture;
the content loss function accepts the difference from the content feature pixel level between the synthesized text image and the real image to optimize the best parameters, and the generator generates I the synthesized image syn The real image is I gt
L content =||I gt -I syn || 1
The perception loss function solves the problem that the generated image is excessively smooth due to the content loss function, and utilizes a pre-trained VGG-19 network on ImageNet to obtain a feature space after a relu activation function after a first convolution layer of five convolution modules:
φ i (·),i={1,2,3,4,5},
calculating a composite map I syn And a real graph I gt The characteristic difference is as follows:
Figure FDA0003767725770000041
MSE represents a mean square error function, and the values of alpha are respectively 1/32,1/16,1/18,1/4 and 1; the countermeasure loss function adopts the method of patchGAN, and the arbiter network of former GAN network has been changed into the full convolution network, and patchGAN's arbiter network differentiates a small region of input image receptive field and outputs, makes the model more can focus on image details, and the convolution that superposes one by one finally outputs N × N's matrix, and each element wherein actually represents a bigger receptive field in the original image, corresponds to the patch in the original image, specifically as follows:
Figure FDA0003767725770000042
d represents a network of discriminators for determining the probability of the comparison between the two being true or false.
CN202011429519.3A 2020-12-09 2020-12-09 Method for improving handwriting OCR performance by utilizing synthesized online text image Active CN112364838B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011429519.3A CN112364838B (en) 2020-12-09 2020-12-09 Method for improving handwriting OCR performance by utilizing synthesized online text image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011429519.3A CN112364838B (en) 2020-12-09 2020-12-09 Method for improving handwriting OCR performance by utilizing synthesized online text image

Publications (2)

Publication Number Publication Date
CN112364838A CN112364838A (en) 2021-02-12
CN112364838B true CN112364838B (en) 2023-04-07

Family

ID=74536790

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011429519.3A Active CN112364838B (en) 2020-12-09 2020-12-09 Method for improving handwriting OCR performance by utilizing synthesized online text image

Country Status (1)

Country Link
CN (1) CN112364838B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113190161A (en) * 2021-04-25 2021-07-30 无锡乐骐科技有限公司 Electronic writing exercise method based on convolutional neural network
CN113554549B (en) * 2021-07-27 2024-03-29 深圳思谋信息科技有限公司 Text image generation method, device, computer equipment and storage medium
CN114973279B (en) * 2022-06-17 2023-02-17 北京百度网讯科技有限公司 Training method and device for handwritten text image generation model and storage medium

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020140421A1 (en) * 2019-01-03 2020-07-09 Boe Technology Group Co., Ltd. Computer-implemented method of training convolutional neural network, convolutional neural network, computer-implemented method using convolutional neural network, apparatus for training convolutional neural network, and computer-program product

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11423282B2 (en) * 2018-10-30 2022-08-23 Huawei Technologies Co., Ltd. Autoencoder-based generative adversarial networks for text generation
CN110443864B (en) * 2019-07-24 2021-03-02 北京大学 Automatic artistic font generation method based on single-stage small-amount sample learning
CN111753493A (en) * 2019-09-29 2020-10-09 西交利物浦大学 Style character generation method containing multiple normalization processes based on small amount of samples
CN110992252B (en) * 2019-11-29 2023-05-12 北京航空航天大学合肥创新研究院 Image multi-grid conversion method based on latent variable feature generation
CN111402179B (en) * 2020-03-12 2022-08-09 南昌航空大学 Image synthesis method and system combining countermeasure autoencoder and generation countermeasure network

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020140421A1 (en) * 2019-01-03 2020-07-09 Boe Technology Group Co., Ltd. Computer-implemented method of training convolutional neural network, convolutional neural network, computer-implemented method using convolutional neural network, apparatus for training convolutional neural network, and computer-program product

Also Published As

Publication number Publication date
CN112364838A (en) 2021-02-12

Similar Documents

Publication Publication Date Title
CN112364838B (en) Method for improving handwriting OCR performance by utilizing synthesized online text image
CN109635883B (en) Chinese character library generation method based on structural information guidance of deep stack network
Jiang et al. Scfont: Structure-guided chinese font generation via deep stacked networks
CN108399419B (en) Method for recognizing Chinese text in natural scene image based on two-dimensional recursive network
CN108537733B (en) Super-resolution reconstruction method based on multi-path deep convolutional neural network
CN111179167B (en) Image super-resolution method based on multi-stage attention enhancement network
Yin et al. FD-SSD: An improved SSD object detection algorithm based on feature fusion and dilated convolution
CN110136062B (en) Super-resolution reconstruction method combining semantic segmentation
Peyrard et al. ICDAR2015 competition on text image super-resolution
CN113096017A (en) Image super-resolution reconstruction method based on depth coordinate attention network model
CN110853039B (en) Sketch image segmentation method, system and device for multi-data fusion and storage medium
CN112101262B (en) Multi-feature fusion sign language recognition method and network model
CN114898284B (en) Crowd counting method based on feature pyramid local difference attention mechanism
CN116309913B (en) Method for generating image based on ASG-GAN text description of generation countermeasure network
CN113076947A (en) RGB-T image significance detection system with cross-guide fusion
CN114693929A (en) Semantic segmentation method for RGB-D bimodal feature fusion
CN111260577B (en) Face image restoration system based on multi-guide image and self-adaptive feature fusion
CN113378949A (en) Dual-generation confrontation learning method based on capsule network and mixed attention
CN116935043A (en) Typical object remote sensing image generation method based on multitasking countermeasure network
CN109766918A (en) Conspicuousness object detecting method based on the fusion of multi-level contextual information
CN112288626A (en) Face illusion method and system based on dual-path depth fusion
CN116152374A (en) Chinese character font generating method
Li et al. Line drawing guided progressive inpainting of mural damages
Tang et al. SRARNet: A unified framework for joint superresolution and aircraft recognition
CN114626984A (en) Super-resolution reconstruction method for Chinese text image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant