CN105893968A - Text-independent end-to-end handwriting recognition method based on deep learning - Google Patents

Text-independent end-to-end handwriting recognition method based on deep learning Download PDF

Info

Publication number
CN105893968A
CN105893968A CN201610202734.7A CN201610202734A CN105893968A CN 105893968 A CN105893968 A CN 105893968A CN 201610202734 A CN201610202734 A CN 201610202734A CN 105893968 A CN105893968 A CN 105893968A
Authority
CN
China
Prior art keywords
character
text
pseudo
described step
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610202734.7A
Other languages
Chinese (zh)
Other versions
CN105893968B (en
Inventor
金连文
杨维信
刘曼飞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing Sign Digital Technology Co ltd
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201610202734.7A priority Critical patent/CN105893968B/en
Publication of CN105893968A publication Critical patent/CN105893968A/en
Application granted granted Critical
Publication of CN105893968B publication Critical patent/CN105893968B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/32Digital ink
    • G06V30/333Preprocessing; Feature extraction
    • G06V30/347Sampling; Contour coding; Stroke extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Multimedia (AREA)
  • Character Discrimination (AREA)

Abstract

The invention provides a text-independent end-to-end handwriting recognition method based on deep learning, comprising the following steps: A, preprocessing an online handwritten text to generate a pseudo character sample; B, calculating the path integral feature image of the pseudo character sample; C, training a deep neural network model of a sample of a known writer; and D, using the deep neural network model in step C to automatically recognize a sample of an unknown writer. Through the method, online text lines can be processed automatically, there is no need to extract character features manually, and text-independent online writer recognition is realized efficiently.

Description

The person's handwriting recognition methods end to end that text based on degree of depth study is unrelated
Technical field
The invention belongs to degree of depth study and artificial intelligence field, particularly relate to computer user's clipboard computer The unrelated hand script Chinese input equipment document of text carries out feature learning to distinguish the technology of writer.
Background technology
Person's handwriting is an important evidence of authentication, law evidence obtaining etc..In recent years, mobile terminal and clipboard are touched The universal attention rate improving hand script Chinese input equipment writer verification.But due to data deficiency, do not find that handwriting samples is special Good feature, the accuracy that writer identifies need to improve.Conventional recognition methods is all based on pattern recognition etc. and is correlated with Computer image processing technology, artificially determine some features statement and distance calculate, artificially extract character feature, identify Accuracy rate the highest.
Summary of the invention
In order to solve the technical problem existing for prior art, the invention provides the end that text based on degree of depth study is unrelated To the person's handwriting recognition methods of end, the method can automatically process online line of text, it is not necessary to artificially extracts character feature, efficiently Achieve the unrelated online writer of text and identify.
The present invention adopts the following technical scheme that and realizes: the person's handwriting identification side end to end that text based on degree of depth study is unrelated Method, it is characterised in that comprise the steps: A, hand script Chinese input equipment text carries out pretreatment, generates pseudo-character sample; B, the path integral characteristic image of the pseudo-character sample of calculating;C, train the deep neural network mould of known writer's sample Type;D, utilize the deep neural network model of step C, the sample of uncertain writer is identified automatically.
Preferably, described step A, particularly as follows: A1, each stroke to hand script Chinese input equipment text carry out resampling, is adopted Sampling point density uniform hand script Chinese input equipment text chunk;
A2, the hand script Chinese input equipment text after resampling is carried out stroke segmentation, obtain the path collection of the most broken little stroke section composition Close;
A3, stroke is split after text carry out Character segmentation, generate pseudo-character;
A4, the pseudo-character after each segmentation is carried out stroke section remove at random;
A5, the size of normalization puppet character;
A6, affine transformation, generate pseudo-character sample.
Preferably, in described step A1, resampling is to calculate new according to set tracing point density parameter and original stroke number The quantity of the tracing point of sample;
In described step A2, stroke segmentation is first to carry out Corner Detection, then this stroke is cut off from corner point, generates new Comparatively short stroke section;
In described step A3, Character segmentation is to take out stroke section the most successively, when these stroke sections of being extracted into are combined into The when that the width of character being just above character average height, currently last stroke section is taken as a fresh character Start;
In described step A4, the quantity removing the pseudo-character obtained at random of each character stroke section is total stroke hop count si's Function;
In described step A5, normalization is to be zoomed in and out by the figure that each pseudo-character is mapped to two dimensional surface, relatively wide and High length, longest edge is transformed to fixing needed for length value, and keep the ratio of width to height constant in the case of, to short Multiple is scaled accordingly while be multiplied by;
In described step A6, affine transformation includes the rotation to whole character path, stretches and tilt.
Preferably, described step B is particularly as follows: B1, calculate one group of path integral feature to each pseudo-character sample;
B2, reassemble into different path integral characteristic patterns by often organizing path integral feature according to the feature of identical dimensional;
B3, path integral characteristic pattern is carried out margin pixel fill up.
Preferably, when described step B1 calculates path integral feature, it is assumed that finite length stroke section P is two-dimensional spacePath, the time that track moves meet0 < τ1< ... < τk< T, τiIn the middle of representing the I time point, and positive integer i meets 1≤i≤k, then calculates the k rank path integral at time [0, T] interior P special LevyWhen P is straight line, use Δ0,TRepresent path displacement,Can be calculated by segmentation and try to achieve;Calculate n rank road Footpath integration feature, it is simply that path integral feature is done k rank and blocks, the feature set obtained is Obtain 2n+1The path integral feature of dimension;
In described step B2, each dimension of path integral feature is individually become a path integral characteristic pattern, so Each pseudo-character sample has 2n+1Open path integral characteristic pattern, including the two-dimensional image in path itself;
In described step B3, first path integral characteristic pattern is set as the size that pixel value is z × z, is placed on a pixel value For the center of the figure of Z × Z, then these path integral characteristic patterns are input to the deep neural network model described in step C In be trained, z < Z≤3z.
Preferably, described step C particularly as follows:
C1, the projected depth neutral net number of plies, the template number of convolutional layer and the neuron number of full articulamentum;
C2, whole sample extraction characteristic images of training set are trained as the input of deep neural network model;
C3, when network converge in training set accuracy rate no longer rise time, deconditioning, preserve deep neural network mould Shape parameter.
Preferably, in described step C1, deep neural network includes five convolutional layers, has one after each convolutional layer Great Chiization layer;
In described step C2, the training of deep neural network includes the successive ignition of two steps of forward and backward, first with front After network obtains network error, using back-propagation algorithm to be updated network parameter, continuous iteration optimization network is joined Number.
Preferably, described step D is particularly as follows: D1, that the path integral characteristic image through pretreatment is inputted the degree of depth is neural Network model, is calculated and each removes the candidate item probability tables that the pseudo-character sample after stroke is corresponding at random;
D2, the candidate item probability tables that the multiple pseudo-character samples of each character are corresponding is added and asks probability average, obtain this puppet The candidate item probability tables of character, is added the candidate item probability tables of all characters of text and asks probability averagely to obtain the text Candidate item probability tables;
D3, the candidate item probability tables of foundation text are selected the candidate item of highest scoring and are judged to writer.
Preferably, in described step D1, each text dividing is multiple character, and each pseudo-character produces multiple pseudo-character Sample, these pseudo-character samples have identical label, calculate candidate item probability tables respectively as an independent sample;
In described step D2, probability averagely includes that the many character probabilities of text are average and multiple puppet character sample probability of character Averagely;The candidate item that each pseudo-character sample candidate item probability tables each dimension addition of one character is obtained this character is put down All probability tableses, obtain the text by being added the candidate item each dimension of average probability table of each character of a text chunk Candidate item probability tables;
In described step D3, the candidate item selecting highest scoring according to the character average candidate item probability tables of text is judged to this The writer of text.
From above technical scheme, the person's handwriting recognition methods end to end that text that the present invention learns based on the degree of depth is unrelated, Mainly include the preprocessing process of hand script Chinese input equipment text, deep neural network model training process and automatically identify process. Stroke section dividing method that wherein pretreatment is used, remove stroke phase method and first path integral feature is used for book at random Writer's identification is the innovation emphasis of the present invention.Compared with prior art, the invention have the advantages that and beneficial effect:
1, the method for pretreatment includes the generalization ability enhancing that text dividing, sample augmentation and text are unrelated;Pretreatment operation Make the present invention be applicable to the text of various length, can be long text or short text, it might even be possible to be individual character.
2, remove stroke section and generate abundant training sample, prevent over-fitting when of being used for training deep neural network, also Generate multiple pseudo-character during for testing to be identified improving discrimination.
3, the present invention proposes for the first time a path integral feature for writer's identification mission, is also that the first time degree of depth is rolled up Long-pending neural fusion writer identifies.Path integral feature can be extracted and can be used for the validity feature that writer identifies, Being learnt by the supplemental characteristic of deep neural network, discrimination is up to 95.72% (Chinese), 98.51% (English).Based on Deep neural network, it is possible to be identified the handwriting samples of the different length of writer, has higher accuracy and Shandong Rod.
Accompanying drawing explanation
Fig. 1 is the flow chart of the present invention;
Fig. 2 is the schematic diagram of random drop stroke section in the present invention, have chosen 30 random drop stroke sections of a character After pseudo-character as example;
Fig. 3 represents the visualized graphs of path integral feature in the present invention.
Detailed description of the invention
Below in conjunction with embodiment and accompanying drawing, the present invention is described in further detail, but the detailed description of the invention of the present invention is not It is confined to this.
Embodiment
Present invention mainly solves the identification of online text written person and implement, have employed and online text is carried out cutting The preprocess method removed at random with stroke section, establishes unrelated end-to-end of complete text based on degree of depth study Person's handwriting recognition methods.The character types that user is inputted by the present invention do not limit, and the most not limit text, it is possible to Allowing user to carry out free text written in big degree, overall flow is as shown in Figure 1.
Seeing Fig. 1, the present invention includes following four process: A, the preprocessing process of hand script Chinese input equipment text;B, known write The deep neural network model training process of person's sample;C, calculating path integral feature;D, the sample of uncertain writer Automatically process is identified.Specifically, first have to the line of text of hand script Chinese input equipment long text is carried out resampling, become sampled point The online line of text that spacing is equal, then the line of text after resampling is divided into smaller stroke section set, by these pens Draw section and be divided into single character based on the ratio of width to height.Then the stroke section of each character is removed at random, generate multiple pseudo-word Symbol.Calculate afterwards after each pseudo-character carries out affine transformation and generate one group of path integral characteristic pattern, and fill null value around Point.It is input in deep neural network carry out degree of depth network model by the path integral characteristic pattern of the pseudo-character sample of training set Training, to close to saturated, preserves degree of depth network training parameter.Test time, training set hand script Chinese input equipment long text is carried out on The degree of depth network model that data prediction described in literary composition being input to preserves calculates, exports each pseudo-character sample Candidate probability table, then calculates the probability tables of each character.Afterwards the probability tables correspondence from same text fragment is waited Option is sued for peace, and obtains final probability tables, and selects, according to this probability tables, the candidate item that probit is maximum, it is determined that for Writer.The labeling requirement of the test item of native system occurred in training set.
Individually below each key step of the present invention is described in detail:
Step A data prediction
The purpose of step A data prediction is that the hand script Chinese input equipment line of text data to user's input are split, and is formed permissible The form utilized, and extract some features, help deep neural network preferably to learn and processing feature, in efficiency and knowledge Good auxiliaring effect is had in other accuracy.The sample method resampling by linear interpolation, by local buckling degree meter Calculate detection angle point.Stroke section after segmentation is combined into character, then the stroke section inside each character is moved at random Removing, obtain substantial amounts of pseudo-character, these pseudo-characters obtain more diversified pseudo-word through size normalization and affine transformation Symbol sample.
A1, sample resampling
Resampling is the quantity of the tracing point calculating new samples according to set tracing point density parameter and original stroke number;Root Calculate the total length of a stroke according to original tracing point, divided by the quantity of the tracing point of new samples, obtain dot density, And then determine whether former tracing point retains and need the number of interpolation two-by-two on line so that it is determined that the tracing point of new samples is sat Mark.
If a stroke has the sampled point { (x of p constant duration1,x2),...,(xp,kp)}.Due to the difference of writing speed, The Euclidean distance of these points is the most different.When integer i meets 1≤i≤p, it is assumed that (xi,yi) and (xi+1,yi+1) point-to-point transmission Euclidean distance be di, (x0,y0) arrive (xi,yi) inter-two-point path is a length ofIf the puppet after interpolation to be obtained Specimen sample point sum is l, and l is the integral multiple of p.After interpolation, first point of each stroke keeps constant, from second Individual point starts, and the position coordinates of i-th point is:
(xi×α+xi+1×(1-α),yi×α+yi+1×(1-α)), (1)
Wherein
α = ( D j + 1 - i l × D j ) D j + 1 - D j , - - - ( 2 )
The set of the point after each resampling still falls within this stroke.
A2, stroke section are split
Stroke segmentation is first to carry out Corner Detection, then this stroke is cut off from corner point, generates new comparatively short stroke Section;Judging that a point is angle point, need to be calculated its flexibility by the coordinate of the point before and after this point, local buckling degree is Big point is considered as angle point;Assume (xi,yi) be the i-th trajectory coordinates point after interpolation, respectively with before this point and after Kth point (the x in facei-k,yi-k) and (xi+k,yi+k) coordinate figure calculate flexibility.
The segmentation of stroke section is first to carry out stroke end points identification according to storage data.Run through end point mark or one online The first coordinate of hand-written long text file, is just defaulted as the starting point of a stroke, i.e. end points.Corner Detection judges angle The principle of point is that local buckling degree is maximum.Its flexibility is calculated: assume (x by the coordinate of point before and after each pointi,yi) it is interpolation After trajectory coordinates point, the kth point of this front and back is (xi-k,yi-k) and (xi+k,yi+k), flexibility is defined as:
β=max (| xi+k+xi-k-2xi|,|yi+k+yi-k-2yi|)/2k, (3)
Then this stroke is cut off from corner point, generate new comparatively short stroke section;For training data, if each character It is highly ymax-ymin, then estimate average height y of each character of a documentaverFor Character segmentation.For Each character of test data traversal text, it is thus achieved that the maximum of local vertical coordinate, minima are ymaxAnd ymin, and then estimate Go out character height y that the text is averageaver
A3, Character segmentation, pseudo-character generates
Character segmentation is the character fixing in order to obtain length-width ratio.Character segmentation is to take out stroke section the most successively, and remembers Record the maximum x of its abscissa occurredmaxWith minima xmin.Width when the character that these stroke sections being extracted into are combined into Just above character average height yaverWhen, last stroke section is taken as the beginning of a fresh character;Character Average height be calculated cutting stroke section when.
A4, stroke section remove at random
Assume that a total stroke number of character is m, the stroke hop count s of i-th strokei, stroke section remove the word obtained at random The sample size of symbol is siFunction with m.If each character is removed d at randomi(0≤di< si) pen, these are remaining Stroke section is reassembled into pseudo-character sample according to original sequencing, then the pseudo-character sample sum obtained is exactly:
N ( m , S ) = ( Π i = 1 m Σ d i = 0 s i C s i d i ) - C s ^ s ^ = 2 s ^ - 1 , - - - ( 4 )
Fig. 2 is shown in by schematic diagram.
A5, size normalization
Normalization is to be zoomed in and out by the figure that each pseudo-character is mapped to two dimensional surface, and relatively wide and high length will be Long limit transforms to fixing required length value, and in the case of keeping the ratio of width to height constant, minor face is multiplied by corresponding contracting Put multiple.
Size normalization is the coordinate (x of the first path point of one character of traversali,yi), find out width w=xmax-xminAnd height H=ymax-ymin, then by long limit max, (w, h) is stretched to fixed value Q, and minor face expands corresponding multiple, obtains normalizing Path point coordinate after change
( x i × Q m a x ( w , h ) , y i × Q m a x ( w , h ) ) , - - - ( 5 )
A6, affine transformation
Affine transformation include the rotation to whole character path, stretch, inclination etc.;The angle rotated depends on certain interval The twiddle factor w of interior random size, the coordinate of postrotational point is:
(xi×cos(w)+yi×sin(w),-xi×sin(w)+yi×cos(x)), (6)
Stretching is the abscissa to path coordinate points or vertical coordinate carries out linear transformation, and drawing coefficient is set to α and β, ((α, β) ∈ [-1,1]), coordinate (xi,yi) coordinate after stretching conversion is:
(xi×(1+α),yi×(1+β)), (7)
Tilt variation includes the tilt variation to horizontal direction and vertical direction.Coordinate (xi,yi) inclination in the horizontal direction Coordinate after change is:
(xi×(1+αx),yi), (8)
Coordinate (xi,yi) in the vertical direction tilt variation after coordinate:
(xi,yi×(1+αy)), (9)
Step B calculates path integral characteristic pattern
B1, calculating path integral feature
Calculating path integral feature is by the method for path integral feature.Assume that finite length stroke section P is that two dimension is empty BetweenPath, middle i-th time point is τi, positive integer i meets 1≤i≤k, the time that track movesAnd 0 < τ1< ... < τk< T, then the k rank path integral feature of P is exactly:
P 0 , T k = ∫ 0 T ∫ 0 τ k ... ∫ 0 τ 2 dP τ 1 ⊗ ... ⊗ dP τ k , - - - ( 10 )
When P is straight line, use Δ0,TRepresenting path displacement, segmentation calculates:
P 0 , T k = { ( P 0 , T k - 1 ⊗ Δ 0 , T ) / k k ≥ 1 1 k = 0 , - - - ( 11 )
Calculating n rank path integral feature, the feature set obtained is expressed as
F 0 , T n = ( P 0 , T 0 , P 0 , T 1 , P 0 , T 2 , ... , P 0 , T n ) , - - - ( 12 )
The dimension of the path integral feature including path itself obtained is 2n+1
B2, generation path integral characteristic pattern
Generating the multidimensional path integral feature of image in step bl is determined., each dimension can corresponding one-tenth one width path integral Characteristic pattern.Removing path image itself, the quantity of the characteristic pattern that each pseudo-character sample obtains is 2n+1-1.Generate path Integration schematic diagram is as shown in Figure 3.
B3, margin pixel are filled up
In order to keep image not lose the rim path integration feature of image because of convolution operation, input layer image is carried out Blank pixel is filled up.The detailed description of the invention filled up is first path integral characteristic pattern to be converted into the big of 54 × 54 pixels Little, then surrounding fills the blank pixel point of 21 layers of pixel, becomes the figure of 96 × 96.
Deep neural network is trained by step C
C1, projected depth neural network model
In the present invention, the deep neural network of setting comprises convolutional layer and maximum pond layer;Its structure is five convolutional layers, Maximum pond layer (MP) is had after each convolutional layer;The size of ground floor convolution kernel is 3 × 3 (being expressed as C3), after The size of the four layers of convolution kernel in face is 2 × 2 (being expressed as C2);Step-length is 2;Finally there are two full articulamentums, are respectively 480 and 512 neurons.Whole network structure is collectively expressed as:
M×96×96Input-80C3-MP2-160C2-MP2-240C2-MP2-320C2-MP2-400C2-MP2-480FC-512FC- Output,
Wherein M represents the port number of input layer, equal with the quantity of the integration characteristic pattern of each pseudo-character sample.
C2, training deep neural network
It is used for training deep neural network by the data of training set.Classification problem is done the when of training.Deep neural network Training includes the successive ignition of two steps of forward and backward.After first obtaining network error with feedforward network, use reversely biography Broadcast algorithm network parameter is updated, continuous iteration optimization network parameter, test for training data after every suboptimization Classification accuracy.
C3, preservation deep neural network model parameter
The accuracy rate of training data presents the trend that concussion rises.When the accuracy rate of training data almost no longer rises, recognize For training close to saturated, then preservation model Parameter File, it is used for testing.
Step D identifies writer automatically
D1, candidate probability calculate
Can generate tens to thousand of pseudo-character samples for each long text, the writer of text is exactly each The label of pseudo-character sample.Each pseudo-character sample can generate 2n+1Path integral characteristic pattern, 2n+1It it is input layer simultaneously Port number.Input layer image is input in the deep neural network model that step C3 preserves carry out forward calculation, obtains deep The output of degree neutral net;I-th that in long text, i-th character generatesjThe probability of individual pseudo-character sample is:
( p i j 1 , p i j 2 , ... , p i j η ) , - - - ( 13 )
D2, probability are average
Probability averagely includes that average and character the multiple pseudo-character sample probability of the many character probabilities of text is average.Assume total η class Hands writer's long text, each long text can generate r character, and each character can generate N number of pseudo-character sample, often Individual pseudo-character sample can obtain the character puppet sample probability average out to of a candidate item probability column i-th character:
( p i 1 , p i 2 , ... , p i η ) = Σ j = 1 N ( p i j 1 , p i j 2 , ... , p i j η ) , - - - ( 14 )
The character candidates item average probability table of each text chunk is:
( p 1 , p 2 , ... , p η ) = Σ i = 1 r ( p i 1 , p i 2 , ... , p i η ) , - - - ( 15 )
Finding the most probable value in formula (14) is λ item
p λ = m a x ( Σ i = 1 r ( p 1 , p 2 , ... , p η ) ) , - - - ( 16 )
D3, writer judge
The candidate item selecting highest scoring according to the candidate item probability tables of pseudo-character is judged to writer;Being understood by step D2 should The classification results of long text is the λ class of candidate item.
Embodiments of the present invention are also not restricted to the described embodiments, other any spirit without departing from the present invention with The change made under principle, modify, substitute, combine, simplify, all should be the substitute mode of equivalence, be included in this Within bright protection domain.

Claims (10)

1. the person's handwriting recognition methods end to end that text based on degree of depth study is unrelated, it is characterised in that comprise the steps:
A, hand script Chinese input equipment text is carried out pretreatment, generate pseudo-character sample;
B, the path integral characteristic image of the pseudo-character sample of calculating;
C, train the deep neural network model of known writer's sample;
D, utilize the deep neural network model of step C, the sample of uncertain writer is identified automatically.
Person's handwriting recognition methods the most according to claim 1, it is characterised in that described step A particularly as follows:
A1, each stroke to hand script Chinese input equipment text carry out resampling, obtain sampling point density uniform hand script Chinese input equipment text chunk;
A2, the hand script Chinese input equipment text after resampling is carried out stroke segmentation, obtain the set of paths of the most broken little stroke section composition;
A3, stroke is split after text carry out Character segmentation, generate pseudo-character;
A4, the pseudo-character after each segmentation is carried out stroke section remove at random;
A5, the size of normalization puppet character;
A6, affine transformation, generate pseudo-character sample.
Person's handwriting recognition methods the most according to claim 2, it is characterised in that
In described step A1, resampling is the track calculating new samples according to set tracing point density parameter and original stroke number The quantity of point;
In described step A2, stroke segmentation is first to carry out Corner Detection, then this stroke is cut off from corner point, generates new Comparatively short stroke section;
In described step A3, Character segmentation is to take out stroke section the most successively, when the word that these stroke sections being extracted into are combined into The when that the width of symbol being just above character average height, currently last stroke section is taken as the beginning of a fresh character;
In described step A4, the quantity removing the pseudo-character obtained at random of each character stroke section is total stroke hop count siLetter Number;
In described step A5, normalization is to be zoomed in and out by the figure that each pseudo-character is mapped to two dimensional surface, relatively wide and high Length, longest edge is transformed to fixing needed for length value, and keep the ratio of width to height constant in the case of, minor face is multiplied by Scale multiple accordingly;
In described step A6, affine transformation includes the rotation to whole character path, stretches and tilt.
Person's handwriting recognition methods the most according to claim 1, it is characterised in that described step B particularly as follows:
B1, each pseudo-character sample is calculated one group of path integral feature;
B2, reassemble into different path integral characteristic patterns by often organizing path integral feature according to the feature of identical dimensional;
B3, path integral characteristic pattern is carried out margin pixel fill up.
Person's handwriting recognition methods the most according to claim 4, it is characterised in that:
When described step B1 calculates path integral feature, it is assumed that finite length stroke section P is two-dimensional spacePath, The time that track moves meets0 < τ1< ... < τk< T, τiI-th time point in the middle of representing, And positive integer i meets 1≤i≤k, then calculate the k rank path integral feature at time [0, T] interior PWhen P is straight During line, use Δ0,TRepresent path displacement,Can be calculated by segmentation and try to achieve;Calculate n rank path integral feature, it is simply that Path integral feature is done k rank and is blocked, and the feature set obtained isObtain 2n+1The path integral of dimension is special Levy;
In described step B2, each dimension of path integral feature is individually become a path integral characteristic pattern, so often One pseudo-character sample has 2n+1Open path integral characteristic pattern, including the two-dimensional image in path itself;
In described step B3, path integral characteristic pattern being first set as the size that pixel value is z × z, being placed on a pixel value is Then these path integral characteristic patterns are input to enter in the deep neural network model described in step C by the center of the figure of Z × Z Row training, z < Z≤3z.
Person's handwriting recognition methods the most according to claim 1, it is characterised in that described step C particularly as follows:
C1, the projected depth neutral net number of plies, the template number of convolutional layer and the neuron number of full articulamentum;
C2, whole sample extraction characteristic images of training set are trained as the input of deep neural network model;
C3, when network converge in training set accuracy rate no longer rise time, deconditioning, preserve deep neural network model ginseng Number.
Person's handwriting recognition methods the most according to claim 6, it is characterised in that:
In described step C1, deep neural network includes five convolutional layers, has a maximum pond layer after each convolutional layer;
In described step C2, the training of deep neural network includes the successive ignition of two steps of forward and backward, first uses forward direction After network obtains network error, use back-propagation algorithm that network parameter is updated, continuous iteration optimization network parameter.
Person's handwriting recognition methods the most according to claim 2, it is characterised in that described step D particularly as follows:
D1, the path integral characteristic image through pretreatment is inputted deep neural network model, be calculated and each remove at random The candidate item probability tables that pseudo-character sample after stroke is corresponding;
D2, the candidate item probability tables that the multiple pseudo-character samples of each character are corresponding is added and asks probability average, obtain this puppet character Candidate item probability tables, the candidate item probability tables of all characters of text is added that to ask probability averagely to obtain the candidate item of the text general Rate table;
D3, the candidate item probability tables of foundation text are selected the candidate item of highest scoring and are judged to writer.
Person's handwriting recognition methods the most according to claim 8, it is characterised in that
In described step D1, each text dividing is multiple character, and each pseudo-character produces multiple pseudo-character sample, these Pseudo-character sample has identical label, calculates candidate item probability tables respectively as an independent sample;
In described step D2, probability averagely includes that average and character the multiple pseudo-character sample probability of the many character probabilities of text is put down All;Each pseudo-character sample candidate item probability tables each dimension addition to a character obtains the candidate item average probability of this character Table, general by the candidate item each dimension of average probability table of each character of a text chunk being added the candidate item obtaining the text Rate table;
In described step D3, the candidate item selecting highest scoring according to the character average candidate item probability tables of text is judged to this article This writer.
Person's handwriting recognition methods the most according to claim 3, it is characterised in that during described step A2 Corner Detection, sentence A disconnected point is angle point, needs to be calculated its flexibility by the coordinate of the point before and after this point, and the point of local buckling degree maximum is recognized For being angle point;Assume (xi,yi) it is the trajectory coordinates point after interpolation, respectively with the kth point (x of this front and backi-k,yi-k) With (xi+k,yi+k) coordinate figure calculate flexibility.
CN201610202734.7A 2016-03-31 2016-03-31 The unrelated person's handwriting recognition methods end to end of text based on deep learning Active CN105893968B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610202734.7A CN105893968B (en) 2016-03-31 2016-03-31 The unrelated person's handwriting recognition methods end to end of text based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610202734.7A CN105893968B (en) 2016-03-31 2016-03-31 The unrelated person's handwriting recognition methods end to end of text based on deep learning

Publications (2)

Publication Number Publication Date
CN105893968A true CN105893968A (en) 2016-08-24
CN105893968B CN105893968B (en) 2019-06-14

Family

ID=57013309

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610202734.7A Active CN105893968B (en) 2016-03-31 2016-03-31 The unrelated person's handwriting recognition methods end to end of text based on deep learning

Country Status (1)

Country Link
CN (1) CN105893968B (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106408038A (en) * 2016-09-09 2017-02-15 华南理工大学 Rotary Chinese character identifying method based on convolution neural network model
CN106997473A (en) * 2016-09-08 2017-08-01 汪润春 A kind of image-recognizing method based on neutral net
CN107292280A (en) * 2017-07-04 2017-10-24 盛世贞观(北京)科技有限公司 A kind of seal automatic font identification method and identifying device
CN107657230A (en) * 2017-09-27 2018-02-02 安徽硕威智能科技有限公司 A kind of bank self-help robot character recognition device
CN108108746A (en) * 2017-09-13 2018-06-01 湖南理工学院 License plate character recognition method based on Caffe deep learning frames
CN108154136A (en) * 2018-01-15 2018-06-12 众安信息技术服务有限公司 For identifying the method, apparatus of writing and computer-readable medium
CN108596168A (en) * 2018-04-20 2018-09-28 北京京东金融科技控股有限公司 For identification in image character method, apparatus and medium
CN108665010A (en) * 2018-05-12 2018-10-16 新疆大学 A kind of hand script Chinese input equipment Uighur words data enhancement methods
CN109740605A (en) * 2018-12-07 2019-05-10 天津大学 A kind of handwritten Chinese text recognition method based on CNN
CN109815809A (en) * 2018-12-19 2019-05-28 天津大学 A kind of English handwriting identification method based on CNN
CN109858488A (en) * 2018-12-28 2019-06-07 众安信息技术服务有限公司 A kind of handwriting samples recognition methods and system based on sample enhancing
CN111723807A (en) * 2019-03-20 2020-09-29 Sap欧洲公司 Recognizing machine-typed and handwritten characters using end-to-end deep learning
CN111738167A (en) * 2020-06-24 2020-10-02 华南理工大学 Method for recognizing unconstrained handwritten text image
CN112989834A (en) * 2021-04-15 2021-06-18 杭州一知智能科技有限公司 Named entity identification method and system based on flat grid enhanced linear converter

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101730898A (en) * 2005-06-23 2010-06-09 微软公司 Adopt the handwriting recognition of neural network
CN104850837A (en) * 2015-05-18 2015-08-19 西南交通大学 Handwritten character recognition method
CN105320961A (en) * 2015-10-16 2016-02-10 重庆邮电大学 Handwriting numeral recognition method based on convolutional neural network and support vector machine
CN105354538A (en) * 2015-10-13 2016-02-24 广东小天才科技有限公司 Chinese character handwriting recognition method and system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101730898A (en) * 2005-06-23 2010-06-09 微软公司 Adopt the handwriting recognition of neural network
CN104850837A (en) * 2015-05-18 2015-08-19 西南交通大学 Handwritten character recognition method
CN105354538A (en) * 2015-10-13 2016-02-24 广东小天才科技有限公司 Chinese character handwriting recognition method and system
CN105320961A (en) * 2015-10-16 2016-02-10 重庆邮电大学 Handwriting numeral recognition method based on convolutional neural network and support vector machine

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
周琳霞 等: "基于前向神经网络的与内容无关的笔迹鉴别", 《南昌航空工业学院学报(自然科学版)》 *

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106997473A (en) * 2016-09-08 2017-08-01 汪润春 A kind of image-recognizing method based on neutral net
CN106408038A (en) * 2016-09-09 2017-02-15 华南理工大学 Rotary Chinese character identifying method based on convolution neural network model
CN107292280A (en) * 2017-07-04 2017-10-24 盛世贞观(北京)科技有限公司 A kind of seal automatic font identification method and identifying device
CN108108746A (en) * 2017-09-13 2018-06-01 湖南理工学院 License plate character recognition method based on Caffe deep learning frames
CN108108746B (en) * 2017-09-13 2021-04-09 湖南理工学院 License plate character recognition method based on Caffe deep learning framework
CN107657230A (en) * 2017-09-27 2018-02-02 安徽硕威智能科技有限公司 A kind of bank self-help robot character recognition device
CN108154136A (en) * 2018-01-15 2018-06-12 众安信息技术服务有限公司 For identifying the method, apparatus of writing and computer-readable medium
CN108154136B (en) * 2018-01-15 2022-04-05 众安信息技术服务有限公司 Method, apparatus and computer readable medium for recognizing handwriting
CN108596168B (en) * 2018-04-20 2020-11-20 京东数字科技控股有限公司 Method, apparatus and medium for recognizing characters in image
CN108596168A (en) * 2018-04-20 2018-09-28 北京京东金融科技控股有限公司 For identification in image character method, apparatus and medium
CN108665010A (en) * 2018-05-12 2018-10-16 新疆大学 A kind of hand script Chinese input equipment Uighur words data enhancement methods
CN108665010B (en) * 2018-05-12 2022-01-04 新疆大学 Online handwriting Uygur language word data enhancement method
CN109740605A (en) * 2018-12-07 2019-05-10 天津大学 A kind of handwritten Chinese text recognition method based on CNN
CN109815809A (en) * 2018-12-19 2019-05-28 天津大学 A kind of English handwriting identification method based on CNN
CN109858488A (en) * 2018-12-28 2019-06-07 众安信息技术服务有限公司 A kind of handwriting samples recognition methods and system based on sample enhancing
CN111723807A (en) * 2019-03-20 2020-09-29 Sap欧洲公司 Recognizing machine-typed and handwritten characters using end-to-end deep learning
CN111723807B (en) * 2019-03-20 2023-12-26 Sap欧洲公司 End-to-end deep learning recognition machine for typing characters and handwriting characters
CN111738167A (en) * 2020-06-24 2020-10-02 华南理工大学 Method for recognizing unconstrained handwritten text image
CN112989834A (en) * 2021-04-15 2021-06-18 杭州一知智能科技有限公司 Named entity identification method and system based on flat grid enhanced linear converter

Also Published As

Publication number Publication date
CN105893968B (en) 2019-06-14

Similar Documents

Publication Publication Date Title
CN105893968A (en) Text-independent end-to-end handwriting recognition method based on deep learning
Le et al. Pattern generation strategies for improving recognition of handwritten mathematical expressions
El Abed et al. On-line Arabic handwriting recognition competition: ADAB database and participating systems
CN107729865A (en) A kind of handwritten form mathematical formulae identified off-line method and system
CN106384094A (en) Chinese word stock automatic generation method based on writing style modeling
El Abed et al. Icdar 2009 online arabic handwriting recognition competition
CN108664975B (en) Uyghur handwritten letter recognition method and system and electronic equipment
Chowdhury et al. Online handwriting recognition using Levenshtein distance metric
CN110334724B (en) Remote sensing object natural language description and multi-scale correction method based on LSTM
Hu et al. MST-based visual parsing of online handwritten mathematical expressions
CN113076900B (en) Test paper head student information automatic detection method based on deep learning
CN113920516B (en) Calligraphy character skeleton matching method and system based on twin neural network
He et al. Context-aware mathematical expression recognition: An end-to-end framework and a benchmark
CN111310820A (en) Foundation meteorological cloud chart classification method based on cross validation depth CNN feature integration
CN112651323B (en) Chinese handwriting recognition method and system based on text line detection
Jiulong et al. Detecting Chinese calligraphy style consistency by deep learning and one-class SVM
Boutarfass et al. Convolutional autoencoder for discriminating handwriting styles
Solanki et al. Printed Gujarati script OCR using hopfield neural network
Hajihashemi et al. A pattern recognition based Holographic Graph Neuron for Persian alphabet recognition
CN113903043B (en) Method for identifying printed Chinese character font based on twin metric model
Khorsheed Recognizing cursive typewritten text using segmentation-free system
CN115359486A (en) Method and system for determining custom information in document image
CN103793720A (en) Method and system for positioning eyes
Teulings et al. An on-line handwriting-recognition system based on unreliable modules
CN106599901A (en) Object segmentation and behavior identification coordinated method based on deep Boltzmann machine

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20220706

Address after: 401120 no.17-1, building 13, No.106, west section of Jinkai Avenue, Yubei District, Chongqing

Patentee after: CHONGQING AOS ONLINE INFORMATION TECHNOLOGY CO.,LTD.

Address before: 510640 South China University of technology, 381 Wushan Road, Tianhe District, Guangzhou City, Guangdong Province

Patentee before: SOUTH CHINA University OF TECHNOLOGY

CP03 Change of name, title or address
CP03 Change of name, title or address

Address after: 401121 no.17-1, building 13, No.106, west section of Jinkai Avenue, Yubei District, Chongqing

Patentee after: Chongqing Sign Digital Technology Co.,Ltd.

Country or region after: China

Address before: 401120 no.17-1, building 13, No.106, west section of Jinkai Avenue, Yubei District, Chongqing

Patentee before: CHONGQING AOS ONLINE INFORMATION TECHNOLOGY CO.,LTD.

Country or region before: China