CN101751565B - Method for character identification through fusing binary image and gray level image - Google Patents

Method for character identification through fusing binary image and gray level image Download PDF

Info

Publication number
CN101751565B
CN101751565B CN 200810239331 CN200810239331A CN101751565B CN 101751565 B CN101751565 B CN 101751565B CN 200810239331 CN200810239331 CN 200810239331 CN 200810239331 A CN200810239331 A CN 200810239331A CN 101751565 B CN101751565 B CN 101751565B
Authority
CN
China
Prior art keywords
image
character
sigma
feature
gray level
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN 200810239331
Other languages
Chinese (zh)
Other versions
CN101751565A (en
Inventor
张树武
杨武夷
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN 200810239331 priority Critical patent/CN101751565B/en
Publication of CN101751565A publication Critical patent/CN101751565A/en
Application granted granted Critical
Publication of CN101751565B publication Critical patent/CN101751565B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Character Discrimination (AREA)

Abstract

The invention relates to a method for character identification through fusing a binary image and a gray level image, which comprises the following steps: processing the binary image of a character image and a fusion image of the gray level image to carry out character identification; fusing the binary image of the character image and the gray level image to obtain the fusion image; normalizing the dimension and the position of the fusion image; extracting the features of a gradient histogram of a normalized image; obtaining a feature dimension reduction transformation matrix by utilizing principle component analysis and linear discriminant analysis; and establishing a character feature template library to carry out the character identification. The invention overcomes the defect that the traditional character identification technology of the binary image based on characters or the gray level image based on characters can not simultaneously identify the degenerated character image or the character image containing complex background.

Description

Merge the method for the character recognition of bianry image and gray level image
Technical field
The invention belongs to character recognition field (being called for short OCR), relate to the method for the character recognition of consistent fusion bianry image and gray level image.
Background technology
Traditional character recognition technologies is based on the bianry image of character or based on the gray level image of character.When the recognition technology based on the bianry image of character is applied to various low-quality images, such as the low-resolution images such as character picture in the degraded character image in the video, ID Card Image, license plate, the natural scene, because the character picture quality after the binaryzation is low, recognition effect is poor.When the recognition technology based on the gray level image of character is applied to comprise the character picture of complex background, such as the character picture in the video, because character picture comprises incomparable inconsistent background, recognition effect is with variation.
Summary of the invention
In order to solve the problem of prior art, the object of the present invention is to provide a kind of bianry image of character and method that gray level image carries out character recognition of merging.
For reaching described purpose, the method for the character recognition of fusion bianry image provided by the invention and gray level image is processed the fused images of described bianry image and gray level image, carries out character recognition, and it may further comprise the steps:
Step 1: the bianry image of establishing the single character picture that obtains after the pre-service is B 0=[b 0(x, y)], the value that wherein is positioned at the pixel of the capable y row of x is b 0(x, y), b 0(x, y) is 0 or 1, and the size of image is W 1* H 1The gray level image of character is G c=[g c(x, y)], the value that is positioned at the pixel of the capable y row of x is g c(x, y), 0≤g c(x, y)≤255; Bianry image B with character picture 0With gray level image G cMerge, get the image G=[g (x, y) after the fusion], the value that is positioned at the pixel of the capable y row of x is g (x, y), 0≤g (x, y)≤255;
Step 2: extracting fused images G=[g (x, y)] feature before, carry out first fused images G=[g (x, y)] the position and the normalized of size; The input picture that image normalization is processed is G=[g (x, y)], the output image after the normalization is F=[f (x ', y ')], its size is respectively W 1* H 1And W 2* H 2Input picture G=[g (x, y)] pixel that is positioned at the capable y of x row will be mapped to F=[f (x ', y ')] being positioned at the pixel of x ' row y ' row, the coordinate by input picture and output image shines upon to realize image normalization:
x ′ = x ′ ( x , y ) y ′ = y ′ ( x , y )
One dimension coordinate is mapped as:
x ′ = x ′ ( x ) y ′ = y ′ ( x ) ;
Step 3: the feature of extracting the histogram of gradients of normalized image based on histogram of gradients;
Step 4: utilize principal component analysis and linear discriminant analysis that the feature of the histogram of gradients of normalized image is carried out dimension-reduction treatment, obtain the transformation matrix of Feature Dimension Reduction;
Step 5: set up character feature template library, read the transformation matrix of Feature Dimension Reduction and character is identified.
Beneficial effect of the present invention: the invention is characterized in the bianry image of character picture and the fused images of gray level image are processed that carry out character recognition, it may further comprise the steps: the fusion of (1) bianry image and gray level image; (2) normalization of image; (3) based on the feature extraction of histogram of gradients; (4) Feature Dimension Reduction; (5) classifier design and character recognition.The present invention has overcome based on the bianry image of character or the shortcoming that can not identify simultaneously the degraded character image and comprise the character picture of complex background based on traditional character recognition technologies of the gray level image of character.The technical field of application of the present invention comprises the character recognition in the video, the character recognition in ID Card Image, license plate, the natural scene image
Description of drawings
Fig. 1 is character recognition system process flow diagram of the present invention;
Fig. 2 is the framework synoptic diagram of the fusion of bianry image of the present invention and gray level image;
Fig. 3 is the normalized framework synoptic diagram of image of the present invention;
Fig. 4 is the framework synoptic diagram that the present invention is based on the feature extraction of histogram of gradients;
Fig. 5 is the configuration diagram that the present invention asks the transformation matrix of Feature Dimension Reduction;
Fig. 6 is classifier design of the present invention and character recognition configuration diagram;
Fig. 7 is Sobel gradient operator template;
Fig. 8 is L reference direction example, left side L=4, the right L=8;
Fig. 9 is that gradient is decomposed example;
Figure 10 be calculating pixel and center, rectangular area in the horizontal direction with vertical direction on distance examples.
Embodiment
Describe each related detailed problem in the technical solution of the present invention in detail below in conjunction with accompanying drawing.Be to be noted that described embodiment only is intended to be convenient to the understanding of the present invention, and it is not played any restriction effect.
As shown in Figure 1, character recognition system process flow diagram of the present invention, recognizer can be divided into two parts: training system and recognition system.Training system merges its bianry image and gray level image to each character training sample, and fused images is carried out the normalization of size and position, extracts the feature of histogram of gradients; The feature that utilization is extracted from training sample is found the solution the transformation matrix that carries out Feature Dimension Reduction, obtains the character recognition storehouse.In recognition system, merge bianry image and the gray level image of character to be identified, fused images is carried out the normalization of size and position, extract the feature of histogram of gradients, the transformation matrix that utilizes training system to obtain carries out dimensionality reduction to feature, then sends into recognizer, obtains recognition result.
Fusion character bianry image and gray level image carry out the realization of character recognition system need to consider following several aspect:
1) realization of training system;
2) realization of recognition system.
The below describes in detail to these two aspects respectively.
The realization of 1 training system
1.1 see also the framework of the fusion of the bianry image shown in Fig. 2 and gray level image.
If the bianry image of the single character picture that obtains after the pre-service is B 0=[b 0(x, y)], the value that wherein is positioned at the pixel of the capable y row of x is b 0(x, y), b 0(x, y) is 0 or 1.The gray level image of character is G c=[g c(x, y)], the value that is positioned at the pixel of the capable y row of x is g c(x, y), 0≤g c(x, y)≤255.Bianry image is B 0With gray level image G cSize be W 1* H 1Carry out the bianry image B of character by following flow process 0Gray level image G with character cFusion, the image G=[g (x, y) after obtaining merging], the value that is positioned at the pixel of the capable y of x row is g (x, y), 0≤g (x, y)≤255:
Ask a total head threshold value th, for example can be to gray level image g c(x, y) utilizes traditional maximum variance between clusters (OSTU method) to try to achieve this total head threshold value th Ostu, make th=a x th Ostu, a is a constant.Utilize global threshold th that gray level image g (x, y) is carried out threshold processing and obtain bianry image B g=[b g(x, y)], b g(x, y) is defined as:
b g ( x , y ) = 1 g c ( x , y ) > th 0 g c ( x , y ) ≤ th , x = 0 , · · · , W 1 - 1 , y = 0 , · · · , H 1 - 1 ;
Utilize two-value morphology to bianry image B oCarry out condition expansion.If D is the image of 3 x 3, the pixel value of its each position is 1.To bianry image B oCarrying out condition expansion is:
B i + 1 = ( B i ⊕ D ) ∩ B g , i = 1,2 , · · · , N , B 1 = B 0
According to following formula to bianry image B iRepeatedly carry out condition expansion, until B I+1=B iOr reach maximum iterations, establishing the bianry image that obtains at last is B=[b (x, y)].
Obtain the fused images G=[g (x, y) for single character recognition], g (x, y) is defined as:
g ( x , y ) = g c ( x , y ) b ( x , y ) = 1 0 b ( x , y ) = 0 , x = 0 , · · · , W 1 - 1 , y = 0 , · · · , H 1 - 1 .
1.2 see also the normalized framework of the image shown in Fig. 3.
Before the feature of extracting fused images, the normalized of the position of advanced line character image and size.The input picture of image normalization is G=[g (x, y)], the output image after the normalization is F=[f (x ', y ')], its size is respectively W 1* H 1And W 2* H 2Input picture G=[g (x, y)] pixel that is positioned at the capable y of x row will be mapped to F=[f (x ', y ')] being positioned at the pixel of x ' row y ' row, the coordinate by input picture and output image shines upon to realize image normalization:
x ′ = x ′ ( x , y ) y ′ = y ′ ( x , y )
One dimension coordinate is mapped as
x ′ = x ′ ( x ) y ′ = y ′ ( x ) ;
Calculate fused images G=[g (x, y)] barycenter (x c, y c), barycenter is adjusted into normalized image F=[f (x ', y ')] center (W 2/ 2, H 2/ 2):
g x ( x ) = Σ y = 0 H 1 - 1 g ( x , y ) / Σ x = 0 W 1 - 1 Σ y = 0 H 1 - 1 g ( x , y ) , x = 0 , · · · , W 1 - 1 ,
g y ( y ) = Σ x = 0 W 1 - 1 g ( x , y ) / Σ x = 0 W 1 - 1 Σ y = 0 H 1 - 1 g ( x , y ) , y = 0 , · · · , H 1 - 1 ,
x c = Σ x = 0 W 1 - 1 x g x ( x ) ,
y c = Σ y = 0 H 1 - 1 y g y ( y ) ,
G wherein x(x) and g y(y) be respectively fused images G=[g (x, y)] in the vertical direction with horizontal direction on picture element density;
According to centroid position (x c, y c), computed image G=[g (x, y)] monolateral second moment
Figure G2008102393315D00055
Figure G2008102393315D00056
Figure G2008102393315D00057
With
Figure G2008102393315D00058
μ x + = Σ x > x c ( x - x c ) 2 g x ( x )
&mu; x + = &Sigma; x < x c ( x - x c ) 2 g x ( x )
&mu; y + = &Sigma; y > y c ( y - y c ) 2 g y ( y )
&mu; y + = &Sigma; y < y c ( y - y c ) 2 g y ( y ) ;
The housing that input picture is set according to the monolateral second moment that calculates is [ x c - 2 &mu; x - , x c + 2 &mu; x + ] With [ y c - 2 &mu; y - , y c + 2 &mu; y + ] . For the x axle, find the solution quadratic function u (x)=ax 2+ bx+c is three points on the x axle ( x c - 2 &mu; x - , x c , x c + 2 &mu; x + ) Be mapped as respectively (0,0.5,1), in like manner obtain the quadratic function u (y) of y axle three points on the y axle ( y c - 2 &mu; y - , y c , y c + 2 &mu; y + ) Be mapped as respectively (0,0.5,1); Obtain input picture G=[g (x, y)] be positioned at the pixel of the capable y of x row and output image F=[f (x ', y ')] be positioned at the coordinate mapping function of the pixel of x ' row y ' row:
x &prime; = W 2 u ( x ) y &prime; = H 2 u ( y ) ;
Determine input picture G=[g (x according to the coordinate mapping function, y)] with normalized image F=[f (x ', y ')] the coordinate mapping relations, the input picture gray-scale value is passed through bilinear interpolation, obtain normalized image F=[f (x ', y ')] value.
1.3 see also the framework based on the feature extraction of histogram of gradients shown in Fig. 4.
Two 3 * 3 templates utilizing the Sobel operator are computed image F=[f (x, y) respectively] in each locational gradient, two 3x3 templates of Sobel operator are as shown in Figure 7.For image F=[f (x, y)], it is tried to achieve by following formula along x axle and the axial first order derivative component of y respectively:
g x(x,y)=f(x+1,y-1)+2f(x+1,y)+f(x+1,y+1)
-f(x-1,y-1)-2f(x-1,y)-f(x-1,y+1),
g y(x,y)=f(x-1,y+1)+2f(x,y+1)+f(x+1,y+1)
-f(x-1,y-1)-2f(x,y-1)-f(x+1,y-1).
x=0,...,W 2-1,y=1,...,H 2-1;
Computed image F=[f (x, y)] gradient intensity mag (x, y) and the deflection of position (x, y)
Figure G2008102393315D0006104115QIETU
(x, y) is respectively:
mag ( x , y ) = [ g x 2 ( x , y ) + g y 2 ( x , y ) ] 1 / 2 ,
Figure G2008102393315D00062
Define L reference direction, the situation of L=4 and L=8 as Fig. 8 and shown in.Utilize parallelogram law to be decomposed into from its two nearest reference direction, as shown in Figure 9 gradient.Be W with size after the normalization 2* H 2Image F=[f (x, y)] be divided into R * R mutually disjoint rectangular area, set up the gradient orientation histogram of L dimension for each rectangular area.Image F=[f (x, y)] in the gradient orientation histogram of nearest 4 rectangular areas of gradient pair and this pixel of each pixel contribution is arranged.Be 4 rectangular areas (from top to bottom, from left to right being numbered respectively 1,2,3 and 4) that pixel is nearest with it as shown in figure 10, wherein each little rectangle frame represents a pixel, and 4 * 4 little rectangle frames form a large rectangle zone.In the horizontal direction, the distance at pixel and center, rectangular area is respectively d HlAnd d HrIn the vertical direction, the distance at pixel and center, rectangular area is respectively d VtAnd d VbIf the intensity of the component of pixel gradient on the l direction is g l, then the gradient of this pixel is to the 1st, 2, and the contribution margin of the l of the gradient orientation histogram of 3 and 4 rectangular areas dimension is respectively g l* d Hr* d Vb/ ((d Hl+ d Hr) * (d Vt+ d Vb)), g l* d Hl* d Vb/ ((d Hl+ d Hr) * (d Vt+ d Vb)), g l* d Hr* d Vt/ ((d Hl+ d Hr) * (d Vt+ d Vb)) and g l* d Hl* d Vt/ ((d Hl+ d Hr) * (d Vt+ d Vb)).Utilize this method to calculate the gradient of each pixel to the contribution of the gradient orientation histogram of rectangular area adjacent thereto, try to achieve each rectangular area gradient orientation histogram, obtained at last the R * R of character picture * L dimensional feature.
1.4 see also the framework of the transformation matrix of asking Feature Dimension Reduction shown in Fig. 5:
1.4.1 principal component analysis (PCA)
The high dimensional feature vector comprises the feature that is mutually related, and it is large that it is processed operand, utilizes principal component analysis that the high dimensional feature vector is carried out principal component analysis (PCA), finds the solution PCA dimensionality reduction matrix P PCAIf the character feature that extracts from n training sample is x i, i=1 ..., n, x iDimension m=R * R * L; The Scatter Matrix of training sample character feature is:
&Sigma; = 1 n &Sigma; i = 1 n ( x i - x &OverBar; ) T ( x i - x &OverBar; ) , x &OverBar; = &Sigma; i = 1 n x i
Scatter Matrix is carried out Eigenvalues Decomposition is:
∑=UΛU T
U=[u wherein 1, u 2..., u m] be orthogonal matrix, Λ=diag (λ 1, λ 2..., λ m) be diagonal matrix, λ 1〉=λ 2〉=... 〉=λ mBe eigenwert.If will preserve the energy of r% behind the principal component analysis PCA dimensionality reduction, then the principal direction number l of principal component analysis preservation is
l = arg min k ( &Sigma; i = 1 k &lambda; i &Sigma; i = 1 m &lambda; i &GreaterEqual; r )
The transformation matrix that principal component analysis obtains is P PCA=[u 1, u 2..., u l], to character feature x i, carry out the l dimension character feature z after dimensionality reduction obtains dimensionality reduction i=(P PCA) Tx i, i=1 ..., n, (P PCA) TExpression P PCATransposed matrix;
1.4.2 the character feature behind the training sample dimensionality reduction is carried out linear discriminant analysis (LDA), finds the solution transformation matrix W:
If character class number to be identified in the recognition system is C, the i class comprises n iIndividual training sample.Calculate i class character sample characteristic mean μ iWith all sample characteristics average μ:
&mu; i = 1 n i &Sigma; k = 1 n i z k i , &mu; i = 1 n &Sigma; i = 1 C &Sigma; k = 1 n i z k i , n = &Sigma; i = 1 C n i
Scatter Matrix S between compute classes bWith Scatter Matrix S in the class w:
S b = &Sigma; i = 1 C n i n ( &mu; i - &mu; ) ( &mu; i - &mu; ) T
S w = &Sigma; i = 1 C ( n i n &Sigma; k = 1 n i ( z k i - &mu; i ) ( z k i - &mu; i ) T )
A transformation matrix W is sought in linear discriminant analysis so that after the conversion between class dispersion as far as possible large, dispersion is as far as possible little in the class simultaneously, utilizes the maximization criterion
J = tr ( W T S b W ) tr ( W T S w W )
Represent.LDA can solve by finding the solution the generalized eigenvector problem:
S bw=λS ww
If vectorial w 1..., w d..., w lBe the solution of generalized eigenvector problem, the generalized eigenvalue λ of their correspondences 1〉=... 〉=λ d〉=... 〉=λ l, the solution proper vector of d generalized eigenvector problem forms W, i.e. W=[w before selecting 1..., w d].
1.5 see also the classifier design shown in Fig. 6 and character recognition framework:
Utilize transformation matrix W to i character type characteristic mean μ iCarry out dimensionality reduction, and the feature behind the normalization dimensionality reduction
&mu; i * = W T &mu; i , &mu; &OverBar; i = &mu; i * / ( &mu; i * ) T &mu; i *
Preserve transformation matrix P=WP PCA, the coding of each character type and characteristic of correspondence μ thereof iIn the File of identification storehouse.
(5.2) character recognition
The realization of 2 recognition systems
From the File of character recognition storehouse, read transformation matrix P, the coding of each character type and characteristic of correspondence μ thereof iBianry image and gray level image to each character to be identified merge, and the image after merging is carried out normalization, carry out the multidimensional characteristic a that feature extraction obtains character picture.Utilize transformation matrix P that the multidimensional characteristic a of character picture is carried out feature b=P after Feature Dimension Reduction obtains dimensionality reduction TA, P TTransposed matrix for transformation matrix P.Feature normalization behind the dimensionality reduction is obtained
b &OverBar; = b / b T b .
Normalization center vector { the μ of order computation b and each character type i} 1≤i≤CCosine distance { d i} 1≤i≤C
d i=1-y Tμ
The minimum class of distance is the recognition result of character picture.
The above; only be the embodiment among the present invention; but protection scope of the present invention is not limited to this; anyly be familiar with the people of this technology in the disclosed technical scope of the present invention; can understand conversion or the replacement expected; all should be encompassed in of the present invention comprising within the scope, therefore, protection scope of the present invention should be as the criterion with the protection domain of claims.

Claims (5)

1. a method that merges the character recognition of bianry image and gray level image is characterized in that, the fused images of described bianry image and gray level image is processed, and carries out character recognition, and it may further comprise the steps:
Step 1: the bianry image of establishing the single character picture that obtains after the pre-service is B 0=[b 0(x, y)], the value that wherein is positioned at the pixel of the capable y row of x is b 0(x, y), b 0(x, y) is 0 or 1, and the size of image is W 1* H 1The gray level image of character is G c=[g c(x, y)], the value that is positioned at the pixel of the capable y row of x is g c(x, y), 0≤g c(x, y)≤255; Bianry image B with character picture 0With gray level image G cMerge the image G=[g (x, y) after obtaining merging], the value that is positioned at the pixel of the capable y row of x is g (x, y), 0≤g (x, y)≤255;
Step 2: extracting fused images G=[g (x, y)] feature before, carry out first fused images G=[g (x, y)] the position and the normalized of size; The input picture that image normalization is processed is G=[g (x, y)], the output image after the normalization is F=[f (x ', y ')], its size is respectively W 1* H 1And W 2* H 2Input picture G=[g (x, y)] pixel that is positioned at the capable y of x row will be mapped to F=[f (x ', y ')] being positioned at the pixel of x ' row y ' row, the coordinate by input picture and output image shines upon to realize image normalization:
x &prime; = x &prime; ( x , y ) y &prime; = y &prime; ( x , y )
One dimension coordinate is mapped as:
x &prime; = x &prime; ( x ) y &prime; = y &prime; ( x ) ;
Step 3: the feature of extracting the histogram of gradients of normalized image based on histogram of gradients;
Step 4: utilize principal component analysis and linear discriminant analysis that the feature of the histogram of gradients of normalized image is carried out dimension-reduction treatment, obtain the transformation matrix of Feature Dimension Reduction;
Step 5: set up character feature template library, read the transformation matrix of Feature Dimension Reduction and character is identified;
The fusion of described bianry image and gray level image comprises:
Step 11: to gray level image G c=[g c(x, y)] pixel point value g c(x, y) utilizes traditional maximum variance between clusters to try to achieve threshold value th Ostu, ask a global threshold th, make th=a * th Ostu, a is a constant; Utilize total head threshold value th to gray level image G c=[g c(x, y)] pixel point value g c(x, y) carries out threshold processing, obtains bianry image B g=[b g(x, y)], the pixel point value b of bianry image g(x, y) is defined as:
b g ( x , y ) = 1 g c ( x , y ) > th 0 g c ( x , y ) &le; th , x=0,...,W 1-1,y=0,...,H 1-1;
Step 12: utilize two-value morphology to bianry image B oCarry out condition expansion, establishing D is the image of a 3x3, and the pixel value of its each position is 1; To bianry image B oCarrying out condition expansion is:
B i + 1 = ( B i &CirclePlus; ) &cap; B g , i=1,2,...,N,B 1=B 0
According to following formula to bianry image B iRepeatedly carry out condition expansion, until B I+1=B iOr reach maximum iterations, establishing the bianry image that obtains at last is B=[b (x, y)];
Step 13: obtain the fused images G=[g (x, y) for single character recognition], g (x, y) is defined as:
g ( x , y ) = g c ( x , y ) b ( x , y ) = 1 0 b ( x , y ) = 0 , x=0,...,W 1-1,y=0,...,H 1-1。
2. the method for the character recognition of described fusion bianry image and gray level image according to claim 1 is characterized in that the normalization of image comprises:
Step 21: calculate fused images G=[g (x, y)] barycenter (x c, y c), barycenter is adjusted into normalized image F=[f (x ', y ')] center (W 2/ 2, H 2/ 2):
g x ( x ) = &Sigma; y = 0 H 1 - 1 g ( x , y ) / &Sigma; x = 0 W 1 - 1 &Sigma; x = 0 H 1 - 1 g ( x , y ) , x=0,...,W 1-1,
g y ( y ) = &Sigma; x = 0 W 1 - 1 g ( x , y ) / &Sigma; x = 0 W 1 - 1 &Sigma; y = 0 H 1 - 1 g ( x , y ) , y=0,...,H 1-1,
x c = &Sigma; x = 0 W 1 - 1 xg x ( x ) ,
y c = &Sigma; y = 0 H 1 - 1 yg y ( y ) ,
G wherein x(x) and g y(y) be respectively fused images G=[g (x, y)] in the vertical direction with horizontal direction on picture element density;
Step 22: according to centroid position (x c, y c), computed image G=[g (x, y)] monolateral second moment
Figure FDA00002237040400028
Figure FDA00002237040400029
With
Figure FDA000022370404000210
&mu; x + = &Sigma; x > x c ( x - x c ) 2 g x ( x )
&mu; x - = &Sigma; x < x c ( x - x c ) 2 g x ( x )
&mu; y + = &Sigma; y > y c ( y - y c ) 2 g y ( y )
&mu; y - = &Sigma; y < y c ( y - y c ) 2 g y ( y )
Step 23: the housing that input picture is set according to the monolateral second moment that calculates is:
[ x c - 2 &mu; x - , x c = 2 &mu; x + ] With [ y c - 2 &mu; y - , y c + 2 &mu; y + ] ;
For the x axle, find the solution quadratic function u (x)=ax 2+ bx+c is three points on the x axle
Figure FDA00002237040400035
Be mapped as respectively (0,0.5,1), in like manner obtain the quadratic function u (y) of y axle three points on the y axle
Figure FDA00002237040400036
Be mapped as respectively (0,0.5,1); Obtain input picture G=[g (x, y)] be positioned at the pixel of the capable y of x row and output image F=[f (x ', y ')] be positioned at the coordinate mapping function of the pixel of x ' row y ' row:
x &prime; = W 2 u ( x ) y &prime; = H 2 u ( y ) , W 2, H 2Be respectively output image F=[f (x ', y ')] wide and high;
Step 24: finally obtain normalized image F=[f (x ', y ') by bilinear interpolation] value.
3. the method for the character recognition of described fusion bianry image and gray level image according to claim 1 is characterized in that, the step of the feature of the described histogram of gradients of extracting normalized image based on histogram of gradients comprises:
Step 31: two 3 * 3 templates utilizing the Sobel operator are computed image F=[f (x, y) respectively] in each locational gradient; For image F=[f (x, y)], it is tried to achieve by following formula along x axle and the axial first order derivative component of y respectively:
g x(x,y)=f(x+1,y-1)+2f(x+1,y)+f(x+1,y+1)
-f(x-1,y-1)-2f(x-1,y)-f(x-1,y+1),
g y(x,y)=f(x-1,y+1)+2f(x,y+1)+f(x+1,y+1),
-f(x-1,y-1)-2f(x,y-1)-f(x+1,y-1).
x=0,...,W 2-1,y=0,...,H 2-1;
Step 32: image F=[f (x, y)] gradient intensity mag (x, y) and the deflection of position (x, y) Be respectively:
mag ( x , y ) = [ g x 2 ( x , y ) + g x 2 ( x , y ) ] 1 / 2 ,
Figure FDA000022370404000310
Step 33: defining L reference direction, utilize parallelogram law to be decomposed into from its two nearest reference direction gradient, is W with size after the normalization 2* H 2Image F=[f (x, y)] be divided into R * R mutually disjoint rectangular area, set up the gradient orientation histogram of L dimension for each rectangular area; Image F=[f (x, y)] in the gradient orientation histogram of nearest 4 rectangular areas of gradient pair and this pixel of each pixel contribution is arranged; Calculate the gradient of each pixel to the contribution of the gradient orientation histogram of rectangular area adjacent thereto, try to achieve each rectangular area gradient orientation histogram, obtained at last the R * R of character picture * L dimensional feature.
4. the method for the character recognition of described fusion bianry image and gray level image according to claim 1 is characterized in that the described step of utilizing principal component analysis and linear discriminant analysis that the feature of the histogram of gradients of normalized image is carried out dimension-reduction treatment comprises:
Step 41: the high dimensional feature vector is carried out principal component analysis (PCA), find the solution PCA dimensionality reduction matrix P PCA:
The high dimensional feature vector comprises the feature that is mutually related, and it is large that it is processed operand, utilizes principal component analysis that the high dimensional feature vector is carried out principal component analysis (PCA), finds the solution PCA dimensionality reduction matrix P PCAIf the character feature that extracts from n training sample is x i, i=1 ..., n, x iDimension m=R * R * L; The Scatter Matrix of training sample character feature is:
&Sigma; = 1 n &Sigma; i = 1 n ( x i - x &OverBar; ) T ( x i - x &OverBar; ) , x &OverBar; = &Sigma; i = 1 n x i ,
Scatter Matrix is carried out Eigenvalues Decomposition is:
∑=UΛU T
U=[u wherein 1, u 2..., u m] be orthogonal matrix, Λ=diag (λ 1, λ 2..., λ m) be diagonal matrix, λ 1〉=λ 2〉=... 〉=λ mBe eigenwert, establish the energy that will preserve r% behind the principal component analysis PCA dimensionality reduction, then the principal direction number l that preserves of principal component analysis is:
l = arg min k ( &Sigma; i = 1 k &lambda; i &Sigma; i = 1 m &lambda; i &GreaterEqual; r )
The transformation matrix that principal component analysis obtains is P PCA=[u 1, u 2..., u l], to character feature x i, carry out the l dimension character feature z after dimensionality reduction obtains dimensionality reduction i=(P PCA) Tx i, i=1 ..., n, (P PCA) TExpression P PCATransposed matrix;
Step 42: the character feature behind the training sample dimensionality reduction is carried out linear discriminant analysis, find the solution transformation matrix W:
If character class number to be identified in the recognition system is C, the i class comprises n iIndividual training sample; Calculate i class character sample characteristic mean μ iWith all sample characteristics average μ:
&mu; i = 1 n i &Sigma; k = 1 n i z k i , &mu; = 1 n &Sigma; i = 1 C &Sigma; k = 1 n i z k i , n = &Sigma; i = 1 C n i
Scatter Matrix S between compute classes bWith Scatter Matrix S in the class w:
S b = &Sigma; i = 1 C n i n ( &mu; i - &mu; ) ( &mu; i - &mu; ) T
S w = &Sigma; i = 1 C ( n i n &Sigma; k = 1 n i ( z k i - &mu; i ) ( z k i - &mu; i ) T )
A transformation matrix W is sought in linear discriminant analysis so that after the conversion between class dispersion as far as possible large, dispersion is as far as possible little in the class simultaneously, utilizes the maximization criterion
J = tr ( W T S b W ) tr ( W T S w W ) Represent;
Solve linear discriminant analysis by finding the solution the generalized eigenvector problem:
S bw=λS ww
If vectorial w 1..., w d..., w lBe the solution of generalized eigenvector problem, the generalized eigenvalue λ of their correspondences 1〉=... 〉=λ d〉=... 〉=λ l, the solution proper vector of d generalized eigenvector problem forms W, i.e. W=[w before selecting 1..., w d].
5. the method for the character recognition of described fusion bianry image and gray level image according to claim 4 is characterized in that described character recognition comprises:
Step 51: design category device
Utilize transformation matrix W to i character type characteristic mean μ iCarry out dimensionality reduction, and the feature behind the normalization dimensionality reduction &mu; i * = W T &mu; i , &mu; &OverBar; i = &mu; i * / ( &mu; i * ) T &mu; i * ;
Preserve transformation matrix P=WP PCA, the coding of each character type and corresponding character type feature thereof
Figure FDA00002237040400059
In the file in the character recognition storehouse;
Step 52: character recognition
Read coding and the corresponding character type feature thereof of transformation matrix P, each character type in the file from the character recognition storehouse
Figure FDA00002237040400061
Bianry image and gray level image to each character to be identified merge, and the image after merging is carried out normalization, carry out the multidimensional characteristic a that feature extraction obtains character picture; Utilize transformation matrix P that the multidimensional characteristic a of character picture is carried out feature b=P after Feature Dimension Reduction obtains dimensionality reduction TA, P TTransposed matrix for transformation matrix P;
Feature b normalization behind the dimensionality reduction is obtained
Figure FDA00002237040400062
Order computation character normalization feature
Figure FDA00002237040400063
Normalization center vector with each character type Cosine distance { d i} 1≤i≤C
d i = 1 - y &OverBar; T &mu; &OverBar;
The minimum class of distance is the recognition result of character picture.
CN 200810239331 2008-12-10 2008-12-10 Method for character identification through fusing binary image and gray level image Expired - Fee Related CN101751565B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 200810239331 CN101751565B (en) 2008-12-10 2008-12-10 Method for character identification through fusing binary image and gray level image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 200810239331 CN101751565B (en) 2008-12-10 2008-12-10 Method for character identification through fusing binary image and gray level image

Publications (2)

Publication Number Publication Date
CN101751565A CN101751565A (en) 2010-06-23
CN101751565B true CN101751565B (en) 2013-01-02

Family

ID=42478527

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 200810239331 Expired - Fee Related CN101751565B (en) 2008-12-10 2008-12-10 Method for character identification through fusing binary image and gray level image

Country Status (1)

Country Link
CN (1) CN101751565B (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102011075275A1 (en) * 2011-05-04 2012-11-08 Bundesdruckerei Gmbh Method and device for recognizing a character
CN102750530B (en) * 2012-05-31 2014-11-26 贺江涛 Character recognition method and device
CN103854020B (en) * 2012-11-29 2018-11-30 捷讯平和(北京)科技发展有限公司 Character recognition method and device
CN103679208A (en) * 2013-11-27 2014-03-26 北京中科模识科技有限公司 Broadcast and television caption recognition based automatic training data generation and deep learning method
CN106257495A (en) * 2015-06-19 2016-12-28 阿里巴巴集团控股有限公司 A kind of digit recognition method and device
CN106203434B (en) * 2016-07-08 2019-07-19 中国科学院自动化研究所 File and picture binary coding method based on stroke structure symmetry
CN108319958A (en) * 2018-03-16 2018-07-24 福州大学 A kind of matched driving license of feature based fusion detects and recognition methods
CN108830138B (en) * 2018-04-26 2021-05-07 平安科技(深圳)有限公司 Livestock identification method, device and storage medium
CN109520706B (en) * 2018-11-21 2020-10-09 云南师范大学 Screw hole coordinate extraction method of automobile fuse box
CN109919253A (en) * 2019-03-27 2019-06-21 北京爱数智慧科技有限公司 Character identifying method, device, equipment and computer-readable medium
CN111583217A (en) * 2020-04-30 2020-08-25 深圳开立生物医疗科技股份有限公司 Tumor ablation curative effect prediction method, device, equipment and computer medium
CN112200247B (en) * 2020-10-12 2021-07-02 西安泽塔云科技股份有限公司 Image processing system and method based on multi-dimensional image mapping

Also Published As

Publication number Publication date
CN101751565A (en) 2010-06-23

Similar Documents

Publication Publication Date Title
CN101751565B (en) Method for character identification through fusing binary image and gray level image
Chen et al. Accurate and efficient traffic sign detection using discriminative adaboost and support vector regression
Chen et al. Fast human detection using a novel boosted cascading structure with meta stages
Mahadevan et al. Saliency-based discriminant tracking
Fukui et al. The kernel orthogonal mutual subspace method and its application to 3D object recognition
CN103870811B (en) A kind of front face Quick method for video monitoring
CN103198303B (en) A kind of gender identification method based on facial image
Riggan et al. Thermal to visible synthesis of face images using multiple regions
CN105023008A (en) Visual saliency and multiple characteristics-based pedestrian re-recognition method
CN102663413A (en) Multi-gesture and cross-age oriented face image authentication method
CN102609716A (en) Pedestrian detecting method based on improved HOG feature and PCA (Principal Component Analysis)
CN102163281B (en) Real-time human body detection method based on AdaBoost frame and colour of head
CN103413119A (en) Single sample face recognition method based on face sparse descriptors
Redondo-Cabrera et al. All together now: Simultaneous object detection and continuous pose estimation using a hough forest with probabilistic locally enhanced voting
CN105894018B (en) Classification of Polarimetric SAR Image method based on depth multi-instance learning
CN103605993B (en) Image-to-video face identification method based on distinguish analysis oriented to scenes
CN105138951B (en) Human face portrait-photo array the method represented based on graph model
Kruthi et al. Offline signature verification using support vector machine
Li et al. Multi-view face detection using support vector machines and eigenspace modelling
CN104008404B (en) Pedestrian detection method and system based on significant histogram features
CN111259780A (en) Single-sample face recognition method based on block linear reconstruction discriminant analysis
CN114612980A (en) Deformed face detection based on multi-azimuth fusion attention
Luzardo et al. Head pose estimation for sign language video
Hanani et al. Evaluating new variants of motion interchange patterns
Zhu et al. Car detection based on multi-cues integration

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20130102