CN101751565B - Method for character identification through fusing binary image and gray level image - Google Patents
Method for character identification through fusing binary image and gray level image Download PDFInfo
- Publication number
- CN101751565B CN101751565B CN 200810239331 CN200810239331A CN101751565B CN 101751565 B CN101751565 B CN 101751565B CN 200810239331 CN200810239331 CN 200810239331 CN 200810239331 A CN200810239331 A CN 200810239331A CN 101751565 B CN101751565 B CN 101751565B
- Authority
- CN
- China
- Prior art keywords
- image
- character
- sigma
- feature
- gray level
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000000034 method Methods 0.000 title claims abstract description 14
- 239000011159 matrix material Substances 0.000 claims abstract description 44
- 230000009466 transformation Effects 0.000 claims abstract description 25
- 230000004927 fusion Effects 0.000 claims abstract description 15
- 238000012545 processing Methods 0.000 claims abstract description 3
- 238000000513 principal component analysis Methods 0.000 claims description 33
- 238000010606 normalization Methods 0.000 claims description 22
- 238000012549 training Methods 0.000 claims description 15
- 238000000605 extraction Methods 0.000 claims description 5
- 238000013461 design Methods 0.000 claims description 4
- 239000006185 dispersion Substances 0.000 claims description 4
- 239000000284 extract Substances 0.000 claims description 4
- 238000013507 mapping Methods 0.000 claims description 4
- 238000012887 quadratic function Methods 0.000 claims description 4
- 238000006243 chemical reaction Methods 0.000 claims description 3
- 230000003760 hair shine Effects 0.000 claims description 3
- 238000000354 decomposition reaction Methods 0.000 claims description 2
- 238000005516 engineering process Methods 0.000 abstract description 7
- 230000007547 defect Effects 0.000 abstract 1
- 238000010586 diagram Methods 0.000 description 7
- 230000000694 effects Effects 0.000 description 3
- 230000026676 system process Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004321 preservation Methods 0.000 description 1
Images
Landscapes
- Character Discrimination (AREA)
Abstract
The invention relates to a method for character identification through fusing a binary image and a gray level image, which comprises the following steps: processing the binary image of a character image and a fusion image of the gray level image to carry out character identification; fusing the binary image of the character image and the gray level image to obtain the fusion image; normalizing the dimension and the position of the fusion image; extracting the features of a gradient histogram of a normalized image; obtaining a feature dimension reduction transformation matrix by utilizing principle component analysis and linear discriminant analysis; and establishing a character feature template library to carry out the character identification. The invention overcomes the defect that the traditional character identification technology of the binary image based on characters or the gray level image based on characters can not simultaneously identify the degenerated character image or the character image containing complex background.
Description
Technical field
The invention belongs to character recognition field (being called for short OCR), relate to the method for the character recognition of consistent fusion bianry image and gray level image.
Background technology
Traditional character recognition technologies is based on the bianry image of character or based on the gray level image of character.When the recognition technology based on the bianry image of character is applied to various low-quality images, such as the low-resolution images such as character picture in the degraded character image in the video, ID Card Image, license plate, the natural scene, because the character picture quality after the binaryzation is low, recognition effect is poor.When the recognition technology based on the gray level image of character is applied to comprise the character picture of complex background, such as the character picture in the video, because character picture comprises incomparable inconsistent background, recognition effect is with variation.
Summary of the invention
In order to solve the problem of prior art, the object of the present invention is to provide a kind of bianry image of character and method that gray level image carries out character recognition of merging.
For reaching described purpose, the method for the character recognition of fusion bianry image provided by the invention and gray level image is processed the fused images of described bianry image and gray level image, carries out character recognition, and it may further comprise the steps:
Step 1: the bianry image of establishing the single character picture that obtains after the pre-service is B
0=[b
0(x, y)], the value that wherein is positioned at the pixel of the capable y row of x is b
0(x, y), b
0(x, y) is 0 or 1, and the size of image is W
1* H
1The gray level image of character is G
c=[g
c(x, y)], the value that is positioned at the pixel of the capable y row of x is g
c(x, y), 0≤g
c(x, y)≤255; Bianry image B with character picture
0With gray level image G
cMerge, get the image G=[g (x, y) after the fusion], the value that is positioned at the pixel of the capable y row of x is g (x, y), 0≤g (x, y)≤255;
Step 2: extracting fused images G=[g (x, y)] feature before, carry out first fused images G=[g (x, y)] the position and the normalized of size; The input picture that image normalization is processed is G=[g (x, y)], the output image after the normalization is F=[f (x ', y ')], its size is respectively W
1* H
1And W
2* H
2Input picture G=[g (x, y)] pixel that is positioned at the capable y of x row will be mapped to F=[f (x ', y ')] being positioned at the pixel of x ' row y ' row, the coordinate by input picture and output image shines upon to realize image normalization:
One dimension coordinate is mapped as:
Step 3: the feature of extracting the histogram of gradients of normalized image based on histogram of gradients;
Step 4: utilize principal component analysis and linear discriminant analysis that the feature of the histogram of gradients of normalized image is carried out dimension-reduction treatment, obtain the transformation matrix of Feature Dimension Reduction;
Step 5: set up character feature template library, read the transformation matrix of Feature Dimension Reduction and character is identified.
Beneficial effect of the present invention: the invention is characterized in the bianry image of character picture and the fused images of gray level image are processed that carry out character recognition, it may further comprise the steps: the fusion of (1) bianry image and gray level image; (2) normalization of image; (3) based on the feature extraction of histogram of gradients; (4) Feature Dimension Reduction; (5) classifier design and character recognition.The present invention has overcome based on the bianry image of character or the shortcoming that can not identify simultaneously the degraded character image and comprise the character picture of complex background based on traditional character recognition technologies of the gray level image of character.The technical field of application of the present invention comprises the character recognition in the video, the character recognition in ID Card Image, license plate, the natural scene image
Description of drawings
Fig. 1 is character recognition system process flow diagram of the present invention;
Fig. 2 is the framework synoptic diagram of the fusion of bianry image of the present invention and gray level image;
Fig. 3 is the normalized framework synoptic diagram of image of the present invention;
Fig. 4 is the framework synoptic diagram that the present invention is based on the feature extraction of histogram of gradients;
Fig. 5 is the configuration diagram that the present invention asks the transformation matrix of Feature Dimension Reduction;
Fig. 6 is classifier design of the present invention and character recognition configuration diagram;
Fig. 7 is Sobel gradient operator template;
Fig. 8 is L reference direction example, left side L=4, the right L=8;
Fig. 9 is that gradient is decomposed example;
Figure 10 be calculating pixel and center, rectangular area in the horizontal direction with vertical direction on distance examples.
Embodiment
Describe each related detailed problem in the technical solution of the present invention in detail below in conjunction with accompanying drawing.Be to be noted that described embodiment only is intended to be convenient to the understanding of the present invention, and it is not played any restriction effect.
As shown in Figure 1, character recognition system process flow diagram of the present invention, recognizer can be divided into two parts: training system and recognition system.Training system merges its bianry image and gray level image to each character training sample, and fused images is carried out the normalization of size and position, extracts the feature of histogram of gradients; The feature that utilization is extracted from training sample is found the solution the transformation matrix that carries out Feature Dimension Reduction, obtains the character recognition storehouse.In recognition system, merge bianry image and the gray level image of character to be identified, fused images is carried out the normalization of size and position, extract the feature of histogram of gradients, the transformation matrix that utilizes training system to obtain carries out dimensionality reduction to feature, then sends into recognizer, obtains recognition result.
Fusion character bianry image and gray level image carry out the realization of character recognition system need to consider following several aspect:
1) realization of training system;
2) realization of recognition system.
The below describes in detail to these two aspects respectively.
The realization of 1 training system
1.1 see also the framework of the fusion of the bianry image shown in Fig. 2 and gray level image.
If the bianry image of the single character picture that obtains after the pre-service is B
0=[b
0(x, y)], the value that wherein is positioned at the pixel of the capable y row of x is b
0(x, y), b
0(x, y) is 0 or 1.The gray level image of character is G
c=[g
c(x, y)], the value that is positioned at the pixel of the capable y row of x is g
c(x, y), 0≤g
c(x, y)≤255.Bianry image is B
0With gray level image G
cSize be W
1* H
1Carry out the bianry image B of character by following flow process
0Gray level image G with character
cFusion, the image G=[g (x, y) after obtaining merging], the value that is positioned at the pixel of the capable y of x row is g (x, y), 0≤g (x, y)≤255:
Ask a total head threshold value th, for example can be to gray level image g
c(x, y) utilizes traditional maximum variance between clusters (OSTU method) to try to achieve this total head threshold value th
Ostu, make th=a x th
Ostu, a is a constant.Utilize global threshold th that gray level image g (x, y) is carried out threshold processing and obtain bianry image B
g=[b
g(x, y)], b
g(x, y) is defined as:
Utilize two-value morphology to bianry image B
oCarry out condition expansion.If D is the image of 3 x 3, the pixel value of its each position is 1.To bianry image B
oCarrying out condition expansion is:
According to following formula to bianry image B
iRepeatedly carry out condition expansion, until B
I+1=B
iOr reach maximum iterations, establishing the bianry image that obtains at last is B=[b (x, y)].
Obtain the fused images G=[g (x, y) for single character recognition], g (x, y) is defined as:
1.2 see also the normalized framework of the image shown in Fig. 3.
Before the feature of extracting fused images, the normalized of the position of advanced line character image and size.The input picture of image normalization is G=[g (x, y)], the output image after the normalization is F=[f (x ', y ')], its size is respectively W
1* H
1And W
2* H
2Input picture G=[g (x, y)] pixel that is positioned at the capable y of x row will be mapped to F=[f (x ', y ')] being positioned at the pixel of x ' row y ' row, the coordinate by input picture and output image shines upon to realize image normalization:
One dimension coordinate is mapped as
Calculate fused images G=[g (x, y)] barycenter (x
c, y
c), barycenter is adjusted into normalized image F=[f (x ', y ')] center (W
2/ 2, H
2/ 2):
G wherein
x(x) and g
y(y) be respectively fused images G=[g (x, y)] in the vertical direction with horizontal direction on picture element density;
According to centroid position (x
c, y
c), computed image G=[g (x, y)] monolateral second moment
With
The housing that input picture is set according to the monolateral second moment that calculates is
With
For the x axle, find the solution quadratic function u (x)=ax
2+ bx+c is three points on the x axle
Be mapped as respectively (0,0.5,1), in like manner obtain the quadratic function u (y) of y axle three points on the y axle
Be mapped as respectively (0,0.5,1); Obtain input picture G=[g (x, y)] be positioned at the pixel of the capable y of x row and output image F=[f (x ', y ')] be positioned at the coordinate mapping function of the pixel of x ' row y ' row:
Determine input picture G=[g (x according to the coordinate mapping function, y)] with normalized image F=[f (x ', y ')] the coordinate mapping relations, the input picture gray-scale value is passed through bilinear interpolation, obtain normalized image F=[f (x ', y ')] value.
1.3 see also the framework based on the feature extraction of histogram of gradients shown in Fig. 4.
Two 3 * 3 templates utilizing the Sobel operator are computed image F=[f (x, y) respectively] in each locational gradient, two 3x3 templates of Sobel operator are as shown in Figure 7.For image F=[f (x, y)], it is tried to achieve by following formula along x axle and the axial first order derivative component of y respectively:
g
x(x,y)=f(x+1,y-1)+2f(x+1,y)+f(x+1,y+1)
-f(x-1,y-1)-2f(x-1,y)-f(x-1,y+1),
g
y(x,y)=f(x-1,y+1)+2f(x,y+1)+f(x+1,y+1)
-f(x-1,y-1)-2f(x,y-1)-f(x+1,y-1).
x=0,...,W
2-1,y=1,...,H
2-1;
Computed image F=[f (x, y)] gradient intensity mag (x, y) and the deflection of position (x, y)
(x, y) is respectively:
Define L reference direction, the situation of L=4 and L=8 as Fig. 8 and shown in.Utilize parallelogram law to be decomposed into from its two nearest reference direction, as shown in Figure 9 gradient.Be W with size after the normalization
2* H
2Image F=[f (x, y)] be divided into R * R mutually disjoint rectangular area, set up the gradient orientation histogram of L dimension for each rectangular area.Image F=[f (x, y)] in the gradient orientation histogram of nearest 4 rectangular areas of gradient pair and this pixel of each pixel contribution is arranged.Be 4 rectangular areas (from top to bottom, from left to right being numbered respectively 1,2,3 and 4) that pixel is nearest with it as shown in figure 10, wherein each little rectangle frame represents a pixel, and 4 * 4 little rectangle frames form a large rectangle zone.In the horizontal direction, the distance at pixel and center, rectangular area is respectively d
HlAnd d
HrIn the vertical direction, the distance at pixel and center, rectangular area is respectively d
VtAnd d
VbIf the intensity of the component of pixel gradient on the l direction is g
l, then the gradient of this pixel is to the 1st, 2, and the contribution margin of the l of the gradient orientation histogram of 3 and 4 rectangular areas dimension is respectively g
l* d
Hr* d
Vb/ ((d
Hl+ d
Hr) * (d
Vt+ d
Vb)), g
l* d
Hl* d
Vb/ ((d
Hl+ d
Hr) * (d
Vt+ d
Vb)), g
l* d
Hr* d
Vt/ ((d
Hl+ d
Hr) * (d
Vt+ d
Vb)) and g
l* d
Hl* d
Vt/ ((d
Hl+ d
Hr) * (d
Vt+ d
Vb)).Utilize this method to calculate the gradient of each pixel to the contribution of the gradient orientation histogram of rectangular area adjacent thereto, try to achieve each rectangular area gradient orientation histogram, obtained at last the R * R of character picture * L dimensional feature.
1.4 see also the framework of the transformation matrix of asking Feature Dimension Reduction shown in Fig. 5:
1.4.1 principal component analysis (PCA)
The high dimensional feature vector comprises the feature that is mutually related, and it is large that it is processed operand, utilizes principal component analysis that the high dimensional feature vector is carried out principal component analysis (PCA), finds the solution PCA dimensionality reduction matrix P
PCAIf the character feature that extracts from n training sample is x
i, i=1 ..., n, x
iDimension m=R * R * L; The Scatter Matrix of training sample character feature is:
Scatter Matrix is carried out Eigenvalues Decomposition is:
∑=UΛU
T
U=[u wherein
1, u
2..., u
m] be orthogonal matrix, Λ=diag (λ
1, λ
2..., λ
m) be diagonal matrix, λ
1〉=λ
2〉=... 〉=λ
mBe eigenwert.If will preserve the energy of r% behind the principal component analysis PCA dimensionality reduction, then the principal direction number l of principal component analysis preservation is
The transformation matrix that principal component analysis obtains is P
PCA=[u
1, u
2..., u
l], to character feature x
i, carry out the l dimension character feature z after dimensionality reduction obtains dimensionality reduction
i=(P
PCA)
Tx
i, i=1 ..., n, (P
PCA)
TExpression P
PCATransposed matrix;
1.4.2 the character feature behind the training sample dimensionality reduction is carried out linear discriminant analysis (LDA), finds the solution transformation matrix W:
If character class number to be identified in the recognition system is C, the i class comprises n
iIndividual training sample.Calculate i class character sample characteristic mean μ
iWith all sample characteristics average μ:
Scatter Matrix S between compute classes
bWith Scatter Matrix S in the class
w:
A transformation matrix W is sought in linear discriminant analysis so that after the conversion between class dispersion as far as possible large, dispersion is as far as possible little in the class simultaneously, utilizes the maximization criterion
Represent.LDA can solve by finding the solution the generalized eigenvector problem:
S
bw=λS
ww
If vectorial w
1..., w
d..., w
lBe the solution of generalized eigenvector problem, the generalized eigenvalue λ of their correspondences
1〉=... 〉=λ
d〉=... 〉=λ
l, the solution proper vector of d generalized eigenvector problem forms W, i.e. W=[w before selecting
1..., w
d].
1.5 see also the classifier design shown in Fig. 6 and character recognition framework:
Utilize transformation matrix W to i character type characteristic mean μ
iCarry out dimensionality reduction, and the feature behind the normalization dimensionality reduction
Preserve transformation matrix P=WP
PCA, the coding of each character type and characteristic of correspondence μ thereof
iIn the File of identification storehouse.
(5.2) character recognition
The realization of 2 recognition systems
From the File of character recognition storehouse, read transformation matrix P, the coding of each character type and characteristic of correspondence μ thereof
iBianry image and gray level image to each character to be identified merge, and the image after merging is carried out normalization, carry out the multidimensional characteristic a that feature extraction obtains character picture.Utilize transformation matrix P that the multidimensional characteristic a of character picture is carried out feature b=P after Feature Dimension Reduction obtains dimensionality reduction
TA, P
TTransposed matrix for transformation matrix P.Feature normalization behind the dimensionality reduction is obtained
Normalization center vector { the μ of order computation b and each character type
i}
1≤i≤CCosine distance { d
i}
1≤i≤C
d
i=1-y
Tμ
The minimum class of distance is the recognition result of character picture.
The above; only be the embodiment among the present invention; but protection scope of the present invention is not limited to this; anyly be familiar with the people of this technology in the disclosed technical scope of the present invention; can understand conversion or the replacement expected; all should be encompassed in of the present invention comprising within the scope, therefore, protection scope of the present invention should be as the criterion with the protection domain of claims.
Claims (5)
1. a method that merges the character recognition of bianry image and gray level image is characterized in that, the fused images of described bianry image and gray level image is processed, and carries out character recognition, and it may further comprise the steps:
Step 1: the bianry image of establishing the single character picture that obtains after the pre-service is B
0=[b
0(x, y)], the value that wherein is positioned at the pixel of the capable y row of x is b
0(x, y), b
0(x, y) is 0 or 1, and the size of image is W
1* H
1The gray level image of character is G
c=[g
c(x, y)], the value that is positioned at the pixel of the capable y row of x is g
c(x, y), 0≤g
c(x, y)≤255; Bianry image B with character picture
0With gray level image G
cMerge the image G=[g (x, y) after obtaining merging], the value that is positioned at the pixel of the capable y row of x is g (x, y), 0≤g (x, y)≤255;
Step 2: extracting fused images G=[g (x, y)] feature before, carry out first fused images G=[g (x, y)] the position and the normalized of size; The input picture that image normalization is processed is G=[g (x, y)], the output image after the normalization is F=[f (x ', y ')], its size is respectively W
1* H
1And W
2* H
2Input picture G=[g (x, y)] pixel that is positioned at the capable y of x row will be mapped to F=[f (x ', y ')] being positioned at the pixel of x ' row y ' row, the coordinate by input picture and output image shines upon to realize image normalization:
One dimension coordinate is mapped as:
Step 3: the feature of extracting the histogram of gradients of normalized image based on histogram of gradients;
Step 4: utilize principal component analysis and linear discriminant analysis that the feature of the histogram of gradients of normalized image is carried out dimension-reduction treatment, obtain the transformation matrix of Feature Dimension Reduction;
Step 5: set up character feature template library, read the transformation matrix of Feature Dimension Reduction and character is identified;
The fusion of described bianry image and gray level image comprises:
Step 11: to gray level image G
c=[g
c(x, y)] pixel point value g
c(x, y) utilizes traditional maximum variance between clusters to try to achieve threshold value th
Ostu, ask a global threshold th, make th=a * th
Ostu, a is a constant; Utilize total head threshold value th to gray level image G
c=[g
c(x, y)] pixel point value g
c(x, y) carries out threshold processing, obtains bianry image B
g=[b
g(x, y)], the pixel point value b of bianry image
g(x, y) is defined as:
Step 12: utilize two-value morphology to bianry image B
oCarry out condition expansion, establishing D is the image of a 3x3, and the pixel value of its each position is 1; To bianry image B
oCarrying out condition expansion is:
According to following formula to bianry image B
iRepeatedly carry out condition expansion, until B
I+1=B
iOr reach maximum iterations, establishing the bianry image that obtains at last is B=[b (x, y)];
Step 13: obtain the fused images G=[g (x, y) for single character recognition], g (x, y) is defined as:
2. the method for the character recognition of described fusion bianry image and gray level image according to claim 1 is characterized in that the normalization of image comprises:
Step 21: calculate fused images G=[g (x, y)] barycenter (x
c, y
c), barycenter is adjusted into normalized image F=[f (x ', y ')] center (W
2/ 2, H
2/ 2):
G wherein
x(x) and g
y(y) be respectively fused images G=[g (x, y)] in the vertical direction with horizontal direction on picture element density;
Step 22: according to centroid position (x
c, y
c), computed image G=[g (x, y)] monolateral second moment
With
Step 23: the housing that input picture is set according to the monolateral second moment that calculates is:
For the x axle, find the solution quadratic function u (x)=ax
2+ bx+c is three points on the x axle
Be mapped as respectively (0,0.5,1), in like manner obtain the quadratic function u (y) of y axle three points on the y axle
Be mapped as respectively (0,0.5,1); Obtain input picture G=[g (x, y)] be positioned at the pixel of the capable y of x row and output image F=[f (x ', y ')] be positioned at the coordinate mapping function of the pixel of x ' row y ' row:
Step 24: finally obtain normalized image F=[f (x ', y ') by bilinear interpolation] value.
3. the method for the character recognition of described fusion bianry image and gray level image according to claim 1 is characterized in that, the step of the feature of the described histogram of gradients of extracting normalized image based on histogram of gradients comprises:
Step 31: two 3 * 3 templates utilizing the Sobel operator are computed image F=[f (x, y) respectively] in each locational gradient; For image F=[f (x, y)], it is tried to achieve by following formula along x axle and the axial first order derivative component of y respectively:
g
x(x,y)=f(x+1,y-1)+2f(x+1,y)+f(x+1,y+1)
-f(x-1,y-1)-2f(x-1,y)-f(x-1,y+1),
g
y(x,y)=f(x-1,y+1)+2f(x,y+1)+f(x+1,y+1),
-f(x-1,y-1)-2f(x,y-1)-f(x+1,y-1).
x=0,...,W
2-1,y=0,...,H
2-1;
Step 32: image F=[f (x, y)] gradient intensity mag (x, y) and the deflection of position (x, y)
Be respectively:
Step 33: defining L reference direction, utilize parallelogram law to be decomposed into from its two nearest reference direction gradient, is W with size after the normalization
2* H
2Image F=[f (x, y)] be divided into R * R mutually disjoint rectangular area, set up the gradient orientation histogram of L dimension for each rectangular area; Image F=[f (x, y)] in the gradient orientation histogram of nearest 4 rectangular areas of gradient pair and this pixel of each pixel contribution is arranged; Calculate the gradient of each pixel to the contribution of the gradient orientation histogram of rectangular area adjacent thereto, try to achieve each rectangular area gradient orientation histogram, obtained at last the R * R of character picture * L dimensional feature.
4. the method for the character recognition of described fusion bianry image and gray level image according to claim 1 is characterized in that the described step of utilizing principal component analysis and linear discriminant analysis that the feature of the histogram of gradients of normalized image is carried out dimension-reduction treatment comprises:
Step 41: the high dimensional feature vector is carried out principal component analysis (PCA), find the solution PCA dimensionality reduction matrix P
PCA:
The high dimensional feature vector comprises the feature that is mutually related, and it is large that it is processed operand, utilizes principal component analysis that the high dimensional feature vector is carried out principal component analysis (PCA), finds the solution PCA dimensionality reduction matrix P
PCAIf the character feature that extracts from n training sample is x
i, i=1 ..., n, x
iDimension m=R * R * L; The Scatter Matrix of training sample character feature is:
Scatter Matrix is carried out Eigenvalues Decomposition is:
∑=UΛU
T
U=[u wherein
1, u
2..., u
m] be orthogonal matrix, Λ=diag (λ
1, λ
2..., λ
m) be diagonal matrix, λ
1〉=λ
2〉=... 〉=λ
mBe eigenwert, establish the energy that will preserve r% behind the principal component analysis PCA dimensionality reduction, then the principal direction number l that preserves of principal component analysis is:
The transformation matrix that principal component analysis obtains is P
PCA=[u
1, u
2..., u
l], to character feature x
i, carry out the l dimension character feature z after dimensionality reduction obtains dimensionality reduction
i=(P
PCA)
Tx
i, i=1 ..., n, (P
PCA)
TExpression P
PCATransposed matrix;
Step 42: the character feature behind the training sample dimensionality reduction is carried out linear discriminant analysis, find the solution transformation matrix W:
If character class number to be identified in the recognition system is C, the i class comprises n
iIndividual training sample; Calculate i class character sample characteristic mean μ
iWith all sample characteristics average μ:
Scatter Matrix S between compute classes
bWith Scatter Matrix S in the class
w:
A transformation matrix W is sought in linear discriminant analysis so that after the conversion between class dispersion as far as possible large, dispersion is as far as possible little in the class simultaneously, utilizes the maximization criterion
Solve linear discriminant analysis by finding the solution the generalized eigenvector problem:
S
bw=λS
ww
If vectorial w
1..., w
d..., w
lBe the solution of generalized eigenvector problem, the generalized eigenvalue λ of their correspondences
1〉=... 〉=λ
d〉=... 〉=λ
l, the solution proper vector of d generalized eigenvector problem forms W, i.e. W=[w before selecting
1..., w
d].
5. the method for the character recognition of described fusion bianry image and gray level image according to claim 4 is characterized in that described character recognition comprises:
Step 51: design category device
Utilize transformation matrix W to i character type characteristic mean μ
iCarry out dimensionality reduction, and the feature behind the normalization dimensionality reduction
Preserve transformation matrix P=WP
PCA, the coding of each character type and corresponding character type feature thereof
In the file in the character recognition storehouse;
Step 52: character recognition
Read coding and the corresponding character type feature thereof of transformation matrix P, each character type in the file from the character recognition storehouse
Bianry image and gray level image to each character to be identified merge, and the image after merging is carried out normalization, carry out the multidimensional characteristic a that feature extraction obtains character picture; Utilize transformation matrix P that the multidimensional characteristic a of character picture is carried out feature b=P after Feature Dimension Reduction obtains dimensionality reduction
TA, P
TTransposed matrix for transformation matrix P;
Order computation character normalization feature
Normalization center vector with each character type
Cosine distance { d
i}
1≤i≤C
The minimum class of distance is the recognition result of character picture.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 200810239331 CN101751565B (en) | 2008-12-10 | 2008-12-10 | Method for character identification through fusing binary image and gray level image |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 200810239331 CN101751565B (en) | 2008-12-10 | 2008-12-10 | Method for character identification through fusing binary image and gray level image |
Publications (2)
Publication Number | Publication Date |
---|---|
CN101751565A CN101751565A (en) | 2010-06-23 |
CN101751565B true CN101751565B (en) | 2013-01-02 |
Family
ID=42478527
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN 200810239331 Expired - Fee Related CN101751565B (en) | 2008-12-10 | 2008-12-10 | Method for character identification through fusing binary image and gray level image |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN101751565B (en) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102011075275A1 (en) * | 2011-05-04 | 2012-11-08 | Bundesdruckerei Gmbh | Method and device for recognizing a character |
CN102750530B (en) * | 2012-05-31 | 2014-11-26 | 贺江涛 | Character recognition method and device |
CN103854020B (en) * | 2012-11-29 | 2018-11-30 | 捷讯平和(北京)科技发展有限公司 | Character recognition method and device |
CN103679208A (en) * | 2013-11-27 | 2014-03-26 | 北京中科模识科技有限公司 | Broadcast and television caption recognition based automatic training data generation and deep learning method |
CN106257495A (en) * | 2015-06-19 | 2016-12-28 | 阿里巴巴集团控股有限公司 | A kind of digit recognition method and device |
CN106203434B (en) * | 2016-07-08 | 2019-07-19 | 中国科学院自动化研究所 | File and picture binary coding method based on stroke structure symmetry |
CN108319958A (en) * | 2018-03-16 | 2018-07-24 | 福州大学 | A kind of matched driving license of feature based fusion detects and recognition methods |
CN108830138B (en) * | 2018-04-26 | 2021-05-07 | 平安科技(深圳)有限公司 | Livestock identification method, device and storage medium |
CN109520706B (en) * | 2018-11-21 | 2020-10-09 | 云南师范大学 | Screw hole coordinate extraction method of automobile fuse box |
CN109919253A (en) * | 2019-03-27 | 2019-06-21 | 北京爱数智慧科技有限公司 | Character identifying method, device, equipment and computer-readable medium |
CN111583217A (en) * | 2020-04-30 | 2020-08-25 | 深圳开立生物医疗科技股份有限公司 | Tumor ablation curative effect prediction method, device, equipment and computer medium |
CN112200247B (en) * | 2020-10-12 | 2021-07-02 | 西安泽塔云科技股份有限公司 | Image processing system and method based on multi-dimensional image mapping |
-
2008
- 2008-12-10 CN CN 200810239331 patent/CN101751565B/en not_active Expired - Fee Related
Also Published As
Publication number | Publication date |
---|---|
CN101751565A (en) | 2010-06-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN101751565B (en) | Method for character identification through fusing binary image and gray level image | |
Chen et al. | Accurate and efficient traffic sign detection using discriminative adaboost and support vector regression | |
Chen et al. | Fast human detection using a novel boosted cascading structure with meta stages | |
Mahadevan et al. | Saliency-based discriminant tracking | |
Fukui et al. | The kernel orthogonal mutual subspace method and its application to 3D object recognition | |
CN103870811B (en) | A kind of front face Quick method for video monitoring | |
CN103198303B (en) | A kind of gender identification method based on facial image | |
Riggan et al. | Thermal to visible synthesis of face images using multiple regions | |
CN105023008A (en) | Visual saliency and multiple characteristics-based pedestrian re-recognition method | |
CN102663413A (en) | Multi-gesture and cross-age oriented face image authentication method | |
CN102609716A (en) | Pedestrian detecting method based on improved HOG feature and PCA (Principal Component Analysis) | |
CN102163281B (en) | Real-time human body detection method based on AdaBoost frame and colour of head | |
CN103413119A (en) | Single sample face recognition method based on face sparse descriptors | |
Redondo-Cabrera et al. | All together now: Simultaneous object detection and continuous pose estimation using a hough forest with probabilistic locally enhanced voting | |
CN105894018B (en) | Classification of Polarimetric SAR Image method based on depth multi-instance learning | |
CN103605993B (en) | Image-to-video face identification method based on distinguish analysis oriented to scenes | |
CN105138951B (en) | Human face portrait-photo array the method represented based on graph model | |
Kruthi et al. | Offline signature verification using support vector machine | |
Li et al. | Multi-view face detection using support vector machines and eigenspace modelling | |
CN104008404B (en) | Pedestrian detection method and system based on significant histogram features | |
CN111259780A (en) | Single-sample face recognition method based on block linear reconstruction discriminant analysis | |
CN114612980A (en) | Deformed face detection based on multi-azimuth fusion attention | |
Luzardo et al. | Head pose estimation for sign language video | |
Hanani et al. | Evaluating new variants of motion interchange patterns | |
Zhu et al. | Car detection based on multi-cues integration |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20130102 |