CN102855479A - Printed Chinese character recognition system - Google Patents
Printed Chinese character recognition system Download PDFInfo
- Publication number
- CN102855479A CN102855479A CN2012102574590A CN201210257459A CN102855479A CN 102855479 A CN102855479 A CN 102855479A CN 2012102574590 A CN2012102574590 A CN 2012102574590A CN 201210257459 A CN201210257459 A CN 201210257459A CN 102855479 A CN102855479 A CN 102855479A
- Authority
- CN
- China
- Prior art keywords
- chinese characters
- sample
- sigma
- image
- gray
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 claims abstract description 19
- 238000009499 grossing Methods 0.000 claims abstract description 11
- 238000013528 artificial neural network Methods 0.000 claims abstract description 9
- 238000012545 processing Methods 0.000 claims abstract description 5
- 238000006243 chemical reaction Methods 0.000 claims abstract description 4
- 238000005192 partition Methods 0.000 claims abstract description 4
- 238000010187 selection method Methods 0.000 claims abstract description 4
- 230000002708 enhancing effect Effects 0.000 claims abstract description 3
- 230000011218 segmentation Effects 0.000 claims description 14
- 230000002776 aggregation Effects 0.000 claims description 12
- 238000004220 aggregation Methods 0.000 claims description 12
- 230000005484 gravity Effects 0.000 claims description 9
- 230000004520 agglutination Effects 0.000 claims description 6
- 238000012549 training Methods 0.000 claims description 6
- 241000282414 Homo sapiens Species 0.000 claims description 3
- 238000011835 investigation Methods 0.000 claims description 3
- 235000019557 luminance Nutrition 0.000 claims description 3
- 230000006870 function Effects 0.000 claims description 2
- 238000000638 solvent extraction Methods 0.000 abstract 3
- 241001270131 Agaricus moelleri Species 0.000 abstract 2
- 238000013473 artificial intelligence Methods 0.000 description 5
- 239000013598 vector Substances 0.000 description 2
- 240000002627 Cordeauxia edulis Species 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000003706 image smoothing Methods 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Landscapes
- Character Input (AREA)
Abstract
The invention relates to a printed Chinese character recognition system. By adopting a novel algorithm in the artificial intelligent field, the Chinese character recognition quality is effectively improved. The system comprises a scanning input for converting Chinese characters printed on a paper surface into electric signals through photoelectric conversion equipment to form a plurality of gray scales of digital signals and inputting the signals into a computer for processing, a fuzzy enhancement and cluster partitioning module for enhancing and smoothing fuzzy and performing optimal partition on multiple gray scales, an image data binarization module for carrying out binaryzation on an image subjected to smoothing, enhancement and multi-gray-scale cluster partitioning by an overall threshold value selection method, counting a gray scale histogram of the image and partitioning the histogram into two groups at a certain threshold value, and determining the threshold value when a variance between the two divided groups is the largest, and a Chinese character recognition module for performing rough classification on a sample through a control network (CN) by a parallel neural network method, and performing fine classification on the rough classifications through a recognition network (RN) to recognize Chinese characters.
Description
Technical Field
The invention belongs to the technical field of pattern recognition, and particularly relates to a printed Chinese character recognition system adopting a new algorithm in the field of artificial intelligence.
Background
The study of automatic abstracts is an important area of natural language processing. However, all automatic abstract models currently use the built-in representation of characters as system input, and obviously have a considerable gap from the practical goal, because a large amount of literature still exists in the traditional paper printing form. Therefore, a parallel neural network method for identifying the Chinese characters in the printed form is provided and experimental simulation is carried out. In the test of 2500 Chinese characters, the recognition rate is 97%, the false recognition rate is 1% and the rejection rate is 2%. It should be noted that although our system is designed for print Chinese characters, the basic implementation principles and methods are equally applicable to recognition of print English, as well as symbols such as handwritten numbers, English, and Chinese characters.
The system for identifying the Chinese characters in the printed form mainly comprises the following steps: scanner input → fuzzy enhancement and cluster segmentation → image data binarization → Chinese character matching by parallel neural network.
Disclosure of Invention
The invention aims to provide a tunable band-pass filter system which outputs one or more accurate central frequency points, bandwidth control, out-of-band signal suppression and signal gain functions to an input radio frequency digital signal.
The technical scheme for realizing the purpose of the invention is as follows: a printed Chinese character recognition system adopts a new algorithm in the field of artificial intelligence, effectively improves the quality of Chinese character recognition, and comprises:
scanning input, converting the Chinese characters printed on the paper surface into electric signals through a photoelectric conversion device to form digital signals with multiple gray levels, and inputting the digital signals into a computer for processing;
the fuzzy enhancement and clustering segmentation module comprises fuzzy enhancement, smoothing and multi-gray-level optimal segmentation;
the image data binarization module is used for carrying out binarization by adopting an integral threshold value selection method after smoothing and enhancing the image and clustering segmentation of multiple gray levels, firstly counting a gray level histogram of the image, then segmenting the histogram into two groups at a certain threshold value, and determining a threshold value when the variance between the two groups is maximum;
the system adopts a parallel neural network method, roughly classifies samples through a control network CN, and finely classifies various rough classes through an identification network RN, so that Chinese characters are identified.
As a further improvement of the present invention, the fuzzy enhancement and smoothing module adopts a model proposed by s.k.pal, etc., from input to output, and two times uses a fuzzy contrast enhancement operator for fuzzy enhancement processing, and one smoothing operation therebetween, to prepare for the next enhancement.
As a further improvement of the invention, the cluster segmentation performs the following steps:
(1) defining sample interval, taking K, C and R (K < S)3 values, and taking K samples as aggregation points;
(2) calculating the distance between every two of the K aggregation points, if the minimum distance is less than C, merging the corresponding two points, and taking the arithmetic mean of the two points as a new aggregation point; repeating the steps until all the intervals are more than or equal to C;
(3) when another (S-K) samples are examined, the distance between the sample and all the agglutination points is calculated every time one sample is examined, and if the minimum distance is greater than R, the sample is used as a new agglutination point; if the minimum distance < R falls within the class of nearest agglutinated points, then recalculating the center of gravity of this class and using this as a new agglutinated point, and if the distance between agglutinated points > C, examining the next sample; otherwise, step 2 is carried out, after the aggregation points are combined, the next sample is inspected until the classification is finished;
(4) inspecting all the samples one by one, clustering according to the step 3, and if the final classification result is the same as the original result, not calculating the gravity center; otherwise, the gravity center is recalculated, and if the classification result is the same as the original result after one-time investigation, the clustering is finished; otherwise, repeating the step 4 until the classification result is completely the same as the previous classification result;
(5) the above clustering process ends with the state of class number 3, so that the class numbers m can be generated1,m2,…,me(me=3), and the respective target functional can be determined in accordance with equation (1)
(6) Decision J = min { J }1(m1,C),J2(m2,C),…,Je(meThe partition corresponding to C) is the best;
(7) setting thresholds of all levels, setting class CiU luminances: r isi1,ri2,…,riuCorresponding density is P1,P2,…,PiuThen threshold value ri **Is composed of
Wherein m is*Is the number of classes that is optimally partitioned.
As a further improvement of the invention, the image data binarization performs the following steps:
(1) counting the total number of pixels in the imageAnd the probability of each gray level p (i) = N (i)/N;
(2) dividing the 1, …, M level of image gray into two groups C by a certain gray level k0= 1 → K and C1= { (K +1) → M }, the probability of each group occurring being respectively
C0Average value of (2)
C1Average value of (2)
Where μ = ∑ ipiIs the average of the gray levels of the overall image;is the average value of the gray scale when the image threshold is k;
(3) changing the k value from 1 to M to obtain the above formula delta2(k) Is the maximum k, the value k at this time*Is the threshold value sought
As a further improvement of the invention, a new additional learning algorithm is adopted, and the learning of all Chinese characters is finally realized by simulating the character learning process of human beings, namely, learning part of Chinese characters first and then gradually learning the rest Chinese characters.
As a further refinement of the invention, the supplementary learning algorithm performs the following steps:
(1) let the original PNN recognize all samples in V, and set the correctly recognized sample set as V0;
(2) The samples in the set V-V0 are divided as follows: VT is V-V0Sample subset in PNN learned Chinese characters; VF being V-V0Sample subsets not in the PNN learned Chinese characters;
(3) to VTIf it corresponds to an identified network PNiCan correctly identify it without the need of going to RNiCarry out training whether or notThen these samples are added to the RNiOriginal sample concentration, retraining RNi;
(4) To VFTraining new recognition networks RN ', RN', … and RN ', and setting the divided Chinese character subsets as U1', U2 ', … and Uq';
(5) and retraining the control network CN by using the U.V sample set.
The invention mainly comprises four steps of scanning input, fuzzy enhancement and clustering segmentation, image data binarization, Chinese character matching through a parallel neural network and the like, and aiming at the characteristics of Chinese character recognition, new algorithms in the field of artificial intelligence are adopted, and the system can realize the recognition of the printed Chinese characters.
The invention has the beneficial effects that:
(1) the method adopts new algorithms in the field of artificial intelligence, thereby effectively improving the quality of Chinese character recognition;
(2) the designed network can store training mode vectors with strong correlation; the memory capacity of the network is not limited and even all 2's can be storedNTraining mode vectors, and meanwhile, the connection weight of the network only takes 1, 0 or-1, so that the network is easy to realize optically;
(3) the system adopts a maximum variance threshold setting method, and the method can obtain more satisfactory results regardless of whether the image histogram has obvious double peaks or not.
Drawings
FIG. 1 is a schematic structural diagram of a printed Chinese character recognition system in embodiment 1 of the present invention.
Detailed Description
The following further description is made in conjunction with the accompanying drawings and examples.
As shown in FIG. 1, a system for recognizing Chinese characters in printed form, which adopts a new algorithm in the field of artificial intelligence to effectively improve the quality of Chinese character recognition, comprises:
scanning input, converting the Chinese characters printed on the paper surface into electric signals through a photoelectric conversion device to form digital signals with multiple gray levels, and inputting the digital signals into a computer for processing; the fuzzy enhancement and clustering segmentation module comprises fuzzy enhancement, smoothing and multi-gray-level optimal segmentation; after image data binarization, image smoothing and enhancement and multi-gray-level cluster segmentation, carrying out binarization by adopting an integral threshold value selection method, firstly counting a gray histogram of the image, then segmenting the histogram into two groups at a certain threshold value, and determining a threshold value when the variance between the two groups is maximum; the system adopts a parallel neural network method, roughly classifies samples through a control network CN, finely classifies each rough class through an identification network RN, and accordingly identifies the Chinese characters.
The clustering segmentation performs the following steps:
(1) defining sample interval, taking K, C and R (K < S)3 values, and taking K samples as aggregation points;
(2) calculating the distance between every two of the K aggregation points, if the minimum distance is less than C, merging the corresponding two points, and taking the arithmetic mean of the two points as a new aggregation point; repeating the steps until all the intervals are more than or equal to C;
(3) when another (S-K) samples are examined, the distance between the sample and all the agglutination points is calculated every time one sample is examined, and if the minimum distance is greater than R, the sample is used as a new agglutination point; if the minimum distance < R falls within the class of nearest agglutinated points, then recalculating the center of gravity of this class and using this as a new agglutinated point, and if the distance between agglutinated points > C, examining the next sample; otherwise, step 2 is carried out, after the aggregation points are combined, the next sample is inspected until the classification is finished;
(4) inspecting all the samples one by one, clustering according to the step 3, and if the final classification result is the same as the original result, not calculating the gravity center; otherwise, the gravity center is recalculated, and if the classification result is the same as the original result after one-time investigation, the clustering is finished; otherwise, repeating the step 4 until the classification result is completely the same as the previous classification result;
(5) the above clustering process ends with the state of class number 3, so that the class numbers m can be generated1,m2,…,me(me=3), and the respective target functional can be determined in accordance with equation (1)
(6) Decision J = min { J }1(m1,C),J2(m2,C),…,Je(meThe partition corresponding to C) is the best;
(7) setting thresholds of all levels, setting class CiU luminances: r isi1,ri2,…,riuCorresponding density is P1,P2,…,PiuThen threshold value ri **Is composed of
Wherein m is*Is the number of classes that is optimally partitioned.
The image data is binarized to perform the following steps:
(1) counting the total number of pixels in the imageAnd the probability of each gray level p (i) = N (i)/N;
(2) dividing the 1, …, M level of image gray into two groups C by a certain gray level k0= 1 → K and C1= { (K +1) → M }, the probability of each group occurring being respectively
C0Average value of (2)
C1Average value of (2)
Where μ = ∑ ipiIs the average of the gray levels of the overall image;is the average value of the gray scale when the image threshold is k;
(3) changing the k value from 1 to M to obtain the above formula delta2(k) Is the maximum k, the value k at this time*Is the threshold value sought
The method for recognizing Chinese characters in printed form adopts a parallel neural network method, PNN is specially provided for recognizing a huge sample set of Chinese characters, a control network CN is used for roughly classifying samples, and a recognition network RN is used for finely classifying coarse classes so as to recognize Chinese characters. Both CN and RN adopt Hopfield network, and use the 'outer product equal criterion' as learning rule of associative memory instead of the common outer product and criterion or Hebb learning rule. The system also adopts a new additional learning algorithm which simulates the character learning process of human beings, namely, learning part of Chinese characters first, then gradually learning the rest Chinese characters, and finally realizing the learning of all Chinese characters.
Claims (8)
1. A system for identifying printed chinese characters, the system comprising:
scanning input, converting the Chinese characters printed on the paper surface into electric signals through a photoelectric conversion device to form digital signals with multiple gray levels, and inputting the digital signals into a computer for processing;
the fuzzy enhancement and clustering segmentation module comprises fuzzy enhancement, smoothing and multi-gray-level optimal segmentation;
the image data binarization module is used for carrying out binarization by adopting an integral threshold value selection method after smoothing and enhancing the image and clustering segmentation of multiple gray levels, firstly counting a gray level histogram of the image, then segmenting the histogram into two groups at a certain threshold value, and determining a threshold value when the variance between the two groups is maximum;
the system adopts a parallel neural network method, roughly classifies samples through a control network CN, and finely classifies various rough classes through an identification network RN, so that Chinese characters are identified.
2. The system for identifying printed chinese characters of claim 1, wherein said fuzzy enhancement and smoothing module employs a model proposed by s.k.pal, etc. from input to output, applying fuzzy contrast enhancement operator twice for fuzzy enhancement, and smoothing one time in between, in preparation for the next enhancement.
3. The system for identifying printed chinese characters of claim 1, wherein an image enhanced histogram: having S brightness levels r1,r2,…,rsWith a corresponding probability density of P1,P2,…,PsThe system rewrites the objective function as:
4. The system for identifying printed Chinese characters as claimed in claim 1 or 3, wherein said clustering segmentation of S intensity clusters is performed by:
(1) defining sample interval, taking K, C and R (K < S)3 values, and taking K samples as aggregation points;
(2) calculating the distance between every two of the K aggregation points, if the minimum distance is less than C, merging the corresponding two points, and taking the arithmetic mean of the two points as a new aggregation point; repeating the steps until all the intervals are more than or equal to C;
(3) when another (S-K) samples are examined, the distance between the sample and all the agglutination points is calculated every time one sample is examined, and if the minimum distance is greater than R, the sample is used as a new agglutination point; if the minimum distance < R falls within the class of nearest agglutinated points, then recalculating the center of gravity of this class and using this as a new agglutinated point, and if the distance between agglutinated points > C, examining the next sample; otherwise, step 2 is carried out, after the aggregation points are combined, the next sample is inspected until the classification is finished;
(4) inspecting all the samples one by one, clustering according to the step 3, and if the final classification result is the same as the original result, not calculating the gravity center; otherwise, the gravity center is recalculated, and if the classification result is the same as the original result after one-time investigation, the clustering is finished; otherwise, repeating the step 4 until the classification result is completely the same as the previous classification result;
(5) the above clustering process ends with the state of class number 3, so that the class numbers m can be generated1,m2,…,me(me=3), and the respective target functional can be determined in accordance with equation (1)
(6) Decision J = min { J }1(m1,C),J2(m2,C),…,Je(meThe partition corresponding to C) is the best;
(7) setting thresholds of all levels, setting class CiU luminances: r isi1,ri2,…,riuCorresponding density is P1,P2,…,PiuThen threshold value ri **Is composed of
Wherein m is*Is the number of classes that is optimally partitioned.
5. The system for identifying printed Chinese characters as claimed in claim 1, wherein the binarization of image data employs a maximum variance threshold setting method, and comprises the steps of firstly counting a gray histogram of the image, then dividing the histogram into two groups at a certain threshold, and determining the threshold when the variance between the two groups is maximum, and performing the following steps:
(1) counting the total number of pixels in the imageAnd the probability of each gray level p (i) = N (i)/N;
(2) dividing the 1, …, M level of image gray into two groups C by a certain gray level k0= 1 → K and C1= { (K +1) → M }, the probability of each group occurring being respectively
C0Average value of (2)
C1Average value of (2)
Where μ = ∑ ipiIs the average of the gray levels of the overall image;is the average value of the gray scale when the image threshold is k;
(3) changing the k value from 1 to M to obtain the above formula delta2(k) Is the maximum k, the value k at this time*It is the desired threshold.
6. The system for identifying printed Chinese characters as recited in claim 1, wherein said Chinese character identification uses parallel neural network method to coarsely classify samples by control network CN, and uses identification network RN to finely classify coarse classes, thereby identifying Chinese characters, and CN and RN use Hopfield network.
7. The system for recognizing printed Chinese characters as claimed in claim 1 or 6, wherein a new additional learning algorithm is used to simulate the process of learning Chinese characters by human, i.e. learning some Chinese characters first, then learning the rest of Chinese characters gradually, and finally realizing the learning of all Chinese characters.
8. The system for recognition of typed chinese characters according to claim 1 or 7, wherein the additional learning algorithm performs the following steps:
(1) let the original PNN recognize all samples in V, and set the correctly recognized sample set as V0;
(2) The samples in the set V-V0 are divided as follows: VT is V-V0Sample subset in PNN learned Chinese characters; VF being V-V0Sample subsets not in the PNN learned Chinese characters;
(3) to VTIf it corresponds to an identified network PNiCan correctly identify it without the need of going to RNiTraining is performed, otherwise the samples are added to RNiOriginal sample concentration, retraining RNi;
(4) To VFTraining new recognition networks RN ', RN', … and RN ', and setting the divided Chinese character subsets as U1', U2 ', … and Uq';
(5) and retraining the control network CN by using the U.V sample set.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2012102574590A CN102855479A (en) | 2012-07-24 | 2012-07-24 | Printed Chinese character recognition system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2012102574590A CN102855479A (en) | 2012-07-24 | 2012-07-24 | Printed Chinese character recognition system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN102855479A true CN102855479A (en) | 2013-01-02 |
Family
ID=47402056
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2012102574590A Pending CN102855479A (en) | 2012-07-24 | 2012-07-24 | Printed Chinese character recognition system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102855479A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105938558A (en) * | 2015-03-06 | 2016-09-14 | 松下知识产权经营株式会社 | Learning method |
CN110728526A (en) * | 2019-08-19 | 2020-01-24 | 阿里巴巴集团控股有限公司 | Address recognition method, apparatus and computer readable medium |
-
2012
- 2012-07-24 CN CN2012102574590A patent/CN102855479A/en active Pending
Non-Patent Citations (4)
Title |
---|
何汉华: "《基于Hopfield的脱机手写数字识别理论及算法》", 《通信技术》 * |
张云杰等: "《改进的模糊增强算法》", 《辽宁工程技术大学学报(自然科学版)》 * |
王先旺等: "《智能神经网络系统原理在印刷体汉字识别中的应用》", 《四川大学学报(工程科学版)》 * |
王国胤等: "《汉字识别的并行神经网络方法》", 《模式识别与人工智能》 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105938558A (en) * | 2015-03-06 | 2016-09-14 | 松下知识产权经营株式会社 | Learning method |
CN105938558B (en) * | 2015-03-06 | 2021-02-09 | 松下知识产权经营株式会社 | Learning method |
CN110728526A (en) * | 2019-08-19 | 2020-01-24 | 阿里巴巴集团控股有限公司 | Address recognition method, apparatus and computer readable medium |
CN110728526B (en) * | 2019-08-19 | 2024-04-02 | 创新先进技术有限公司 | Address recognition method, device and computer readable medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113378632B (en) | Pseudo-label optimization-based unsupervised domain adaptive pedestrian re-identification method | |
CN107169504B (en) | A kind of hand-written character recognition method based on extension Non-linear Kernel residual error network | |
CN102842032B (en) | Method for recognizing pornography images on mobile Internet based on multi-mode combinational strategy | |
CN103514456B (en) | Image classification method and device based on compressed sensing multi-core learning | |
Wu et al. | End-to-end chromosome Karyotyping with data augmentation using GAN | |
CN109410184B (en) | Live broadcast pornographic image detection method based on dense confrontation network semi-supervised learning | |
Wang et al. | Radical-based Chinese character recognition via multi-labeled learning of deep residual networks | |
CN103077399B (en) | Based on the biological micro-image sorting technique of integrated cascade | |
CN112734775A (en) | Image annotation, image semantic segmentation and model training method and device | |
CN104504365A (en) | System and method for smiling face recognition in video sequence | |
CN108776774A (en) | A kind of human facial expression recognition method based on complexity categorization of perception algorithm | |
Burie et al. | ICFHR2016 competition on the analysis of handwritten text in images of balinese palm leaf manuscripts | |
CN105718889A (en) | Human face identity recognition method based on GB(2D)2PCANet depth convolution model | |
CN109522924A (en) | A kind of broad-leaf forest wood recognition method based on single photo | |
CN113011357A (en) | Depth fake face video positioning method based on space-time fusion | |
CN112990282B (en) | Classification method and device for fine-granularity small sample images | |
CN109086794B (en) | Driving behavior pattern recognition method based on T-LDA topic model | |
CN101364263A (en) | Method and system for detecting skin texture to image | |
CN109614869A (en) | A kind of pathological image classification method based on multi-scale compress rewards and punishments network | |
Bawane et al. | Object and character recognition using spiking neural network | |
CN103971106A (en) | Multi-view human facial image gender identification method and device | |
CN106874825A (en) | The training method of Face datection, detection method and device | |
CN104268587B (en) | False fingerprint detection method based on finger wave conversion and SVM | |
CN107067022B (en) | Method, device and equipment for establishing image classification model | |
CN115310491A (en) | Class-imbalance magnetic resonance whole brain data classification method based on deep learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C05 | Deemed withdrawal (patent law before 1993) | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20130102 |