CN102855479A - Printed Chinese character recognition system - Google Patents

Printed Chinese character recognition system Download PDF

Info

Publication number
CN102855479A
CN102855479A CN2012102574590A CN201210257459A CN102855479A CN 102855479 A CN102855479 A CN 102855479A CN 2012102574590 A CN2012102574590 A CN 2012102574590A CN 201210257459 A CN201210257459 A CN 201210257459A CN 102855479 A CN102855479 A CN 102855479A
Authority
CN
China
Prior art keywords
chinese characters
sample
sigma
image
gray
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2012102574590A
Other languages
Chinese (zh)
Inventor
陶军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
SUZHOU INDUSTRIAL PARK SEVEN STAR ELECTRONICS CO LTD
Original Assignee
SUZHOU INDUSTRIAL PARK SEVEN STAR ELECTRONICS CO LTD
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by SUZHOU INDUSTRIAL PARK SEVEN STAR ELECTRONICS CO LTD filed Critical SUZHOU INDUSTRIAL PARK SEVEN STAR ELECTRONICS CO LTD
Priority to CN2012102574590A priority Critical patent/CN102855479A/en
Publication of CN102855479A publication Critical patent/CN102855479A/en
Pending legal-status Critical Current

Links

Landscapes

  • Character Input (AREA)

Abstract

The invention relates to a printed Chinese character recognition system. By adopting a novel algorithm in the artificial intelligent field, the Chinese character recognition quality is effectively improved. The system comprises a scanning input for converting Chinese characters printed on a paper surface into electric signals through photoelectric conversion equipment to form a plurality of gray scales of digital signals and inputting the signals into a computer for processing, a fuzzy enhancement and cluster partitioning module for enhancing and smoothing fuzzy and performing optimal partition on multiple gray scales, an image data binarization module for carrying out binaryzation on an image subjected to smoothing, enhancement and multi-gray-scale cluster partitioning by an overall threshold value selection method, counting a gray scale histogram of the image and partitioning the histogram into two groups at a certain threshold value, and determining the threshold value when a variance between the two divided groups is the largest, and a Chinese character recognition module for performing rough classification on a sample through a control network (CN) by a parallel neural network method, and performing fine classification on the rough classifications through a recognition network (RN) to recognize Chinese characters.

Description

Printed Chinese character recognition system
Technical Field
The invention belongs to the technical field of pattern recognition, and particularly relates to a printed Chinese character recognition system adopting a new algorithm in the field of artificial intelligence.
Background
The study of automatic abstracts is an important area of natural language processing. However, all automatic abstract models currently use the built-in representation of characters as system input, and obviously have a considerable gap from the practical goal, because a large amount of literature still exists in the traditional paper printing form. Therefore, a parallel neural network method for identifying the Chinese characters in the printed form is provided and experimental simulation is carried out. In the test of 2500 Chinese characters, the recognition rate is 97%, the false recognition rate is 1% and the rejection rate is 2%. It should be noted that although our system is designed for print Chinese characters, the basic implementation principles and methods are equally applicable to recognition of print English, as well as symbols such as handwritten numbers, English, and Chinese characters.
The system for identifying the Chinese characters in the printed form mainly comprises the following steps: scanner input → fuzzy enhancement and cluster segmentation → image data binarization → Chinese character matching by parallel neural network.
Disclosure of Invention
The invention aims to provide a tunable band-pass filter system which outputs one or more accurate central frequency points, bandwidth control, out-of-band signal suppression and signal gain functions to an input radio frequency digital signal.
The technical scheme for realizing the purpose of the invention is as follows: a printed Chinese character recognition system adopts a new algorithm in the field of artificial intelligence, effectively improves the quality of Chinese character recognition, and comprises:
scanning input, converting the Chinese characters printed on the paper surface into electric signals through a photoelectric conversion device to form digital signals with multiple gray levels, and inputting the digital signals into a computer for processing;
the fuzzy enhancement and clustering segmentation module comprises fuzzy enhancement, smoothing and multi-gray-level optimal segmentation;
the image data binarization module is used for carrying out binarization by adopting an integral threshold value selection method after smoothing and enhancing the image and clustering segmentation of multiple gray levels, firstly counting a gray level histogram of the image, then segmenting the histogram into two groups at a certain threshold value, and determining a threshold value when the variance between the two groups is maximum;
the system adopts a parallel neural network method, roughly classifies samples through a control network CN, and finely classifies various rough classes through an identification network RN, so that Chinese characters are identified.
As a further improvement of the present invention, the fuzzy enhancement and smoothing module adopts a model proposed by s.k.pal, etc., from input to output, and two times uses a fuzzy contrast enhancement operator for fuzzy enhancement processing, and one smoothing operation therebetween, to prepare for the next enhancement.
As a further improvement of the invention, the cluster segmentation performs the following steps:
(1) defining sample interval, taking K, C and R (K < S)3 values, and taking K samples as aggregation points;
(2) calculating the distance between every two of the K aggregation points, if the minimum distance is less than C, merging the corresponding two points, and taking the arithmetic mean of the two points as a new aggregation point; repeating the steps until all the intervals are more than or equal to C;
(3) when another (S-K) samples are examined, the distance between the sample and all the agglutination points is calculated every time one sample is examined, and if the minimum distance is greater than R, the sample is used as a new agglutination point; if the minimum distance < R falls within the class of nearest agglutinated points, then recalculating the center of gravity of this class and using this as a new agglutinated point, and if the distance between agglutinated points > C, examining the next sample; otherwise, step 2 is carried out, after the aggregation points are combined, the next sample is inspected until the classification is finished;
(4) inspecting all the samples one by one, clustering according to the step 3, and if the final classification result is the same as the original result, not calculating the gravity center; otherwise, the gravity center is recalculated, and if the classification result is the same as the original result after one-time investigation, the clustering is finished; otherwise, repeating the step 4 until the classification result is completely the same as the previous classification result;
(5) the above clustering process ends with the state of class number 3, so that the class numbers m can be generated1,m2,…,me(me=3), and the respective target functional can be determined in accordance with equation (1)
J d ( m d , C ) = &Sigma; i = 1 md &Sigma; j = 1 r j - r i * j &Element; C i d = 1,2 , &CenterDot; &CenterDot; &CenterDot; , e
(6) Decision J = min { J }1(m1,C),J2(m2,C),…,Je(meThe partition corresponding to C) is the best;
(7) setting thresholds of all levels, setting class CiU luminances: r isi1,ri2,…,riuCorresponding density is P1,P2,…,PiuThen threshold value ri **Is composed of
r i * * = &Sigma; i = 1 u P ij r ij / &Sigma; i = 1 P ij i = 1,2 , &CenterDot; &CenterDot; &CenterDot; , m *
Wherein m is*Is the number of classes that is optimally partitioned.
As a further improvement of the invention, the image data binarization performs the following steps:
(1) counting the total number of pixels in the image
Figure BDA00001925624900023
And the probability of each gray level p (i) = N (i)/N;
(2) dividing the 1, …, M level of image gray into two groups C by a certain gray level k0= 1 → K and C1= { (K +1) → M }, the probability of each group occurring being respectively
C 0 : &omega; 0 = &Sigma; i = 1 k P i = &omega; ( k )
C 1 : &omega; 1 = &Sigma; i = k + 1 M P i = 1 - &omega; ( k )
C0Average value of (2) &mu; 0 = &Sigma; i = 1 k ( i * P i ) / &omega; 0 = &mu; ( k ) / &omega; ( k )
C1Average value of (2) &mu; 1 = &Sigma; i = k + 1 k ( i * P i ) / &omega; 1 = ( &mu; - &mu; ( k ) ) / ( 1 - &omega; ( k ) )
Where μ = ∑ ipiIs the average of the gray levels of the overall image;
Figure BDA00001925624900028
is the average value of the gray scale when the image threshold is k;
(3) changing the k value from 1 to M to obtain the above formula delta2(k) Is the maximum k, the value k at this time*Is the threshold value sought
As a further improvement of the invention, a new additional learning algorithm is adopted, and the learning of all Chinese characters is finally realized by simulating the character learning process of human beings, namely, learning part of Chinese characters first and then gradually learning the rest Chinese characters.
As a further refinement of the invention, the supplementary learning algorithm performs the following steps:
(1) let the original PNN recognize all samples in V, and set the correctly recognized sample set as V0
(2) The samples in the set V-V0 are divided as follows: VT is V-V0Sample subset in PNN learned Chinese characters; VF being V-V0Sample subsets not in the PNN learned Chinese characters;
(3) to VTIf it corresponds to an identified network PNiCan correctly identify it without the need of going to RNiCarry out training whether or notThen these samples are added to the RNiOriginal sample concentration, retraining RNi
(4) To VFTraining new recognition networks RN ', RN', … and RN ', and setting the divided Chinese character subsets as U1', U2 ', … and Uq';
(5) and retraining the control network CN by using the U.V sample set.
The invention mainly comprises four steps of scanning input, fuzzy enhancement and clustering segmentation, image data binarization, Chinese character matching through a parallel neural network and the like, and aiming at the characteristics of Chinese character recognition, new algorithms in the field of artificial intelligence are adopted, and the system can realize the recognition of the printed Chinese characters.
The invention has the beneficial effects that:
(1) the method adopts new algorithms in the field of artificial intelligence, thereby effectively improving the quality of Chinese character recognition;
(2) the designed network can store training mode vectors with strong correlation; the memory capacity of the network is not limited and even all 2's can be storedNTraining mode vectors, and meanwhile, the connection weight of the network only takes 1, 0 or-1, so that the network is easy to realize optically;
(3) the system adopts a maximum variance threshold setting method, and the method can obtain more satisfactory results regardless of whether the image histogram has obvious double peaks or not.
Drawings
FIG. 1 is a schematic structural diagram of a printed Chinese character recognition system in embodiment 1 of the present invention.
Detailed Description
The following further description is made in conjunction with the accompanying drawings and examples.
As shown in FIG. 1, a system for recognizing Chinese characters in printed form, which adopts a new algorithm in the field of artificial intelligence to effectively improve the quality of Chinese character recognition, comprises:
scanning input, converting the Chinese characters printed on the paper surface into electric signals through a photoelectric conversion device to form digital signals with multiple gray levels, and inputting the digital signals into a computer for processing; the fuzzy enhancement and clustering segmentation module comprises fuzzy enhancement, smoothing and multi-gray-level optimal segmentation; after image data binarization, image smoothing and enhancement and multi-gray-level cluster segmentation, carrying out binarization by adopting an integral threshold value selection method, firstly counting a gray histogram of the image, then segmenting the histogram into two groups at a certain threshold value, and determining a threshold value when the variance between the two groups is maximum; the system adopts a parallel neural network method, roughly classifies samples through a control network CN, finely classifies each rough class through an identification network RN, and accordingly identifies the Chinese characters.
The clustering segmentation performs the following steps:
(1) defining sample interval, taking K, C and R (K < S)3 values, and taking K samples as aggregation points;
(2) calculating the distance between every two of the K aggregation points, if the minimum distance is less than C, merging the corresponding two points, and taking the arithmetic mean of the two points as a new aggregation point; repeating the steps until all the intervals are more than or equal to C;
(3) when another (S-K) samples are examined, the distance between the sample and all the agglutination points is calculated every time one sample is examined, and if the minimum distance is greater than R, the sample is used as a new agglutination point; if the minimum distance < R falls within the class of nearest agglutinated points, then recalculating the center of gravity of this class and using this as a new agglutinated point, and if the distance between agglutinated points > C, examining the next sample; otherwise, step 2 is carried out, after the aggregation points are combined, the next sample is inspected until the classification is finished;
(4) inspecting all the samples one by one, clustering according to the step 3, and if the final classification result is the same as the original result, not calculating the gravity center; otherwise, the gravity center is recalculated, and if the classification result is the same as the original result after one-time investigation, the clustering is finished; otherwise, repeating the step 4 until the classification result is completely the same as the previous classification result;
(5) the above clustering process ends with the state of class number 3, so that the class numbers m can be generated1,m2,…,me(me=3), and the respective target functional can be determined in accordance with equation (1)
J d ( m d , C ) = &Sigma; i = 1 md &Sigma; j = 1 r j - r i * j &Element; C i d = 1,2 , &CenterDot; &CenterDot; &CenterDot; , e
(6) Decision J = min { J }1(m1,C),J2(m2,C),…,Je(meThe partition corresponding to C) is the best;
(7) setting thresholds of all levels, setting class CiU luminances: r isi1,ri2,…,riuCorresponding density is P1,P2,…,PiuThen threshold value ri **Is composed of
r i * * = &Sigma; i = 1 u P ij r ij / &Sigma; i = 1 P ij i = 1,2 , &CenterDot; &CenterDot; &CenterDot; , m *
Wherein m is*Is the number of classes that is optimally partitioned.
The image data is binarized to perform the following steps:
(1) counting the total number of pixels in the image
Figure BDA00001925624900043
And the probability of each gray level p (i) = N (i)/N;
(2) dividing the 1, …, M level of image gray into two groups C by a certain gray level k0= 1 → K and C1= { (K +1) → M }, the probability of each group occurring being respectively
C 0 : &omega; 0 = &Sigma; i = 1 k P i = &omega; ( k )
C 1 : &omega; 1 = &Sigma; i = k + 1 M P i = 1 - &omega; ( k )
C0Average value of (2) &mu; 0 = &Sigma; i = 1 k ( i * P i ) / &omega; 0 = &mu; ( k ) / &omega; ( k )
C1Average value of (2) &mu; 1 = &Sigma; i = k + 1 k ( i * P i ) / &omega; 1 = ( &mu; - &mu; ( k ) ) / ( 1 - &omega; ( k ) )
Where μ = ∑ ipiIs the average of the gray levels of the overall image;is the average value of the gray scale when the image threshold is k;
(3) changing the k value from 1 to M to obtain the above formula delta2(k) Is the maximum k, the value k at this time*Is the threshold value sought
The method for recognizing Chinese characters in printed form adopts a parallel neural network method, PNN is specially provided for recognizing a huge sample set of Chinese characters, a control network CN is used for roughly classifying samples, and a recognition network RN is used for finely classifying coarse classes so as to recognize Chinese characters. Both CN and RN adopt Hopfield network, and use the 'outer product equal criterion' as learning rule of associative memory instead of the common outer product and criterion or Hebb learning rule. The system also adopts a new additional learning algorithm which simulates the character learning process of human beings, namely, learning part of Chinese characters first, then gradually learning the rest Chinese characters, and finally realizing the learning of all Chinese characters.

Claims (8)

1. A system for identifying printed chinese characters, the system comprising:
scanning input, converting the Chinese characters printed on the paper surface into electric signals through a photoelectric conversion device to form digital signals with multiple gray levels, and inputting the digital signals into a computer for processing;
the fuzzy enhancement and clustering segmentation module comprises fuzzy enhancement, smoothing and multi-gray-level optimal segmentation;
the image data binarization module is used for carrying out binarization by adopting an integral threshold value selection method after smoothing and enhancing the image and clustering segmentation of multiple gray levels, firstly counting a gray level histogram of the image, then segmenting the histogram into two groups at a certain threshold value, and determining a threshold value when the variance between the two groups is maximum;
the system adopts a parallel neural network method, roughly classifies samples through a control network CN, and finely classifies various rough classes through an identification network RN, so that Chinese characters are identified.
2. The system for identifying printed chinese characters of claim 1, wherein said fuzzy enhancement and smoothing module employs a model proposed by s.k.pal, etc. from input to output, applying fuzzy contrast enhancement operator twice for fuzzy enhancement, and smoothing one time in between, in preparation for the next enhancement.
3. The system for identifying printed chinese characters of claim 1, wherein an image enhanced histogram: having S brightness levels r1,r2,…,rsWith a corresponding probability density of P1,P2,…,PsThe system rewrites the objective function as:
J ( m , c ) = &Sigma; i = 1 m &Sigma; j = 1 r j - r i * j &Element; C i
wherein,
Figure FDA00001925624800012
is CiThe cluster center of (a); r isjIs of class CiThe jth sample in (a); m is the number of clusters, 2<m<s; if we want to group S samples into m classes, the selection of the cluster center is the brightness corresponding to the first m larger probability densities in the histogram.
4. The system for identifying printed Chinese characters as claimed in claim 1 or 3, wherein said clustering segmentation of S intensity clusters is performed by:
(1) defining sample interval, taking K, C and R (K < S)3 values, and taking K samples as aggregation points;
(2) calculating the distance between every two of the K aggregation points, if the minimum distance is less than C, merging the corresponding two points, and taking the arithmetic mean of the two points as a new aggregation point; repeating the steps until all the intervals are more than or equal to C;
(3) when another (S-K) samples are examined, the distance between the sample and all the agglutination points is calculated every time one sample is examined, and if the minimum distance is greater than R, the sample is used as a new agglutination point; if the minimum distance < R falls within the class of nearest agglutinated points, then recalculating the center of gravity of this class and using this as a new agglutinated point, and if the distance between agglutinated points > C, examining the next sample; otherwise, step 2 is carried out, after the aggregation points are combined, the next sample is inspected until the classification is finished;
(4) inspecting all the samples one by one, clustering according to the step 3, and if the final classification result is the same as the original result, not calculating the gravity center; otherwise, the gravity center is recalculated, and if the classification result is the same as the original result after one-time investigation, the clustering is finished; otherwise, repeating the step 4 until the classification result is completely the same as the previous classification result;
(5) the above clustering process ends with the state of class number 3, so that the class numbers m can be generated1,m2,…,me(me=3), and the respective target functional can be determined in accordance with equation (1)
J d ( m d , C ) = &Sigma; i = 1 md &Sigma; j = 1 r j - r i * j &Element; C i d = 1,2 , &CenterDot; &CenterDot; &CenterDot; , e ;
(6) Decision J = min { J }1(m1,C),J2(m2,C),…,Je(meThe partition corresponding to C) is the best;
(7) setting thresholds of all levels, setting class CiU luminances: r isi1,ri2,…,riuCorresponding density is P1,P2,…,PiuThen threshold value ri **Is composed of
r i * * = &Sigma; i = 1 u P ij r ij / &Sigma; i = 1 P ij i = 1,2 , &CenterDot; &CenterDot; &CenterDot; , m * ;
Wherein m is*Is the number of classes that is optimally partitioned.
5. The system for identifying printed Chinese characters as claimed in claim 1, wherein the binarization of image data employs a maximum variance threshold setting method, and comprises the steps of firstly counting a gray histogram of the image, then dividing the histogram into two groups at a certain threshold, and determining the threshold when the variance between the two groups is maximum, and performing the following steps:
(1) counting the total number of pixels in the image
Figure FDA00001925624800023
And the probability of each gray level p (i) = N (i)/N;
(2) dividing the 1, …, M level of image gray into two groups C by a certain gray level k0= 1 → K and C1= { (K +1) → M }, the probability of each group occurring being respectively
C 0 : &omega; 0 = &Sigma; i = 1 k P i = &omega; ( k )
C 1 : &omega; 1 = &Sigma; i = k + 1 M P i = 1 - &omega; ( k )
C0Average value of (2) &mu; 0 = &Sigma; i = 1 k ( i * P i ) / &omega; 0 = &mu; ( k ) / &omega; ( k )
C1Average value of (2) &mu; 1 = &Sigma; i = k + 1 k ( i * P i ) / &omega; 1 = ( &mu; - &mu; ( k ) ) / ( 1 - &omega; ( k ) ) ;
Where μ = ∑ ipiIs the average of the gray levels of the overall image;
Figure FDA00001925624800028
is the average value of the gray scale when the image threshold is k;
(3) changing the k value from 1 to M to obtain the above formula delta2(k) Is the maximum k, the value k at this time*It is the desired threshold.
6. The system for identifying printed Chinese characters as recited in claim 1, wherein said Chinese character identification uses parallel neural network method to coarsely classify samples by control network CN, and uses identification network RN to finely classify coarse classes, thereby identifying Chinese characters, and CN and RN use Hopfield network.
7. The system for recognizing printed Chinese characters as claimed in claim 1 or 6, wherein a new additional learning algorithm is used to simulate the process of learning Chinese characters by human, i.e. learning some Chinese characters first, then learning the rest of Chinese characters gradually, and finally realizing the learning of all Chinese characters.
8. The system for recognition of typed chinese characters according to claim 1 or 7, wherein the additional learning algorithm performs the following steps:
(1) let the original PNN recognize all samples in V, and set the correctly recognized sample set as V0
(2) The samples in the set V-V0 are divided as follows: VT is V-V0Sample subset in PNN learned Chinese characters; VF being V-V0Sample subsets not in the PNN learned Chinese characters;
(3) to VTIf it corresponds to an identified network PNiCan correctly identify it without the need of going to RNiTraining is performed, otherwise the samples are added to RNiOriginal sample concentration, retraining RNi
(4) To VFTraining new recognition networks RN ', RN', … and RN ', and setting the divided Chinese character subsets as U1', U2 ', … and Uq';
(5) and retraining the control network CN by using the U.V sample set.
CN2012102574590A 2012-07-24 2012-07-24 Printed Chinese character recognition system Pending CN102855479A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2012102574590A CN102855479A (en) 2012-07-24 2012-07-24 Printed Chinese character recognition system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2012102574590A CN102855479A (en) 2012-07-24 2012-07-24 Printed Chinese character recognition system

Publications (1)

Publication Number Publication Date
CN102855479A true CN102855479A (en) 2013-01-02

Family

ID=47402056

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2012102574590A Pending CN102855479A (en) 2012-07-24 2012-07-24 Printed Chinese character recognition system

Country Status (1)

Country Link
CN (1) CN102855479A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105938558A (en) * 2015-03-06 2016-09-14 松下知识产权经营株式会社 Learning method
CN110728526A (en) * 2019-08-19 2020-01-24 阿里巴巴集团控股有限公司 Address recognition method, apparatus and computer readable medium

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
何汉华: "《基于Hopfield的脱机手写数字识别理论及算法》", 《通信技术》 *
张云杰等: "《改进的模糊增强算法》", 《辽宁工程技术大学学报(自然科学版)》 *
王先旺等: "《智能神经网络系统原理在印刷体汉字识别中的应用》", 《四川大学学报(工程科学版)》 *
王国胤等: "《汉字识别的并行神经网络方法》", 《模式识别与人工智能》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105938558A (en) * 2015-03-06 2016-09-14 松下知识产权经营株式会社 Learning method
CN105938558B (en) * 2015-03-06 2021-02-09 松下知识产权经营株式会社 Learning method
CN110728526A (en) * 2019-08-19 2020-01-24 阿里巴巴集团控股有限公司 Address recognition method, apparatus and computer readable medium
CN110728526B (en) * 2019-08-19 2024-04-02 创新先进技术有限公司 Address recognition method, device and computer readable medium

Similar Documents

Publication Publication Date Title
CN113378632B (en) Pseudo-label optimization-based unsupervised domain adaptive pedestrian re-identification method
CN107169504B (en) A kind of hand-written character recognition method based on extension Non-linear Kernel residual error network
CN102842032B (en) Method for recognizing pornography images on mobile Internet based on multi-mode combinational strategy
CN103514456B (en) Image classification method and device based on compressed sensing multi-core learning
Wu et al. End-to-end chromosome Karyotyping with data augmentation using GAN
CN109410184B (en) Live broadcast pornographic image detection method based on dense confrontation network semi-supervised learning
Wang et al. Radical-based Chinese character recognition via multi-labeled learning of deep residual networks
CN103077399B (en) Based on the biological micro-image sorting technique of integrated cascade
CN112734775A (en) Image annotation, image semantic segmentation and model training method and device
CN104504365A (en) System and method for smiling face recognition in video sequence
CN108776774A (en) A kind of human facial expression recognition method based on complexity categorization of perception algorithm
Burie et al. ICFHR2016 competition on the analysis of handwritten text in images of balinese palm leaf manuscripts
CN105718889A (en) Human face identity recognition method based on GB(2D)2PCANet depth convolution model
CN109522924A (en) A kind of broad-leaf forest wood recognition method based on single photo
CN113011357A (en) Depth fake face video positioning method based on space-time fusion
CN112990282B (en) Classification method and device for fine-granularity small sample images
CN109086794B (en) Driving behavior pattern recognition method based on T-LDA topic model
CN101364263A (en) Method and system for detecting skin texture to image
CN109614869A (en) A kind of pathological image classification method based on multi-scale compress rewards and punishments network
Bawane et al. Object and character recognition using spiking neural network
CN103971106A (en) Multi-view human facial image gender identification method and device
CN106874825A (en) The training method of Face datection, detection method and device
CN104268587B (en) False fingerprint detection method based on finger wave conversion and SVM
CN107067022B (en) Method, device and equipment for establishing image classification model
CN115310491A (en) Class-imbalance magnetic resonance whole brain data classification method based on deep learning

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C05 Deemed withdrawal (patent law before 1993)
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20130102