CN113903043B - Method for identifying printed Chinese character font based on twin metric model - Google Patents

Method for identifying printed Chinese character font based on twin metric model Download PDF

Info

Publication number
CN113903043B
CN113903043B CN202111512767.9A CN202111512767A CN113903043B CN 113903043 B CN113903043 B CN 113903043B CN 202111512767 A CN202111512767 A CN 202111512767A CN 113903043 B CN113903043 B CN 113903043B
Authority
CN
China
Prior art keywords
twin
font
embedded
metric model
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111512767.9A
Other languages
Chinese (zh)
Other versions
CN113903043A (en
Inventor
钟乐海
闫飞
李礁
王荣海
赵红军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mianyang Polytechnic
Original Assignee
Mianyang Polytechnic
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mianyang Polytechnic filed Critical Mianyang Polytechnic
Priority to CN202111512767.9A priority Critical patent/CN113903043B/en
Publication of CN113903043A publication Critical patent/CN113903043A/en
Application granted granted Critical
Publication of CN113903043B publication Critical patent/CN113903043B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Character Discrimination (AREA)

Abstract

The invention discloses a method for identifying a printed Chinese character font based on a twin metric model, which comprises the steps of manufacturing a data set of the printed Chinese character font, including a series of images with different fonts, and constructing a training set and a testing set; training a twin metric model by using a training set, wherein the twin metric model comprises two twin subnetworks with the same structure and shared weight, and acquiring a contrast loss function according to a characteristic vector output by the twin subnetworks; continuously sending different images in the training set into any twin subnetwork of the twin metric model to obtain a plurality of characteristic vectors, and constructing a font space by using a contrast loss function; inputting the font image to be detected into the twin metric model to obtain the characteristic vector, and acquiring the font type of the font to be detected from the font space by using the K nearest neighbor algorithm. The method can improve the recognition precision and robustness of the Chinese character font, and simultaneously ensure that the recognition model is smaller and the operation speed is higher.

Description

Method for identifying printed Chinese character font based on twin metric model
Technical Field
The invention belongs to the technical field of font identification methods, and particularly relates to a method for identifying a font of a printed Chinese character based on a twin metric model.
Background
A Chinese character image can be converted into a Chinese character code by using an OCR (optical character recognition) technology, so that the Chinese character code is input into a computer to realize the digitization of a paper document. OCR has therefore been a focus of research in pattern recognition, image processing, and machine learning since the 80's of the last century. Through the development of many years, the recognition of the printed Chinese characters has achieved better effect at present, and the wide application is started, such as Hanwang OCR, Baidu OCR and Ali OCR. However, the existing Chinese OCR system can only read characters, the format information is lost in the input process, and the original layout cannot be completely recovered. In order to automatically recover the font when the layout is restored, the font needs to be identified, and the font identification is an important premise for realizing 'original text reproduction'. The Chinese character font is one of the difficulties in the mode recognition field because of the characteristics of multiple Chinese character categories, variable styles, multiple similar characters and the like.
The method adopted for Chinese character font identification at present mainly comprises an integral analysis method and an individual analysis method. The entire analysis method uses a layout area as an analysis target, and this method cannot recognize that a plurality of types of font information are contained in the layout area. The individual analysis method takes a single character as an analysis object, and the method can avoid the problems of the overall analysis method in multi-font identification, but the characteristics based on manual screening are difficult to be optimal in the aspects of identification accuracy and efficiency. With the continuous development of deep learning technology, CNN is highlighted in the field of computer vision application research, such as target detection, image classification, and the like, as a mainstream network architecture of deep learning. However, most of the existing methods are used for positioning the character region, extracting and identifying the character region of the natural scene, and cannot accurately and reliably identify the font of the printed Chinese character. Moreover, because the structure of the Chinese characters is complex, and different types of calligraphy fonts are evolved in different calligraphy players, the existing deep learning mode is difficult to deal with the various types of classification tasks, and even if the same accuracy can be achieved, the model is quite huge.
Disclosure of Invention
In order to solve the problems, the invention provides a method for identifying the font of the printed Chinese character based on a twin metric model, which can improve the identification precision and robustness of the font of the Chinese character and ensure that the identification model is smaller and the operation speed is higher.
In order to achieve the purpose, the invention adopts the technical scheme that: a method for identifying a font of a printed Chinese character based on a twin metric model comprises the following steps:
s10, making a printed Chinese font data set comprising a series of images with different fonts, and constructing a training set and a testing set;
s20, training a twin metric model by using a training set, wherein the twin metric model comprises two twin subnetworks with the same structure and shared weight, and acquiring a contrast loss function according to a feature vector output by the twin subnetworks; continuously sending different images in the training set into any twin subnetwork of the twin metric model to obtain a plurality of characteristic vectors, and constructing a font space by using a contrast loss function;
and S30, inputting the font image to be detected into the twin metric model to obtain the characteristic vector, and acquiring the font type of the font to be detected from the font space by using the K nearest algorithm.
Further, a training set is used for training a twin metric model, two font images simultaneously pass through two twin subnetworks which have the same structure and are shared by weight, and a first feature vector and a second feature vector are output; calculating the distance between the first characteristic vector and the second characteristic vector to obtain a contrast loss function; judging whether the two images are similar according to a contrast loss function, and enabling the first feature vector and the second feature vector output by the two twin subnetworks to be close to each other in the font space, otherwise, enabling the first feature vector and the second feature vector to be far away; continuously sending different images in the training set into any twin subnetwork of the twin metric model, obtaining a plurality of characteristic vectors through forward propagation, and constructing a font space.
The twin subnetwork comprises a decomposition convolution character characteristic extraction network and an embedding module, wherein the decomposition convolution character characteristic extraction network is responsible for extracting a bottleneck characteristic vector of a character image and inputting the bottleneck characteristic vector into the embedding module to obtain an embedded characteristic vector.
Furthermore, the decomposed convolution font feature extraction network does not have a full connection layer and an output layer, and is directly connected with the embedded module with the full connection structure.
Further, two twin subnetworks with the same structure and shared weight are utilized in the twin metric model to output a first embedded feature vector and a second embedded feature vector; performing distance calculation on the first embedded characteristic vector and the second embedded characteristic vector to obtain a contrast loss function; and judging whether the two images are similar according to the contrast loss function, and enabling the embedded feature vectors output by the two twin subnetworks to be close to each other in the embedded font space, otherwise, enabling the embedded feature vectors to be far away.
Further, training again, continuously sending different images in the training set into any twin subnetwork of the twin metric model to obtain a plurality of embedded characteristic vectors, obtaining a plurality of embedded characteristic vectors through forward propagation, and constructing an embedded font space.
Furthermore, in the embedded font space, if the input images belong to the same font, the distance between the two embedded feature vectors is close in the embedded space, otherwise, the distance is far; each embedded feature vector is classified into a block type according to the font to which the embedded feature vector belongs.
Further, the decomposed convolution font feature extraction network comprises a normalization processing module, a feature extraction module, multiple types of inclusion modules and an average pooling layer which are connected in sequence, and the input font image sequentially passes through the normalization processing module, the feature extraction module, the multiple types of inclusion modules and the average pooling layer to output bottleneck feature vectors.
Further, the embedded module is a fully-connected neural network and comprises an input layer, a hidden layer and an output layer; the input layer receives a one-dimensional characteristic vector obtained by leveling the bottleneck characteristic vector; and outputting the two-dimensional embedded characteristic vector through the hidden layer and the output layer.
Further, the distance acquisition loss function between the feature vectors E1 and E2 of the two twin subnetwork outputs is:
Figure 436912DEST_PATH_IMAGE001
where N is the number of sample pairs, y =0 indicates that the two twin subnetwork input samples are different samples, and y =1 indicates that the input samples are similar samples;
Figure 922293DEST_PATH_IMAGE002
representing the inter-sample distance.
The beneficial effects of the technical scheme are as follows:
in the font identification task, the method utilizes the twin network to judge the similarity, simultaneously can obtain the embedded characteristics fed back by the twin network to be cached in the embedded space, only needs to measure the Euclidean distance between the judgment target and the embedded space characteristics, has quite excellent performance for the original text reproduction needing high-frequency execution, and has good generalization capability and higher accuracy and robustness than other convolution neural networks.
The method establishes the twin metric model to participate in font identification, can obtain quite good learning effect based on a small sample set, and obtains accurate Chinese character font identification.
The method can cope with the complex structure of the Chinese character, can identify different types of calligraphy fonts evolved from different calligraphy owners, can cope with various types of classification tasks, and can ensure smaller identification model and faster operation rate while improving the identification precision and robustness of the Chinese character fonts.
Drawings
FIG. 1 is a schematic flow chart of a method for identifying a font of a printed Chinese character based on a twin metric model according to the present invention;
FIG. 2 is a schematic diagram illustrating a method for identifying a font of a printed Chinese character based on a twin metric model according to an embodiment of the present invention;
FIG. 3 is a schematic structural diagram of a decomposed convolutional font feature extraction network according to an embodiment of the present invention;
fig. 4 is a schematic structural diagram of an embedded module according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described with reference to the accompanying drawings.
In this embodiment, referring to fig. 1 and fig. 2, the present invention provides a method for identifying a printed chinese character font based on a twin metric model, comprising the steps of:
s10, making a printed Chinese font data set comprising a series of images with different fonts, and constructing a training set and a testing set;
s20, training a twin metric model by using a training set, wherein the twin metric model comprises two twin subnetworks with the same structure and shared weight, and acquiring a contrast loss function according to a feature vector output by the twin subnetworks; continuously sending different images in the training set into any twin subnetwork of the twin metric model to obtain a plurality of characteristic vectors, and constructing a font space by using a contrast loss function;
and S30, inputting the font image to be detected into the twin metric model to obtain the characteristic vector, and acquiring the font type of the font to be detected from the font space by using the K nearest algorithm.
When creating a production printed Chinese font data set comprising a series of images having different fonts, a training set and a test set are constructed: and (3) making a data set of the character volume of the printing Chinese characters, generating a printing character picture by using a PIL tool according to 3755 common Chinese characters in a primary character library and a Chinese character coding table and a character file, wherein the image size is 299 x 299 pixels. The model training and testing can be carried out by adopting printing fonts such as regular script, Song style, black body, Song style imitation, clerical script, young circle, Xinwei Huawen and the like. And randomly selecting 80% of pictures in the printed Chinese character font data set to form a training data set, and forming a test data set by using pictures which are not selected in all font subclasses for evaluating the later algorithm performance.
As an optimization scheme 1 of the above embodiment, a training set is used to train a twin metric model, two font images simultaneously pass through two twin subnetworks with the same structure and shared weight, and a first feature vector and a second feature vector are output; calculating the distance between the first characteristic vector and the second characteristic vector to obtain a contrast loss function; judging whether the two images are similar according to a contrast loss function, and enabling the first feature vector and the second feature vector output by the two twin subnetworks to be close to each other in the font space, otherwise, enabling the first feature vector and the second feature vector to be far away; continuously sending different images in the training set into any twin subnetwork of the twin metric model, obtaining a plurality of characteristic vectors through forward propagation, and constructing a font space.
As an optimization scheme 2 of the above embodiment, the twin subnetwork includes a decomposed convolutional font feature extraction network and an embedding module, where the decomposed convolutional font feature extraction network is responsible for extracting a bottleneck feature vector of a font image, and inputs the bottleneck feature vector to the embedding module to obtain an embedded feature vector. The decomposed convolution font characteristic extraction network does not have a full connection layer and an output layer and is directly connected with an embedded module of a full connection structure.
Outputting a first embedded characteristic vector and a second embedded characteristic vector by utilizing two twin subnetworks which have the same structure and are shared by weights in the twin metric model; performing distance calculation on the first embedded characteristic vector and the second embedded characteristic vector to obtain a contrast loss function; and judging whether the two images are similar according to the contrast loss function, and enabling the embedded feature vectors output by the two twin subnetworks to be close to each other in the embedded font space, otherwise, enabling the embedded feature vectors to be far away.
And (4) training again, continuously sending different images in the training set into any twin subnetwork of the twin metric model to obtain a plurality of embedded characteristic vectors, obtaining a plurality of embedded characteristic vectors through forward propagation, and constructing an embedded font space. If the input images belong to the same font, the distance between the two embedded characteristic vectors in the embedded space is close, otherwise, the distance is far; each embedded feature vector is classified into a block type according to the font to which the embedded feature vector belongs.
As an optimization scheme 3 of the foregoing embodiment, as shown in fig. 3, the decomposed convolution font feature extraction network includes a normalization processing module, a feature extraction module, multiple types of inclusion modules, and an average pooling layer, which are connected in sequence, and the input font image sequentially passes through the normalization processing module, the feature extraction module, the multiple types of inclusion modules, and the average pooling layer to output a bottleneck feature vector.
In the decomposed convolution font feature extraction network, images 299 x 3 are normalized, three convolution layers are sequentially passed through, convolution kernels and step sizes are respectively 3 × 3/2, 3 × 3/1 and 3 × 3/1, and then a feature map C1 is obtained through 3 × 3/2 pooling operation, wherein the size of the feature map C1 is 73 × 73 × 64.
And sequentially passing the feature map C1 through three convolutional layers, wherein the convolutional cores and the step sizes are respectively 3 × 3/1, 3 × 3/2 and 3 × 3/1, so as to obtain a feature map C2 with the size of 35 × 35 × 288.
And sequentially passing the characteristic diagram C2 through three types of inclusion modules to obtain the characteristic diagram C3 with the size of 8 × 8 × 2048.
Finally, C3 passes through the average pooling layer to obtain a bottleneck characteristic of 1 × 1 × 2048.
As an optimization scheme 4 of the above embodiment, as shown in fig. 4, the embedded module is a fully-connected neural network, and includes an input layer, a hidden layer, and an output layer; the input layer receives a one-dimensional characteristic vector obtained by leveling the bottleneck characteristic vector; and outputting the two-dimensional embedded characteristic vector through the hidden layer and the output layer.
The embedded module is a three-layer simple fully-connected neural network. The input layer is 2048 neural units, and receives one-dimensional feature vectors obtained by performing flatten operation on the bottleneck feature vectors. The hidden layer is 512 nerve units, and the output layer is 2 nerve units.
As an optimization scheme 4 of the above embodiment, the distance acquisition loss function between the feature vectors E1 and E2 output by the two twin subnetworks is:
Figure 923747DEST_PATH_IMAGE001
where N is the number of sample pairs, y =0 indicates that the two twin subnetwork input samples are different samples, and y =1 indicates that the input samples are similar samples;
Figure 919384DEST_PATH_IMAGE003
the inter-sample distance is expressed by using Euclidean distance:
Figure 198050DEST_PATH_IMAGE004
e is a self-defined threshold, and when the samples are different and the Euclidean distance is smaller than e, the distance is included in a loss function for calculation, and the situations that the glyphs are different and similar are processed. E can be adjusted as needed to control the degree of recognition.
For a better understanding of the invention, the working principle of the various parts of the invention is explained below:
the process of training the twin metric model by using the training set to obtain the twin metric model is as follows:
firstly, a decomposed convolution font feature extraction network is constructed as a bottleneck feature extraction module, and the strong feature extraction capability of a trained convolution neural network model is applied to a printing Chinese character font recognition task. The feature extraction module converts the printed Chinese character font image into a one-dimensional feature vector. And (4) the feature vectors of the printed Chinese character font images are sent into a twin network to obtain two-dimensional embedded space features, so that the dimension reduction of the features is completed. In the embedded space, a KNN classifier is adopted to classify the images of the printed Chinese characters, and finally the Chinese character font identification is realized.
The invention adopts a twin network structure, is convenient for comparing the characteristics of the same type of fonts and different types of fonts when the paired images of the printed Chinese character fonts are trained, provides the recognition capability of the similar fonts and provides the recognition precision of the Chinese character fonts. Meanwhile, a twin network is adopted, and the interaction between the two network nodes is utilized, so that a quite good learning effect can be obtained based on a small sample set. The recognition model has a small structure while guaranteeing the recognition precision, and improves the operation speed.
The method can accurately extract the bottleneck characteristic vector of the font image by utilizing the decomposed and convolved font characteristic extraction network, and can effectively provide the capability of identifying the font similarity by mutually freezing the decomposed and convolved font characteristic extraction networks in the two networks.
According to the method, the bottleneck characteristic is embedded into the characteristic extraction process, so that the high-dimensional characteristic can be effectively reduced in dimension, the representativeness of the characteristic is improved, the complexity of measurement calculation is reduced, and the identification capability of the font similarity can be effectively provided by sharing the weight.
The method utilizes the measurement calculation of the embedded characteristics, evaluates the similarity between the characteristics of the Chinese character font samples through the Euclidean distance, and uses the similarity as the basis of font identification.
According to the method, the contrast loss function is calculated through the Euclidean distance of the samples, so that the characteristics of the similar font samples are as close as possible and the characteristics of the different font samples are as far away as possible in the training process of the model, and the clustering is facilitated.
The invention adopts the training model to transfer the characteristic extraction capability to the recognition task of the printed Chinese character font, thereby reducing the sample scale of the data set. The invention adopts a hybrid model to realize a supervised learning mode of feature compression and clustering.
Under the condition that training is completed, the detection precision reaches 99.19%. The method has higher detection precision, can accurately identify the font format of the printed Chinese characters, and can be applied to the following aspects: the font information is used for assisting the recognition engine, so that a higher recognition rate is obtained; can be used to distinguish different typesetting and article structures; and the recovery of the document format information is facilitated.
The foregoing shows and describes the general principles and features of the present invention, together with the advantages thereof. It will be understood by those skilled in the art that the present invention is not limited to the embodiments described above, which are described in the specification and illustrated only to illustrate the principle of the present invention, but that various changes and modifications may be made therein without departing from the spirit and scope of the present invention, which fall within the scope of the invention as claimed. The scope of the invention is defined by the appended claims and equivalents thereof.

Claims (7)

1. A method for identifying a printed Chinese character font based on a twin metric model is characterized by comprising the following steps:
s10, making a printed Chinese font data set comprising a series of images with different fonts, and constructing a training set and a testing set;
s20, training a twin metric model by using a training set, wherein the twin metric model comprises two twin subnetworks with the same structure and shared weight, and acquiring a contrast loss function according to a feature vector output by the twin subnetworks; continuously sending different images in the training set into any twin subnetwork of the twin metric model to obtain a plurality of characteristic vectors, and constructing a font space by using a contrast loss function;
s30, inputting the font image to be detected into the twin metric model to obtain a characteristic vector, and acquiring the font type of the font to be detected from the font space by using the K nearest algorithm;
the twin subnetwork comprises a decomposed convolution font feature extraction network and an embedding module, wherein the decomposed convolution font feature extraction network is responsible for extracting bottleneck feature vectors of font images and inputting the bottleneck feature vectors into the embedding module to obtain embedded feature vectors;
the character extraction network comprises a normalization processing module, a character extraction module, a plurality of types of inclusion modules and an average pooling layer which are connected in sequence, and an input character image outputs a bottleneck characteristic vector through the normalization processing module, the character extraction module, the plurality of types of inclusion modules and the average pooling layer in sequence;
the embedded module is a fully-connected neural network and comprises an input layer, a hidden layer and an output layer; the input layer receives a one-dimensional characteristic vector obtained by leveling the bottleneck characteristic vector; and outputting the two-dimensional embedded characteristic vector through the hidden layer and the output layer.
2. The method for identifying the fonts of the printed Chinese characters based on the twin metric model as claimed in claim 1, wherein the twin metric model is trained by a training set, two font images simultaneously pass through two twin subnetworks with the same structure and shared weight, and a first feature vector and a second feature vector are output; calculating the distance between the first characteristic vector and the second characteristic vector to obtain a contrast loss function; judging whether the two images are similar according to a contrast loss function, and enabling the first feature vector and the second feature vector output by the two twin subnetworks to be close to each other in the font space, otherwise, enabling the first feature vector and the second feature vector to be far away; continuously sending different images in the training set into any twin subnetwork of the twin metric model, obtaining a plurality of characteristic vectors through forward propagation, and constructing a font space.
3. The method of claim 2, wherein the decomposed convolutional font feature extraction network does not have a fully connected layer and an output layer, and is directly connected to the fully connected embedded module.
4. The method for identifying printed Chinese character fonts based on the twin metric model as claimed in claim 1, wherein two twin subnetworks with the same structure and shared weight are used in the twin metric model to output a first embedded feature vector and a second embedded feature vector; performing distance calculation on the first embedded characteristic vector and the second embedded characteristic vector to obtain a contrast loss function; and judging whether the two images are similar according to the contrast loss function, and enabling the embedded feature vectors output by the two twin subnetworks to be close to each other in the embedded font space, otherwise, enabling the embedded feature vectors to be far away.
5. The method for identifying printed Chinese character fonts based on twin metric models as claimed in claim 4, wherein the training is repeated, different images in the training set are continuously sent into any twin subnetwork of the twin metric models to obtain a plurality of embedded feature vectors, and the embedded feature vectors are obtained through forward propagation to construct the embedded font space.
6. The method for identifying printed Chinese character fonts based on the twin metric model as recited in claim 5, wherein in the embedded font space, the comparison loss function is used for judging, if the input images belong to the same font, the distance between the two embedded feature vectors is close in the embedded space, otherwise, the distance is far; each embedded feature vector is classified into a block type according to the font to which the embedded feature vector belongs.
7. A method for identifying printed Chinese character fonts based on twin metric models as claimed in claim 1, 2 or 4, characterised in that the distance acquisition penalty function between the feature vectors E1 and E2 of the two twin subnetworks is:
Figure 387700DEST_PATH_IMAGE001
where N is the number of sample pairs, y =0 indicates that the two twin subnetwork input samples are different samples, and y =1 indicates that the input samples are similar samples;
Figure 94494DEST_PATH_IMAGE002
representing the distance between samples, e is the custom threshold.
CN202111512767.9A 2021-12-11 2021-12-11 Method for identifying printed Chinese character font based on twin metric model Active CN113903043B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111512767.9A CN113903043B (en) 2021-12-11 2021-12-11 Method for identifying printed Chinese character font based on twin metric model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111512767.9A CN113903043B (en) 2021-12-11 2021-12-11 Method for identifying printed Chinese character font based on twin metric model

Publications (2)

Publication Number Publication Date
CN113903043A CN113903043A (en) 2022-01-07
CN113903043B true CN113903043B (en) 2022-05-06

Family

ID=79026160

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111512767.9A Active CN113903043B (en) 2021-12-11 2021-12-11 Method for identifying printed Chinese character font based on twin metric model

Country Status (1)

Country Link
CN (1) CN113903043B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117727053B (en) * 2024-02-08 2024-04-19 西南科技大学 Multi-category Chinese character single sample font identification method

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109871851A (en) * 2019-03-06 2019-06-11 长春理工大学 A kind of Chinese-character writing normalization determination method based on convolutional neural networks algorithm
CN110175251A (en) * 2019-05-25 2019-08-27 西安电子科技大学 The zero sample Sketch Searching method based on semantic confrontation network
CN110211594A (en) * 2019-06-06 2019-09-06 杭州电子科技大学 A kind of method for distinguishing speek person based on twin network model and KNN algorithm
CN110222792A (en) * 2019-06-20 2019-09-10 杭州电子科技大学 A kind of label defects detection algorithm based on twin network
CN110533057A (en) * 2019-04-29 2019-12-03 浙江科技学院 A kind of Chinese character method for recognizing verification code under list sample and few sample scene
CN111126563A (en) * 2019-11-25 2020-05-08 中国科学院计算技术研究所 Twin network-based space-time data target identification method and system
CN111354076A (en) * 2020-02-29 2020-06-30 北京航空航天大学 Single-image three-dimensional part combined modeling method based on embedding space
CN111797760A (en) * 2020-07-02 2020-10-20 绵阳职业技术学院 Improved crop pest and disease identification method based on Retianet
CN113283584A (en) * 2021-05-21 2021-08-20 北京大学 Knowledge tracking method and system based on twin network
CN113591857A (en) * 2020-04-30 2021-11-02 阿里巴巴集团控股有限公司 Character image processing method and device and ancient Chinese book image identification method

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
ATE354834T1 (en) * 2002-03-15 2007-03-15 Computer Sciences Corp METHOD AND DEVICE FOR ANALYZING WRITING IN DOCUMENTS
CN103235947B (en) * 2013-04-27 2016-12-07 苏州大学 A kind of Handwritten Numeral Recognition Method and device
US10074042B2 (en) * 2015-10-06 2018-09-11 Adobe Systems Incorporated Font recognition using text localization
CN107229932B (en) * 2016-03-25 2021-05-28 阿里巴巴集团控股有限公司 Image text recognition method and device
CN107644006B (en) * 2017-09-29 2020-04-03 北京大学 Automatic generation method of handwritten Chinese character library based on deep neural network
EP3511868A1 (en) * 2018-01-11 2019-07-17 Onfido Ltd Document authenticity determination
CN108710866B (en) * 2018-06-04 2024-02-20 平安科技(深圳)有限公司 Chinese character model training method, chinese character recognition method, device, equipment and medium
CN109800754B (en) * 2018-12-06 2020-11-06 杭州电子科技大学 Ancient font classification method based on convolutional neural network
CN110414633B (en) * 2019-07-04 2022-10-14 东南大学 System and method for recognizing handwritten fonts

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109871851A (en) * 2019-03-06 2019-06-11 长春理工大学 A kind of Chinese-character writing normalization determination method based on convolutional neural networks algorithm
CN110533057A (en) * 2019-04-29 2019-12-03 浙江科技学院 A kind of Chinese character method for recognizing verification code under list sample and few sample scene
CN110175251A (en) * 2019-05-25 2019-08-27 西安电子科技大学 The zero sample Sketch Searching method based on semantic confrontation network
CN110211594A (en) * 2019-06-06 2019-09-06 杭州电子科技大学 A kind of method for distinguishing speek person based on twin network model and KNN algorithm
CN110222792A (en) * 2019-06-20 2019-09-10 杭州电子科技大学 A kind of label defects detection algorithm based on twin network
CN111126563A (en) * 2019-11-25 2020-05-08 中国科学院计算技术研究所 Twin network-based space-time data target identification method and system
CN111354076A (en) * 2020-02-29 2020-06-30 北京航空航天大学 Single-image three-dimensional part combined modeling method based on embedding space
CN113591857A (en) * 2020-04-30 2021-11-02 阿里巴巴集团控股有限公司 Character image processing method and device and ancient Chinese book image identification method
CN111797760A (en) * 2020-07-02 2020-10-20 绵阳职业技术学院 Improved crop pest and disease identification method based on Retianet
CN113283584A (en) * 2021-05-21 2021-08-20 北京大学 Knowledge tracking method and system based on twin network

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
Ehsan Hosseini-Asl 等.Similarity-based Text Recognition by Deeply Supervised Siamese Network.《aiVir:1511.04397》.2016,1-11. *
Learning to Measure Change: Fully Convolutional Siamese Metric Networks for Scene Change Detection;Enqiang Guo 等;《aiXiv:1810.09111》;20181112;1-10 *
Similarity-based Text Recognition by Deeply Supervised Siamese Network;Ehsan Hosseini-Asl 等;《aiVir:1511.04397》;20160705;1-11,正文第3节第2行,第3.2节第3行,第2.1节第2行,第2.2节第2行,图1,图2,图4,公式1,公式2 *
基于Tesseract-OCR的字符识别技术在特定场合的应用;吴鸣;《湖南城市学院学报(自然科学版)》;20200930;第29卷(第5期);58-61 *
基于卷积神经网络的舰船图像增强算法;闫飞 等;《舰船科学技术》;20190831;第41卷(第8A期);163-165 *
基于迁移学习的卷积神经网络印刷汉字字体识别模型研究;闫飞 等;《数字印刷》;20210430(第2期);36-45 *

Also Published As

Publication number Publication date
CN113903043A (en) 2022-01-07

Similar Documents

Publication Publication Date Title
CN109840521B (en) Integrated license plate recognition method based on deep learning
CN109241317B (en) Pedestrian Hash retrieval method based on measurement loss in deep learning network
CN112818862B (en) Face tampering detection method and system based on multi-source clues and mixed attention
CN109063649B (en) Pedestrian re-identification method based on twin pedestrian alignment residual error network
CN110321967B (en) Image classification improvement method based on convolutional neural network
CN112818951B (en) Ticket identification method
Li et al. Category dictionary guided unsupervised domain adaptation for object detection
CN107169485A (en) A kind of method for identifying mathematical formula and device
CN113239954B (en) Attention mechanism-based image semantic segmentation feature fusion method
CN112633350A (en) Multi-scale point cloud classification implementation method based on graph convolution
CN110826462A (en) Human body behavior identification method of non-local double-current convolutional neural network model
CN107704859A (en) A kind of character recognition method based on deep learning training framework
Zhang et al. Automatic discrimination of text and non-text natural images
Yuan et al. Few-shot scene classification with multi-attention deepemd network in remote sensing
CN112036260A (en) Expression recognition method and system for multi-scale sub-block aggregation in natural environment
Sun et al. Brushstroke based sparse hybrid convolutional neural networks for author classification of Chinese ink-wash paintings
CN107609509A (en) A kind of action identification method based on motion salient region detection
CN112528845A (en) Physical circuit diagram identification method based on deep learning and application thereof
CN115620312A (en) Cross-modal character handwriting verification method, system, equipment and storage medium
CN113903043B (en) Method for identifying printed Chinese character font based on twin metric model
CN116091946A (en) Yolov 5-based unmanned aerial vehicle aerial image target detection method
CN115880704A (en) Automatic case cataloging method, system, equipment and storage medium
CN113657414B (en) Object identification method
CN114022703A (en) Efficient vehicle fine-grained identification method based on deep learning
CN111242114B (en) Character recognition method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant