CN112381086A - Method and device for outputting image character recognition result in structured mode - Google Patents

Method and device for outputting image character recognition result in structured mode Download PDF

Info

Publication number
CN112381086A
CN112381086A CN202011229081.4A CN202011229081A CN112381086A CN 112381086 A CN112381086 A CN 112381086A CN 202011229081 A CN202011229081 A CN 202011229081A CN 112381086 A CN112381086 A CN 112381086A
Authority
CN
China
Prior art keywords
character
classifier
anchoring
position information
detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011229081.4A
Other languages
Chinese (zh)
Inventor
汪泰伸
吴婷婷
吴志鹏
陈德意
刘彩玲
高志鹏
赵建强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen Meiya Pico Information Co Ltd
Original Assignee
Xiamen Meiya Pico Information Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen Meiya Pico Information Co Ltd filed Critical Xiamen Meiya Pico Information Co Ltd
Priority to CN202011229081.4A priority Critical patent/CN112381086A/en
Publication of CN112381086A publication Critical patent/CN112381086A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Character Discrimination (AREA)

Abstract

The invention relates to a method and a device for outputting image character recognition results in a structured manner, wherein the method comprises the following steps: s1, acquiring position information of a detection frame by using an optical character recognition algorithm (OCR); s2, constructing a key field data set with labels, wherein the key field is an information category to be acquired; s3, setting an anchoring field; s4, constructing a feature vector, wherein the relative position information and the relative aspect ratio are used for generating the feature vector; s5, training and optimizing a classifier, and training and optimizing a machine learning classifier by using the generated feature vector; s6, classifying the detection frames, namely classifying the detection frames of the character areas of the image to be recognized by using a machine learning classifier after training optimization; and S7, identifying and outputting a structured result, specifically, identifying characters in the detection box, performing key information matching on the identified characters, correcting and outputting character fields with similar formats, and finally outputting structured result data.

Description

Method and device for outputting image character recognition result in structured mode
Technical Field
The invention relates to the technical field of character recognition, in particular to a method and a device for recognizing a structured output image character recognition result of card information.
Background
The image character recognition mainly utilizes an Optical Character Recognition (OCR) technology to recognize and extract characters in an image into character strings, and then the character strings are edited in a post-processing link. The result recognized by OCR technology is simply a string of editable characters, not containing any structured information. For the result, a series of rules are often required to be established to screen each item for entry, or both the rules and the entries are directly manually entered, which have obvious disadvantages in robustness and efficiency, the former cannot establish a complete set of rules to screen each item of information, and the latter causes great waste of labor cost.
In recent years, with the accelerated popularization of internet of things and mobile internet terminals, common certificates, bank cards, entity business cards and the like are usually photographed and stored in a picture form, so that the card and certificate information management requirement is increasingly highlighted. At present, the common card and certificate character structuring methods mainly comprise two methods: the other method is based on layout analysis, and analyzes the adjacency relation between character areas by using the statistical rule of the layout form, and predicts the field attribution of the character area. The common classification method for information attribution of fields based on the grammar mode depends on the coverage degree of a knowledge base, and the effect of diversified names and enterprise names cannot be guaranteed; certain errors exist in the card processing of some personalized typesetting based on the layout analysis method. Therefore, a method for outputting image character recognition results in a structured manner with better robustness is needed to improve the card information management effect.
Disclosure of Invention
The invention aims to provide a method for outputting an image character recognition result in a structured mode, so as to solve the problems. Therefore, the invention adopts the following specific technical scheme:
according to an aspect of the present invention, there is provided a method for outputting a text recognition result of an image in a structured manner, comprising the following steps:
s1, acquiring position information of a detection frame by using an optical character recognition algorithm (OCR), specifically, performing character detection and recognition on an input image by using the OCR, and acquiring a position information set and a character set of a character area, wherein the position information set is a set formed by the vertex coordinates of the upper left corner and the vertex coordinates of the lower right corner of the detection frame of the character area;
s2, constructing a key field data set with labels, wherein the key field is an information category to be acquired;
s3, setting an anchoring field, specifically, setting the anchoring field according to different input data, constructing a data list of information corresponding to the anchoring field, retrieving anchoring information of a character recognition result, taking a detection box of a corresponding character area as an anchoring box, and calculating relative position information and a relative aspect ratio of the detection box and the anchoring box of other character areas, wherein the relative position information refers to a coordinate difference value of a vertex of a lower right corner of the detection box of the other character areas and a vertex of an upper left corner of the anchoring box;
s4, constructing a feature vector, wherein the relative position information and the relative aspect ratio are used for generating the feature vector;
s5, training and optimizing a classifier, and training and optimizing a machine learning classifier by using the generated feature vector;
s6, classifying the detection frames, namely classifying the detection frames of the character areas of the image to be recognized by using a machine learning classifier after training optimization;
and S7, identifying and outputting a structured result, specifically, identifying characters in the detection box, performing key information matching on the identified characters, correcting and outputting character fields with similar formats, and finally outputting structured result data.
Further, the machine learning classifier adopts a support vector machine classifier and a random forest classifier.
Further, the number of decision trees of the random forest classifier is 10, and the out-of-bag data test is set to True; and setting the kernel function of the vector machine classifier as a Gaussian radial basis kernel function, and setting the penalty factor as 90.
According to another aspect of the present invention, there is provided an apparatus for structured output of image text recognition results, comprising:
the detection frame position information acquisition module is used for acquiring detection frame position information by using an optical character recognition algorithm (OCR), specifically, performing character detection and recognition on an input image by using the OCR to acquire a position information set and a character set of a character area, wherein the position information set is a set formed by a top left corner vertex coordinate and a bottom right corner vertex coordinate of a detection frame of the character area;
the system comprises a key field data set construction module, a key field data set identification module and a key field data set identification module, wherein the key field data set identification module is used for identifying the key field data set with labels;
the anchoring field setting module is used for setting anchoring fields, specifically, setting anchoring fields according to different input data, constructing a data list of information corresponding to the anchoring fields, retrieving anchoring information for character recognition results, taking a detection box of a corresponding character area as an anchoring box, and calculating relative position information and relative width-height ratio of the detection box and the anchoring box of other character areas, wherein the relative position information refers to a coordinate difference value of a vertex of a lower right corner of the detection box of the other character areas and a vertex of an upper left corner of the anchoring box;
the characteristic vector construction module is used for using the relative position information and the relative aspect ratio for generating the characteristic vector;
the classifier training optimization module is used for training and optimizing a classifier, and specifically, training and optimizing a machine learning classifier by using the generated feature vector;
the detection frame classification module is used for classifying the detection frames of the character areas of the images to be recognized by using the machine learning classifier after training optimization;
and the recognition output module is used for recognizing and outputting the structured result, specifically recognizing characters in the detection box, performing key information matching on the recognized characters, correcting and outputting character fields with similar formats, and finally outputting structured result data.
Further, the machine learning classifier adopts a support vector machine classifier and a random forest classifier.
Further, the number of decision trees of the random forest classifier is 10, and the out-of-bag data test is set to True; and setting the kernel function of the vector machine classifier as a Gaussian radial basis kernel function, and setting the penalty factor as 90.
By adopting the technical scheme, the invention has the beneficial effects that: the machine learning classifier used by the invention can process high-dimensional data, and has strong generalization capability and good robustness. For the card and certificate pictures with complex and various layout information, the OCR recognition result is structurally processed by using a machine learning classifier and combining a detection frame of a character area, so that the character information after the image character recognition can be accurately output, and an effective solution is provided for card and certificate information management.
Drawings
To further illustrate the various embodiments, the invention provides the accompanying drawings. The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate embodiments of the invention and, together with the description, serve to explain the principles of the embodiments. Those skilled in the art will appreciate still other possible embodiments and advantages of the present invention with reference to these figures. Elements in the figures are not drawn to scale and like reference numerals are generally used to indicate like elements.
FIG. 1 is a flow chart of a method of the present invention;
FIG. 2 illustrates an image to be recognized;
FIG. 3 is a feature vector obtained by identifying the image shown in FIG. 2 using the method of the present invention;
fig. 4 is an output of the recognition of the image shown in fig. 2 using the method of the present invention.
Fig. 5 is a block diagram of the apparatus of the present invention.
Detailed Description
The invention will now be further described with reference to the accompanying drawings and detailed description.
As shown in fig. 1, a method for outputting a text recognition result of an image in a structured manner may include the following steps:
s1 obtains position information of the detection box by using an Optical Character Recognition (OCR) algorithm, specifically, performs character detection and recognition on the input image by using the OCR algorithm, and obtains a position information set and a character set of the character area, where the position information set is a set formed by top left corner vertex coordinates and bottom right corner vertex coordinates of the detection box of the character area.
And S2, constructing a key field data set with labels, wherein the key field is the type of the information to be acquired.
S3, setting an anchoring field, specifically, setting the anchoring field according to different input data, constructing a data list of information corresponding to the anchoring field, retrieving anchoring information of the character recognition result, taking a detection box of a corresponding character area as an anchoring box, and calculating relative position information and relative aspect ratio of the detection box and the anchoring box of other character areas, wherein the relative position information refers to the coordinate difference value of the vertex of the lower right corner of the detection box of the other character areas and the vertex of the upper left corner of the anchoring box.
And S4, constructing a feature vector, and using the relative position information and the relative aspect ratio for feature vector generation.
S5, training and optimizing a classifier, and training and optimizing a machine learning classifier by using the generated feature vector; the machine learning classifier can adopt a support vector machine classifier and a random forest classifier.
And S6, classifying the detection frames, and classifying the detection frames of the character areas of the image to be recognized by using the machine learning classifier after training optimization. The machine learning classifier carries out key field classification steps as follows: 1) calculating a characteristic vector by a key field detection frame and an anchoring frame, 2) manually marking the category of the characteristic vector and inputting the category into a machine learning classifier for training, so that the classifier can distinguish which kind of key field different fields belong to, 3) calculating the characteristic vector of the frame to be detected and the anchoring frame, and inputting the trained classifier for classification.
And S7, identifying and outputting a structured result, specifically, identifying characters in the detection box, performing key information matching on the identified characters, correcting and outputting character fields with similar formats, and finally outputting structured result data.
The method of the present invention is further explained below by taking a business card as an example. Fig. 2 shows a business card image to be recognized, and the specific processes of recognizing and outputting are as follows:
(1) carrying out character detection and recognition on an input business card image by utilizing an optical character recognition algorithm (OCR) to obtain a position information (coordinate) set and a character set of a character area, wherein the position information refers to the top left corner vertex coordinate and the bottom right corner vertex coordinate of a detection frame of the character area;
(2) constructing a keyword field data set with labels, wherein the keyword fields of the business card data are as follows: company name, job title, mobile phone, telephone, mailbox, fax, address and web address, each category is labeled with a number, for example, 0, 1, 2, 3, 4, 5, 6, 7, 8;
(3) searching a detection box matched with the role field in the business card recognition result from a pre-constructed role list, taking the detection box as an anchor box, and then calculating the coordinate difference value x and y of the vertex at the lower right corner of the detection box to be classified and the vertex at the upper left corner of the anchor box and the relative aspect ratio r of the width and the height of the detection box to be classified and the anchor boxw,rh
(4) The relative position information and relative aspect ratio obtained by the calculation are used for generating the characteristic vector [ x, y, rw,rh];
(5) And using the generated feature vectors for training and optimizing a machine learning classifier, and setting parameters of the random forest classifier as follows: the number of the created decision trees is 10, and the out-of-bag data test is set to True; setting the parameters of the support vector machine as follows: setting the kernel function as a radial basis kernel function and setting a penalty factor as 90;
(6) classifying the character region detection boxes in the business card image by using the machine learning classifier after training optimization, wherein the classification result is shown in fig. 3;
(7) recognizing the characters in the detection box area, matching key information of the recognized characters, correcting and outputting character fields with similar formats, and finally outputting structured result data, as shown in fig. 4.
As shown in fig. 5, an apparatus for structured output of image text recognition results includes:
a detection frame position information obtaining module 100, configured to obtain detection frame position information by using an optical character recognition algorithm (OCR), specifically, perform character detection and recognition on an input image by using the OCR, and obtain a position information set and a character set of a character region, where the position information set is a set formed by an upper left corner vertex coordinate and a lower right corner vertex coordinate of a detection frame of the character region;
a key field data set constructing module 200, configured to construct a key field data set with labels, where a key field is an information category to be acquired;
an anchor field setting module 300, configured to set an anchor field, specifically, set the anchor field according to different input data, construct a data list of information corresponding to the anchor field, perform anchor information retrieval on a text recognition result, and calculate relative position information and a relative aspect ratio between a detection frame and an anchor frame of another text region by using a detection frame of a text region corresponding to the detection frame as the anchor frame, where the relative position information is a coordinate difference between a vertex of a lower right corner of the detection frame and a vertex of an upper left corner of the anchor frame of the other text region;
a feature vector construction module 400 for using the relative position information and the relative aspect ratio for feature vector generation;
a classifier training optimization module 500, configured to train and optimize a classifier, specifically, train and optimize a machine learning classifier using the generated feature vectors;
a detection frame classification module 600, configured to classify a detection frame of a text region of an image to be recognized by using a machine learning classifier after training optimization; the machine learning classifier adopts a support vector machine classifier and a random forest classifier; specific parameters of the support vector machine classifier and the random forest classifier can be set according to different recognition objects, for example, for the business card image recognition shown in fig. 2, the number of decision trees of the random forest classifier is 10, and the out-of-bag data test is set to True; setting the kernel function of the vector machine classifier as a Gaussian radial basis kernel function, and setting the penalty factor as 90;
and an identification output module 700, configured to identify and output a structured result, specifically, identify characters in the detection box, perform key information matching on the identified characters, correct and output character fields with similar formats, and finally output structured result data.
The invention applies the image character recognition and machine learning method to card information structured output, the image character recognition technology can efficiently and accurately extract and recognize characters under complex scenes, the character recognition of print forms is developed and has more mature, a plurality of commercial application cases exist, and meanwhile, the machine learning forms a decision method with certain generalization capability after a large amount of data are accumulated and rules are automatically analyzed, and unknown data can be deduced. The machine learning classifier used by the invention can process high-dimensional data, and has strong generalization capability and good robustness. For the card and certificate pictures with complex and various layout information, the OCR recognition result is structurally processed by using a machine learning classifier and combining a detection frame of a character area, so that the character information after the image character recognition can be accurately output, and an effective solution is provided for card and certificate information management.
While the invention has been particularly shown and described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (6)

1. A method for outputting image character recognition results in a structured manner is characterized by comprising the following steps:
s1, acquiring position information of a detection frame by using an optical character recognition algorithm (OCR), specifically, performing character detection and recognition on an input image by using the OCR to acquire a position information set and a character set of a character area, wherein the position information set is a set formed by a vertex coordinate at the upper left corner and a vertex coordinate at the lower right corner of the detection frame of the character area;
s2, constructing a key field data set with labels, wherein the key field is an information category to be acquired;
s3, setting an anchoring field, specifically, setting the anchoring field according to different input data, constructing a data list of information corresponding to the anchoring field, retrieving anchoring information of a character recognition result, taking a detection box of a corresponding character area as an anchoring box, and calculating relative position information and a relative aspect ratio of the detection box and the anchoring box of other character areas, wherein the relative position information refers to a coordinate difference value of a vertex of a lower right corner of the detection box of the other character areas and a vertex of an upper left corner of the anchoring box;
s4, constructing a feature vector, wherein the relative position information and the relative aspect ratio are used for generating the feature vector;
s5, training and optimizing a classifier, and training and optimizing a machine learning classifier by using the generated feature vector;
s6, classifying the detection frames, namely classifying the detection frames of the character areas of the image to be recognized by using a machine learning classifier after training optimization;
and S7, identifying and outputting a structured result, specifically, identifying characters in the detection box, performing key information matching on the identified characters, correcting and outputting character fields with similar formats, and finally outputting structured result data.
2. The method of claim 1, wherein the machine learning classifier employs a support vector machine classifier and a random forest classifier.
3. A method as claimed in claim 2, wherein the number of decision trees for the random forest classifier is 10, the out-of-bag data test is set to True; and setting the kernel function of the vector machine classifier as a Gaussian radial basis kernel function, and setting the penalty factor as 90.
4. An apparatus for structured output of image text recognition results, comprising:
the detection frame position information acquisition module is used for acquiring detection frame position information by using an optical character recognition algorithm (OCR), specifically, performing character detection and recognition on an input image by using the OCR to acquire a position information set and a character set of a character area, wherein the position information set is a set formed by a top left corner vertex coordinate and a bottom right corner vertex coordinate of a detection frame of the character area;
the system comprises a key field data set construction module, a key field data set identification module and a key field data set identification module, wherein the key field data set identification module is used for identifying the key field data set with labels;
the anchoring field setting module is used for setting anchoring fields, specifically, setting anchoring fields according to different input data, constructing a data list of information corresponding to the anchoring fields, retrieving anchoring information for character recognition results, taking a detection box of a corresponding character area as an anchoring box, and calculating relative position information and relative width-height ratio of the detection box and the anchoring box of other character areas, wherein the relative position information refers to a coordinate difference value of a vertex of a lower right corner of the detection box of the other character areas and a vertex of an upper left corner of the anchoring box;
the characteristic vector construction module is used for using the relative position information and the relative aspect ratio for generating the characteristic vector;
the classifier training optimization module is used for training and optimizing a classifier, and specifically, training and optimizing a machine learning classifier by using the generated feature vector;
the detection frame classification module is used for classifying the detection frames of the character areas of the images to be recognized by using the machine learning classifier after training optimization;
and the recognition output module is used for recognizing and outputting the structured result, specifically recognizing characters in the detection box, performing key information matching on the recognized characters, correcting and outputting character fields with similar formats, and finally outputting structured result data.
5. The apparatus of claim 4, in which a machine learning classifier employs a support vector machine classifier and a random forest classifier.
6. Apparatus as claimed in claim 5, wherein the number of decision trees for the random forest classifier is 10, the out-of-bag data test is set to True; and setting the kernel function of the vector machine classifier as a Gaussian radial basis kernel function, and setting the penalty factor as 90.
CN202011229081.4A 2020-11-06 2020-11-06 Method and device for outputting image character recognition result in structured mode Pending CN112381086A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011229081.4A CN112381086A (en) 2020-11-06 2020-11-06 Method and device for outputting image character recognition result in structured mode

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011229081.4A CN112381086A (en) 2020-11-06 2020-11-06 Method and device for outputting image character recognition result in structured mode

Publications (1)

Publication Number Publication Date
CN112381086A true CN112381086A (en) 2021-02-19

Family

ID=74579804

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011229081.4A Pending CN112381086A (en) 2020-11-06 2020-11-06 Method and device for outputting image character recognition result in structured mode

Country Status (1)

Country Link
CN (1) CN112381086A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113269126A (en) * 2021-06-10 2021-08-17 上海云扩信息科技有限公司 Key information extraction method based on coordinate transformation
CN113378710A (en) * 2021-06-10 2021-09-10 平安科技(深圳)有限公司 Layout analysis method and device for image file, computer equipment and storage medium
CN113591657A (en) * 2021-07-23 2021-11-02 京东科技控股股份有限公司 OCR (optical character recognition) layout recognition method and device, electronic equipment and medium
CN113610098A (en) * 2021-08-19 2021-11-05 创优数字科技(广东)有限公司 Tax payment number identification method and device, storage medium and computer equipment

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105260727A (en) * 2015-11-12 2016-01-20 武汉大学 Academic-literature semantic restructuring method based on image processing and sequence labeling
US20170351913A1 (en) * 2016-06-07 2017-12-07 The Neat Company, Inc. d/b/a Neatreceipts, Inc. Document Field Detection And Parsing
CN109086756A (en) * 2018-06-15 2018-12-25 众安信息技术服务有限公司 A kind of text detection analysis method, device and equipment based on deep neural network
CN109800761A (en) * 2019-01-25 2019-05-24 厦门商集网络科技有限责任公司 Method and terminal based on deep learning model creation paper document structural data
CN110689447A (en) * 2019-08-30 2020-01-14 中国科学院自动化研究所南京人工智能芯片创新研究院 Real-time detection method for social software user published content based on deep learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105260727A (en) * 2015-11-12 2016-01-20 武汉大学 Academic-literature semantic restructuring method based on image processing and sequence labeling
US20170351913A1 (en) * 2016-06-07 2017-12-07 The Neat Company, Inc. d/b/a Neatreceipts, Inc. Document Field Detection And Parsing
CN109086756A (en) * 2018-06-15 2018-12-25 众安信息技术服务有限公司 A kind of text detection analysis method, device and equipment based on deep neural network
CN109800761A (en) * 2019-01-25 2019-05-24 厦门商集网络科技有限责任公司 Method and terminal based on deep learning model creation paper document structural data
CN110689447A (en) * 2019-08-30 2020-01-14 中国科学院自动化研究所南京人工智能芯片创新研究院 Real-time detection method for social software user published content based on deep learning

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113269126A (en) * 2021-06-10 2021-08-17 上海云扩信息科技有限公司 Key information extraction method based on coordinate transformation
CN113378710A (en) * 2021-06-10 2021-09-10 平安科技(深圳)有限公司 Layout analysis method and device for image file, computer equipment and storage medium
CN113378710B (en) * 2021-06-10 2024-03-08 平安科技(深圳)有限公司 Layout analysis method and device for image file, computer equipment and storage medium
CN113591657A (en) * 2021-07-23 2021-11-02 京东科技控股股份有限公司 OCR (optical character recognition) layout recognition method and device, electronic equipment and medium
CN113591657B (en) * 2021-07-23 2024-04-09 京东科技控股股份有限公司 OCR layout recognition method and device, electronic equipment and medium
CN113610098A (en) * 2021-08-19 2021-11-05 创优数字科技(广东)有限公司 Tax payment number identification method and device, storage medium and computer equipment

Similar Documents

Publication Publication Date Title
CN110766014B (en) Bill information positioning method, system and computer readable storage medium
US10943105B2 (en) Document field detection and parsing
CN112381086A (en) Method and device for outputting image character recognition result in structured mode
CN108960223B (en) Method for automatically generating voucher based on intelligent bill identification
US7120318B2 (en) Automatic document reading system for technical drawings
US8744196B2 (en) Automatic recognition of images
US7970213B1 (en) Method and system for improving the recognition of text in an image
CN110717366A (en) Text information identification method, device, equipment and storage medium
CN113963147B (en) Key information extraction method and system based on semantic segmentation
CN110413825B (en) Street-clapping recommendation system oriented to fashion electronic commerce
CN111915635A (en) Test question analysis information generation method and system supporting self-examination paper marking
CN111222585A (en) Data processing method, device, equipment and medium
CN113673528B (en) Text processing method, text processing device, electronic equipment and readable storage medium
CN113762257A (en) Identification method and device for marks in makeup brand images
CN110766001B (en) Bank card number positioning and end-to-end identification method based on CNN and RNN
CN114359912B (en) Software page key information extraction method and system based on graph neural network
Sun et al. TemplateFree: product detection on retail store shelves
CN116092100A (en) Text content extraction method and device
Zuo et al. An intelligent knowledge extraction framework for recognizing identification information from real-world ID card images
CN111241955B (en) Bill information extraction method and system
CN115376149A (en) Reimbursement invoice identification method
CN113591657A (en) OCR (optical character recognition) layout recognition method and device, electronic equipment and medium
Xue et al. Location and interpretation of destination addresses on handwritten Chinese envelopes
CN112287763A (en) Image processing method, apparatus, device and medium
CN114202761B (en) Information batch extraction method based on picture information clustering

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210219