CN113554027A - Method for calibrating and extracting text information of reimbursement receipt image - Google Patents

Method for calibrating and extracting text information of reimbursement receipt image Download PDF

Info

Publication number
CN113554027A
CN113554027A CN202110910082.3A CN202110910082A CN113554027A CN 113554027 A CN113554027 A CN 113554027A CN 202110910082 A CN202110910082 A CN 202110910082A CN 113554027 A CN113554027 A CN 113554027A
Authority
CN
China
Prior art keywords
image
reimbursement
connected domain
document
user information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110910082.3A
Other languages
Chinese (zh)
Inventor
胡为民
郑喜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Dib Enterprise Risk Management Technology Co ltd
Original Assignee
Shenzhen Dib Enterprise Risk Management Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Dib Enterprise Risk Management Technology Co ltd filed Critical Shenzhen Dib Enterprise Risk Management Technology Co ltd
Priority to CN202110910082.3A priority Critical patent/CN113554027A/en
Publication of CN113554027A publication Critical patent/CN113554027A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4007Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration using local operators
    • G06T5/30Erosion or dilatation, e.g. thinning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20036Morphological image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30176Document

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Character Input (AREA)
  • Image Analysis (AREA)
  • Financial Or Insurance-Related Operations Such As Payment And Settlement (AREA)

Abstract

A method for calibrating and extracting text information of reimbursement document images comprises the following steps: 1) disclosed is a method for filtering noise of an reimbursement receipt image. The method is based on an image preprocessing algorithm of OTSU threshold segmentation and EDT distance conversion, noise such as seals, ink dots and wrinkles existing in reimbursement document images is filtered, and the filtered images are used as input of an image text information calibration module. 2) A method for reimbursing image text information calibration. The method adopts a maximum connected domain algorithm to extract a user information connected domain from an reimbursement document image, constructs a correlation matrix according to the corresponding relation between the user information and a document field, and marks and represents the relevance between the reimbursement document user information and a template field connected domain; data enhancement is carried out through disturbance processing such as random rotation, scaling, Gaussian noise and the like, and an SSD network is adopted to train a reimbursement image text information calibration model; and identifying optical character information in the document connected domain by Tesseract, and realizing calibration and extraction of the reimbursement document image text information.

Description

Method for calibrating and extracting text information of reimbursement receipt image
Technical Field
The invention relates to the technical fields of image recognition, machine learning and the like, in particular to a method for calibrating and extracting text information of an reimbursement receipt image.
Background
In management activities such as financial reconciliation and the like, the consistency of electronic document data and paper document data in a financial information management system needs to be compared, and the correctness and the authenticity of financial activities are ensured, however, the situation that the electronic document data and the paper document are inconsistent in comparison often occurs in the practical situation, and the efficiency of financial management is serious. At present, electronic document comparison and paper document comparison are mainly executed in a manual auditing mode of financial staff, so that the efficiency is low and errors still exist.
The optical character recognition technology is one of the core technologies of the electronization of the paper documents at present, provides a feasible technical path for completing the comparison between the paper documents and the electronic documents, but has some technical difficulties: (1) the image of the paper reimbursement document has more noises, such as stains generated in the production process of paper, unclear ink marks generated in the printing process of the reimbursement document, seals generated in the auditing process of the reimbursement document and the like, and the noises very affect the identification accuracy of optical characters; (2) the printing process of the paper reimbursement receipt is affected by the placement position of the printing paper, the condition of misalignment and even dislocation often occurs, and great technical difficulty is brought to accurate identification and extraction of receipt user information.
Therefore, aiming at the technical problems, the invention realizes a method for calibrating and extracting the text information of the image of the paper reimbursement receipt based on the image recognition and machine learning technology, realizes the extraction and alignment of the user information of the image of the paper receipt, compares the user information with the electronic receipt of the financial system, automatically checks the consistency of the paper receipt and the electronic receipt, reduces the financial management cost, improves the financial management efficiency and has higher practical value.
Disclosure of Invention
Aiming at the problems of noise caused by formats and seals in the electronization process of paper reimbursement documents and the problem of dislocation of user information and template fields caused by printing, the invention aims to calibrate and extract the user information in reimbursement document images through image processing, identification and other technologies, generate electronic documents to automatically check, and improve the bill checking efficiency of financial staff. The invention comprises the following steps:
(1) method for filtering noise of reimbursement receipt image
(2) Reimbursement image text information calibration method
The method comprises the following specific steps:
(1) method for filtering noise of reimbursement receipt image
The method is based on an image preprocessing algorithm of OTSU threshold segmentation and EDT distance transformation, noise such as seals, ink dots and wrinkles existing in reimbursement document images is filtered, and the filtered images are used as input of an image text information calibration method.
Extracting pixel matrixes of three color channels of RGB of an original image, performing binarization processing by adopting an OTSU threshold segmentation algorithm to generate a binary mask matrix, filtering the original image, reserving a dark color part which may be a reimbursement document character in the image, and removing noise interference of color information such as a seal.
And adopting EDT distance transformation to the filtered image to generate an Euclidean distance gray image which takes the pixels of the reimbursement document characters reserved by the binary image as targets, wherein the pixel value of the gray image is the distance between the gray image and the nearest target point, adopting an OTSU threshold segmentation algorithm again, setting the threshold to be half of the stroke width of the document printing form, obtaining the extracted binary image, and realizing character refinement.
And performing morphological opening operation on the refined binary image, eliminating isolated and narrow ink dots and seal noise in the image, separating and reimbursing the object adhesion at fine connection positions of characters and characters in the document, characters and form frame lines and the like, and eliminating the convex part of a larger area such as the tail end of a stroke.
Extracting characters in the binary image by adopting a contour extraction algorithm, obtaining a minimum external rectangle of a character connected domain by adopting a maximum connected domain algorithm, setting an aspect ratio threshold of a character rectangular frame, filtering the residual non-character connected domains, and obtaining key fields of the reimbursement receipt and user information after noise filtration.
(2) Method for calibrating reimbursement receipt image text information
In order to realize the calibration and extraction of the reimbursement document information, the invention provides an SSD network-based method for semantic correlation registration of an optical character connected domain and reimbursement document image text information identification based on Tesseract optical character identification. The method takes user information in a document as a detection target, adopts a maximum connected domain algorithm to extract a user information connected domain from an reimbursement document image, constructs a correlation matrix according to the corresponding relation between the user information and a document field, and marks and represents the relevance between the reimbursement document user information and a template field connected domain; based on the connected domain correlation matrix, performing data enhancement through disturbance processing such as random rotation, scaling, Gaussian noise, cutting and the like, and training a reimbursement image text information calibration model by adopting an SSD network for aligning and calibrating user information and template fields; and identifying the optical character information in the document text box by Tesseract, comparing the identification text with the electronic document data, and calibrating and extracting the reimbursement document image text information.
The SSD target detection network comprises a backbone network and a multi-scale feature extraction network, wherein the SSD target detection backbone network is formed by removing full connection layers for classification from a convolutional neural network ResNet50, the multi-scale feature extraction network is a multi-layer down-sampling network, and target frames extracted from different scales are realized by outputting the target frames at each down-sampling layer. The correlation extractor is composed of VGG16 and a fully connected output layer, the input of the correlation extractor is a pair of candidate connected domain images, namely feature maps with the channel number of 2, the feature maps are tiled into 1-dimensional feature vectors after passing through VGG16 and input into the fully connected layer, the fully connected layer is composed of a feature mixed layer of 512 neurons and an output layer of 1 neuron, and the final output of the network is a candidate connected domain correlation value.
The SSD network input comprises two images of user information and a template field connected domain image which are extracted from a reimbursement document image and subjected to noise filtering, the two images are scaled to 256x256 size by adopting a bilinear interpolation algorithm, and min-max normalization processing is carried out; the SSD network output comprises n (n represents the number of connected domains) candidate connected domain coordinates to be detected and n-n correlation matrixes among the candidate connected domains.
The SSD network architecture includes two branches: a branch is a correlation extractor, which is used for calculating a correlation coefficient between the user information and the template field connected domain, inputting n x n connected domain images which are paired pairwise, and outputting an n x n correlation matrix M, wherein the meaning of the correlation matrix M is the matching probability of each pair of user information and the template field connected domain; the other branch connects the output layer through the full link layer, and the output text box position is in the center point-width height (x, y, w, h).
And selecting n values with the maximum correlation probability values in the n columns from the output correlation matrix M, namely the n groups of user information and the matching alignment relation of the template field connected domain. Cutting out a rectangular frame local image from the original document of the reimbursement document at the n user information connected domain positions (x, y, w, h), respectively inputting the image into a Tesseract model to identify optical character information in the image, and calibrating the identified text information with corresponding information in the electronic document.
Drawings
FIG. 1 is a schematic overall flow chart of the present invention.
FIG. 2 is a diagram of a network architecture for a reimbursement image text information calibration method.
Detailed Description
The following describes embodiments of the present invention in further detail with reference to the accompanying drawings. As shown in fig. 1, the invention provides an optical character connected domain semantic correlation registration based on an SSD network and a reimbursement document image text information calibration and extraction method based on Tesseract optical character recognition, and the method includes a reimbursement document image noise filtering method and a reimbursement image text information calibration method. The method for filtering the noise of the reimbursement document image comprises the steps of carrying out image denoising on an original reimbursement document image, filtering noises such as seals, ink dots and wrinkles existing in the reimbursement document image, and inputting the filtered image into a reimbursement image text information calibration method. The method is used for calibrating the format of an reimbursement document image, and as shown in figure 2, based on an SSD target detection network and a correlation extractor, a registration model of user information and template fields in the reimbursement document image is trained in a multitask mode by utilizing a template document containing marking information, an reimbursement document image, and calibration is carried out on the basis of Tesseract optical character recognition technology and data in a financial information management system. The method comprises the following specific steps:
the method comprises the following steps: reimbursement document image binarization
Extracting pixel matrixes of three color channels of RGB of an original image, performing binarization processing by adopting an OTSU threshold segmentation algorithm to generate a binary mask matrix, filtering the original image, reserving a dark color part which may be a reimbursement document character in the image, and removing noise interference of color information such as a seal. The threshold segmentation calculation formula is as follows:
Figure BDA0003203194170000031
wherein img (x, y) is the pixel value of the binary image under the x, y coordinates, and thresh is the threshold (according to the actual reimbursement document image, set according to experience).
Step two: text image extraction and refinement
EDT distance transformation is carried out on the filtered image to generate an Euclidean distance gray level image which takes reimbursement document character pixels reserved by a binary image as targets, wherein the pixel value of the gray level image is the distance between the gray level image and the nearest target point, and the formula is as follows:
gray(p)|p=p(x,y)=minq∈T(ED(p,q))
wherein gray (p) is any pixel point in the gray-scale image, T is all pixel points with a value of 1 in the binary image, and ED is Euclidean distance. And after distance transformation is carried out, setting a threshold value to be half of the stroke width of the document print by adopting an OTSU threshold value segmentation algorithm again, obtaining an extracted binary image, and realizing character refinement.
Step three: image noise filtering and text adhesion separation
And performing morphological opening operation on the refined binary image, eliminating isolated and narrow ink dots and seal noise in the image, separating fine connection parts of characters and characters, the characters and form frame lines and the like from adhesion, and eliminating convex parts of a larger area such as the tail end of a stroke.
Step four: character connected domain extraction of key information
Extracting characters in the binary image by adopting a contour extraction algorithm, obtaining a minimum external rectangle of a character connected domain by adopting a maximum connected domain algorithm, setting an aspect ratio threshold of a character rectangular frame, filtering the residual non-character connected domains, and obtaining key fields of the reimbursement receipt and user information after noise filtration.
Step five: correlation mark of reimbursement document image user information and template field connected domain
And constructing a training data set of the reimbursement bill image calibration model, wherein the training data set comprises common reimbursement bill image data, and each bill image comprises a template field, user information and coordinate labels, such as bill ID, date, address, amount and the like. Extracting a user information connected domain from an reimbursement receipt image by adopting a maximum connected domain algorithm, constructing a correlation matrix according to the corresponding relation between the user information and a receipt field, marking and representing the correlation between the reimbursement receipt user information and a template field connected domain, setting the correlation between the user information and the template field corresponding connected domain as 1, setting the correlation between a non-corresponding connected domain as 0, and generating a large number of corresponding connected domain coordinates and correlation labels as a training set of a text information calibration model.
Step six: training data enhancement
And (3) carrying out disturbance processing such as random rotation, scaling, Gaussian noise, cutting and the like on the reimbursement receipt image in the training set to enhance data so as to reduce overfitting and improve the accuracy of model calibration.
Step seven: reimbursement document calibration model training based on SSD network
An SSD network is adopted to train a reimbursement image text information calibration model, SSD network input comprises two images of user information and a template field communication domain image which are extracted from an reimbursement document image and subjected to noise filtering, the two images are scaled to 256x256 size by adopting a bilinear interpolation algorithm, and min-max normalization processing is carried out; the SSD network output comprises n (n represents the number of connected domains) candidate connected domain coordinates to be detected and n-n correlation matrixes among the candidate connected domains.
The SSD network extracts n characteristic graphs from n connected domains, and performs alignment calibration processing on user information and template fields by adding a correlation extractor. The SSD network architecture includes two branches: a branch is a correlation extractor, which is used for calculating a correlation coefficient between the user information and the template field connected domain, inputting n x n connected domain images which are paired pairwise, and outputting an n x n correlation matrix M, wherein the meaning of the correlation matrix M is the matching probability of each pair of user information and the template field connected domain; the other branch connects the output layer through the full link layer, and the output text box position is in the center point-width height (x, y, w, h).
And selecting n values with the maximum correlation probability values in the n columns from the output correlation matrix M, namely the n groups of user information and the matching alignment relation of the template field connected domain. Probability value matching is calculated as follows:
Figure BDA0003203194170000041
wherein
Figure BDA0003203194170000042
For matching user information, the argmax function is the row number index of the maximum value in a column
And carrying out regression training on the candidate connected domain by using a Smooth L1 Loss function, carrying out Softmax function operation on each column by using the correlation matrix, and carrying out classification training by using a cross entropy Loss function. The specific training loss function formula is as follows:
Figure BDA0003203194170000043
Figure BDA0003203194170000044
Figure BDA0003203194170000045
wherein: y represents a symbolic function (0 or 1), if the same text box is 1, otherwise, the same text box is 0;
Figure BDA0003203194170000046
a candidate box representing a prediction; g represents a true candidate box; c represents the predicted relevance; (i, j) denotes the row and column indices of the correlation matrix.
Step eight: extraction and comparison of reimbursement receipt image information
Cutting out a rectangular frame local image from the original image of the reimbursement document at the n user information connected domain positions (x, y, w, h), respectively inputting the local image into a Tesseract model to identify optical character information in the Tesseract model, and calibrating the identified text information with corresponding data in the electronic document of the financial system.

Claims (3)

1. A method for calibrating and extracting text information of reimbursement document images is characterized by comprising the following steps:
1) the method is based on an image preprocessing algorithm of OTSU threshold segmentation and EDT distance transformation, and is used for filtering noises such as seals, ink dots, wrinkles and the like existing in an reimbursement document image, and taking the filtered image as the input of an image text information calibration module;
2) the method comprises the steps of taking user information in a document as a detection target, adopting a maximum connected domain algorithm to extract a user information connected domain from an reimbursed document image, constructing a correlation matrix according to the corresponding relation between the user information and a document field, and marking and representing the correlation between reimbursed document user information and a template field connected domain; based on the connected domain correlation matrix, performing data enhancement through disturbance processing such as random rotation, scaling, Gaussian noise, cutting and the like, and training a reimbursement image text information calibration model by adopting an SSD network for aligning and calibrating user information and template fields; and identifying the optical character information in the document text box by Tesseract, comparing the identification text with the electronic document data, and calibrating and extracting the reimbursement document image text information.
2. The method for calibrating and extracting the text information of the reimbursement document image according to claim 1, comprising the steps of: the method for filtering the noise of the reimbursement receipt image comprises the following steps:
1) extracting pixel matrixes of three color channels of RGB of an original image, performing binarization processing by adopting an OTSU threshold segmentation algorithm to generate a binary mask matrix, filtering the original image, reserving dark character parts in the image, and removing noise interference of color information such as a seal;
2) EDT distance transformation is carried out on the filtered image, an OTSU threshold segmentation algorithm is adopted again, a threshold is set to be half of the stroke width of a document printing form, an extracted binary image is obtained, character refinement is achieved, and adhesion of objects is removed;
3) and extracting characters in the binary image by adopting a contour extraction algorithm, obtaining a minimum circumscribed rectangle of a character connected domain by adopting a maximum connected domain algorithm, and obtaining key fields of the reimbursement receipt and user information after noise filtration.
3. The method for calibrating and extracting the text information of the reimbursement document image according to claim 1, comprising the steps of: the reimbursement image text information calibration method comprises the following steps:
1) and constructing a target connected domain mark training set, and generating document image data by using various common templates, wherein each document image comprises a plurality of fields and coordinate marks thereof, and the fields correspond to user information and comprise document ID, date, address, user information and the like. Extracting a user information connected domain from an reimbursement receipt image by adopting a maximum connected domain algorithm, constructing a correlation matrix according to the corresponding relation between the user information and a receipt field, marking and expressing the correlation between the reimbursement receipt user information and a template field connected domain, setting the correlation between the user information and the corresponding connected domain of the template field as 1, setting the correlation between a non-corresponding connected domain as 0, and generating a large number of coordinates and correlation marks of the corresponding connected domain as a training set of a text information calibration model;
2) carrying out disturbance processing such as random rotation, scaling, Gaussian noise, cutting and the like on the images in the training set to enhance data so as to reduce overfitting and improve the accuracy of model calibration;
3) an SSD network is adopted to train a reimbursement image text information calibration model, SSD network input comprises two images of user information and a template field communication domain image which are extracted from an reimbursement document image and subjected to noise filtering, the two images are scaled to 256x256 size by adopting a bilinear interpolation algorithm, and min-max normalization processing is carried out; the SSD network output comprises n (n represents the number of connected domains) candidate connected domain coordinates to be detected and n-n correlation matrixes among the candidate connected domains. Performing regression training on the candidate connected domain by using a Smooth L1 Loss function;
4) the SSD network extracts n characteristic graphs from n connected domains, and performs alignment calibration processing on user information and template fields by adding a correlation extractor. The SSD network architecture includes two branches: a branch is a correlation extractor, which is used for calculating a correlation coefficient between the user information and the template field connected domain, inputting n x n connected domain images which are paired pairwise, and outputting an n x n correlation matrix M, wherein the meaning of the correlation matrix M is the matching probability of each pair of user information and the template field connected domain; the other branch is connected with an output layer through a full connecting layer, and the position of the output text box is as the center point-width height (x, y, w, h);
5) after the model SSD network training converges, selecting n values with the maximum relevance probability value in the n columns from the output relevance matrix M, namely the matching alignment relation between n groups of user information and the template field connected domain;
6) cutting out a rectangular frame local image from the original document of the reimbursement document at the n user information connected domain positions (x, y, w, h), respectively inputting the image into a Tesseract model to identify optical character information in the image, and calibrating the identified text information with corresponding information in the electronic document.
CN202110910082.3A 2021-08-09 2021-08-09 Method for calibrating and extracting text information of reimbursement receipt image Pending CN113554027A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110910082.3A CN113554027A (en) 2021-08-09 2021-08-09 Method for calibrating and extracting text information of reimbursement receipt image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110910082.3A CN113554027A (en) 2021-08-09 2021-08-09 Method for calibrating and extracting text information of reimbursement receipt image

Publications (1)

Publication Number Publication Date
CN113554027A true CN113554027A (en) 2021-10-26

Family

ID=78106075

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110910082.3A Pending CN113554027A (en) 2021-08-09 2021-08-09 Method for calibrating and extracting text information of reimbursement receipt image

Country Status (1)

Country Link
CN (1) CN113554027A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113850249A (en) * 2021-12-01 2021-12-28 深圳市迪博企业风险管理技术有限公司 Method for formatting and extracting chart information
CN114627730A (en) * 2022-03-31 2022-06-14 北京科技大学 Braille electronic book
CN117237440A (en) * 2023-10-10 2023-12-15 北京惠朗时代科技有限公司 Image calibration method for printing control instrument
CN117373030A (en) * 2023-06-19 2024-01-09 上海简答数据科技有限公司 OCR-based user material identification method, system, device and medium

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006121410A1 (en) * 2005-05-11 2006-11-16 Agency For Science, Technology And Research Method, apparatus and computer software for segmenting the brain from mr data
CN102708356A (en) * 2012-03-09 2012-10-03 沈阳工业大学 Automatic license plate positioning and recognition method based on complex background
CN105184292A (en) * 2015-08-26 2015-12-23 北京云江科技有限公司 Method for analyzing and recognizing structure of handwritten mathematical formula in natural scene image
CN106156761A (en) * 2016-08-10 2016-11-23 北京交通大学 The image form detection of facing moving terminal shooting and recognition methods
US20180211380A1 (en) * 2017-01-25 2018-07-26 Athelas Inc. Classifying biological samples using automated image analysis
CN109376658A (en) * 2018-10-26 2019-02-22 信雅达系统工程股份有限公司 A kind of OCR method based on deep learning
CN110047083A (en) * 2019-04-01 2019-07-23 江西博微新技术有限公司 Image noise recognition methods, server and storage medium
CN110598566A (en) * 2019-08-16 2019-12-20 深圳中兴网信科技有限公司 Image processing method, device, terminal and computer readable storage medium
CN110619327A (en) * 2018-06-20 2019-12-27 湖南省瞬渺通信技术有限公司 Real-time license plate recognition method based on deep learning in complex scene
CN111340024A (en) * 2020-02-27 2020-06-26 深圳市赤狐软件技术有限公司 Electronic document management method and device, computer equipment and storage medium
CN111611933A (en) * 2020-05-22 2020-09-01 中国科学院自动化研究所 Information extraction method and system for document image

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006121410A1 (en) * 2005-05-11 2006-11-16 Agency For Science, Technology And Research Method, apparatus and computer software for segmenting the brain from mr data
CN102708356A (en) * 2012-03-09 2012-10-03 沈阳工业大学 Automatic license plate positioning and recognition method based on complex background
CN105184292A (en) * 2015-08-26 2015-12-23 北京云江科技有限公司 Method for analyzing and recognizing structure of handwritten mathematical formula in natural scene image
CN106156761A (en) * 2016-08-10 2016-11-23 北京交通大学 The image form detection of facing moving terminal shooting and recognition methods
US20180211380A1 (en) * 2017-01-25 2018-07-26 Athelas Inc. Classifying biological samples using automated image analysis
CN110619327A (en) * 2018-06-20 2019-12-27 湖南省瞬渺通信技术有限公司 Real-time license plate recognition method based on deep learning in complex scene
CN109376658A (en) * 2018-10-26 2019-02-22 信雅达系统工程股份有限公司 A kind of OCR method based on deep learning
CN110047083A (en) * 2019-04-01 2019-07-23 江西博微新技术有限公司 Image noise recognition methods, server and storage medium
CN110598566A (en) * 2019-08-16 2019-12-20 深圳中兴网信科技有限公司 Image processing method, device, terminal and computer readable storage medium
CN111340024A (en) * 2020-02-27 2020-06-26 深圳市赤狐软件技术有限公司 Electronic document management method and device, computer equipment and storage medium
CN111611933A (en) * 2020-05-22 2020-09-01 中国科学院自动化研究所 Information extraction method and system for document image

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
李沛原: "出租车发票信息识别算法的设计与实现", 《中国优秀硕士学位论文全文数据库信息科技辑》, no. 2021, 15 May 2021 (2021-05-15), pages 138 - 1684 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113850249A (en) * 2021-12-01 2021-12-28 深圳市迪博企业风险管理技术有限公司 Method for formatting and extracting chart information
CN114627730A (en) * 2022-03-31 2022-06-14 北京科技大学 Braille electronic book
CN117373030A (en) * 2023-06-19 2024-01-09 上海简答数据科技有限公司 OCR-based user material identification method, system, device and medium
CN117237440A (en) * 2023-10-10 2023-12-15 北京惠朗时代科技有限公司 Image calibration method for printing control instrument
CN117237440B (en) * 2023-10-10 2024-03-15 北京惠朗时代科技有限公司 Image calibration method for printing control instrument

Similar Documents

Publication Publication Date Title
CN113554027A (en) Method for calibrating and extracting text information of reimbursement receipt image
Khan et al. Urdu optical character recognition systems: Present contributions and future directions
Sharma et al. Character recognition using neural network
Shelke et al. A multistage handwritten Marathi compound character recognition scheme using neural networks and wavelet features
Manigandan et al. Tamil character recognition from ancient epigraphical inscription using OCR and NLP
CN112069900A (en) Bill character recognition method and system based on convolutional neural network
CN111523622B (en) Method for simulating handwriting by mechanical arm based on characteristic image self-learning
Lutf et al. Arabic font recognition based on diacritics features
CN114898472B (en) Signature identification method and system based on twin vision transducer network
CN115620312A (en) Cross-modal character handwriting verification method, system, equipment and storage medium
CN115880566A (en) Intelligent marking system based on visual analysis
CN113971805A (en) Intelligent marking and scoring method combining machine vision and semantic analysis
Mehran et al. A front-end OCR for omni-font Persian/Arabic cursive printed documents
Al Ghamdi A novel approach to printed Arabic optical character recognition
Aravinda et al. Template matching method for Kannada handwritten recognition based on correlation analysis
Patel et al. Multiresolution technique to handwritten English character recognition using learning rule and Euclidean distance metric
Bhuvneswari et al. Recognition of ancient stone inscription characters using histogram of oriented gradients
Kibria Bengali optical character recognition using self organizing map
CN111325270B (en) Dongba text recognition method based on template matching and BP neural network
Islam et al. Towards building a bangla text recognition solution with a multi-headed cnn architecture
Ajao et al. Yoruba handwriting word recognition quality evaluation of preprocessing attributes using information theory approach
CN111046874A (en) Single number identification method based on template matching
Sharma et al. CDRAMM: Character And Digit Recognition Aided by Mathematical Morphology
Sahu et al. Survey and analysis of devnagari character recognition techniques using neural networks
Fadl et al. Automatic fake document identification and localization using DE-Net and color-based features of foreign inks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination