CN110008956A - Invoice key message localization method, device, computer equipment and storage medium - Google Patents

Invoice key message localization method, device, computer equipment and storage medium Download PDF

Info

Publication number
CN110008956A
CN110008956A CN201910256914.7A CN201910256914A CN110008956A CN 110008956 A CN110008956 A CN 110008956A CN 201910256914 A CN201910256914 A CN 201910256914A CN 110008956 A CN110008956 A CN 110008956A
Authority
CN
China
Prior art keywords
convolutional neural
neural networks
invoice
key message
net
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910256914.7A
Other languages
Chinese (zh)
Other versions
CN110008956B (en
Inventor
张欢
李爱林
张仕洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Smart Shield Technology Co.,Ltd.
Original Assignee
Shenzhen Huafu Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Huafu Information Technology Co Ltd filed Critical Shenzhen Huafu Information Technology Co Ltd
Priority to CN201910256914.7A priority Critical patent/CN110008956B/en
Publication of CN110008956A publication Critical patent/CN110008956A/en
Application granted granted Critical
Publication of CN110008956B publication Critical patent/CN110008956B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • G06V10/245Aligning, centring, orientation detection or correction of the image by locating a pattern; Special marks for positioning

Abstract

The present invention relates to invoice key message localization method, device, computer equipment and storage medium, this method includes obtaining invoice image to be positioned;Invoice image to be positioned is inputted in convolutional neural networks model and carries out information extraction, to obtain invoice key message;Wherein, convolutional neural networks model is resulting as training data training U-Net convolutional neural networks by the invoice image with feature tag.The present invention is by obtaining invoice image to be positioned, invoice image to be positioned is input to the convolutional neural networks model with U-Net convolutional neural networks for basic network, using convolutional neural networks model first to invoice image to be positioned carry out it is sub-category after, it repositions to the key message part of invoice image to be positioned, and export invoice key message, realization quickly locates out invoice key message, and locating accuracy is high.

Description

Invoice key message localization method, device, computer equipment and storage medium
Technical field
The present invention relates to Method for text detection, more specifically refer to invoice key message localization method, device, computer Equipment and storage medium.
Background technique
Value-added tax common invoice is that the value-added tax general taxpayer in addition to retailing is included in forgery prevention for value-added tax tax control System is issued and is managed, that is to say, that same set of forgery prevention for value-added tax taxation control system can be used in general taxpayer, and to issue value-added tax special With invoice, value-added tax common invoice etc..Enterprise and society, personal, government etc. has more and more invoices to need to arrange, increasingly More time consumptions are manually entered and manual retrieval on invoice, are not only wasted time, but also are easy error.It is badly in need of automation Equipment identified.
Existing character identification system mostly uses traditional computer vision algorithm, does not use neural network, accuracy rate is lower, greatly It needs that the equipment such as scanner is cooperated to be scanned more, it is clean in bill, neatly, and it is apparent Shi Caineng and plays certain identification Effect, it is but mostly helpless to fuzzy bills various under natural scene.Existing most of algorithms in reality mobile phone compared with It is ineffective on any angle truthful data that arbitrarily shoots.Especially when considering invoice and its easy fold, fade, And multi-light and weather condition are faced when mobile phone shooting, and multi-angle, more equipment, the difficulties such as perspective.
Now there are two types of the methods about text information positioning, first is that by the position of scanner positioning key message, and pass through Traditional computer vision means are crossed to be finely adjusted position, but this method will depend on special scanner, very not It is convenient.And require bill fold less, fold invoice in practice cannot be coped with;Second is that requiring user that bill is existed according to instruction Picture pre-align is ajusted, and all literal lines are detected, and is sent into identifying system to all texts of detection and is identified, and with non- Often complicated post-processing means go to carry out printed page analysis, what information analyzes the information recognized on earth is.In general, using such Mode is difficult to carry out information point a large amount of on bill into structure alignment, and due to having detected many extra information more, these The extraction and positioning that mistake in redundant information unexpected will influence itself and pay close attention to information, and subsequent identification and processing, point In analysis, many computing resources will be needed.
Therefore, it is necessary to design a kind of new method, realization quickly locates out invoice key message, and locating accuracy It is high.
Summary of the invention
It is an object of the invention to overcome the deficiencies of existing technologies, invoice key message localization method, device, calculating are provided Machine equipment and storage medium.
To achieve the above object, the invention adopts the following technical scheme: invoice key message localization method, comprising:
Obtain invoice image to be positioned;
Invoice image to be positioned is inputted in convolutional neural networks model and carries out information extraction, to obtain invoice key letter Breath;
Wherein, the convolutional neural networks model is by the invoice image with feature tag as training data training U-Net convolutional neural networks are resulting.
Its further technical solution are as follows: the convolutional neural networks model is made by the invoice image with feature tag It is resulting for training data training U-Net convolutional neural networks, comprising:
The invoice image for having feature tag is obtained, to obtain training data;
Construct U-Net convolutional neural networks;
It is based on deep learning frame using training data to learn U-Net convolutional neural networks, to obtain convolution mind Through network model.
Its further technical solution are as follows: described to be based on deep learning frame to U-Net convolutional Neural net using training data Network is learnt, to obtain convolutional neural networks model, comprising:
Training data is inputted in U-Net convolutional neural networks, to obtain sample key message;
Penalty values are calculated according to sample key message;
It is based on deep learning frame according to penalty values to learn U-Net convolutional neural networks, to obtain convolutional Neural Network model.
Its further technical solution are as follows: the U-Net convolutional neural networks are the invoice image a quarters for generating multilayer The network of the characteristic pattern of resolution ratio.
Its further technical solution are as follows: it is described to input training data in U-Net convolutional neural networks, to obtain sample pass Key information, comprising:
Training data is inputted in U-Net convolutional neural networks and carries out process of convolution, to obtain text box;
Merge the text box, to form sample key message.
The present invention also provides invoice key message positioning devices, comprising:
Data capture unit, for obtaining invoice image to be positioned;
Extraction unit carries out information extraction for inputting invoice image to be positioned in convolutional neural networks model, with To invoice key message.
Its further technical solution are as follows: described device further include:
Model training unit, for training U-Net convolution as training data by the invoice image with feature tag Neural network, to obtain convolutional neural networks model.
Its further technical solution are as follows: the model training unit includes:
Training data obtains subelement, for obtaining the invoice image for having feature tag, to obtain training data;
Network struction subelement, for constructing U-Net convolutional neural networks;
Learn subelement, for being based on deep learning frame to U-Net convolutional neural networks using training data It practises, to obtain convolutional neural networks model.
The present invention also provides a kind of computer equipment, the computer equipment includes memory and processor, described to deposit Computer program is stored on reservoir, the processor realizes above-mentioned method when executing the computer program.
The present invention also provides a kind of storage medium, the storage medium is stored with computer program, the computer journey Sequence can realize above-mentioned method when being executed by processor.
Compared with the prior art, the invention has the advantages that: the present invention, will be undetermined by obtaining invoice image to be positioned Position invoice image is input to the convolutional neural networks model with U-Net convolutional neural networks for basic network, utilizes convolutional Neural After network model is first sub-category to invoice image to be positioned progress, repositioning to the key message part of invoice image to be positioned, And invoice key message is exported, realization quickly locates out invoice key message, and locating accuracy is high.
The invention will be further described in the following with reference to the drawings and specific embodiments.
Detailed description of the invention
Technical solution in order to illustrate the embodiments of the present invention more clearly, below will be to needed in embodiment description Attached drawing is briefly described, it should be apparent that, drawings in the following description are some embodiments of the invention, general for this field For logical technical staff, without creative efforts, it is also possible to obtain other drawings based on these drawings.
Fig. 1 is the application scenarios schematic diagram of invoice key message localization method provided in an embodiment of the present invention;
Fig. 2 is the flow diagram of invoice key message localization method provided in an embodiment of the present invention;
Fig. 3 is the sub-process schematic diagram of invoice key message localization method provided in an embodiment of the present invention;
Fig. 4 is the sub-process schematic diagram of invoice key message localization method provided in an embodiment of the present invention;
Fig. 5 is the sub-process schematic diagram of invoice key message localization method provided in an embodiment of the present invention;
Fig. 6 is provided in an embodiment of the present invention using the typical lane segmentation Pixel-level knot of U-Ne convolutional neural networks processing The schematic diagram of fruit;
Fig. 7 is the schematic diagram of visual characteristic pattern provided in an embodiment of the present invention;
Fig. 8 is the schematic diagram of sample key message provided in an embodiment of the present invention;
Fig. 9 is the schematic block diagram of invoice key message positioning device provided in an embodiment of the present invention;
Figure 10 is the schematic block diagram of computer equipment provided in an embodiment of the present invention.
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete Site preparation description, it is clear that described embodiments are some of the embodiments of the present invention, instead of all the embodiments.Based on this hair Embodiment in bright, every other implementation obtained by those of ordinary skill in the art without making creative efforts Example, shall fall within the protection scope of the present invention.
It should be appreciated that ought use in this specification and in the appended claims, term " includes " and "comprising" instruction Described feature, entirety, step, operation, the presence of element and/or component, but one or more of the other feature, whole is not precluded Body, step, operation, the presence or addition of element, component and/or its set.
It is also understood that mesh of the term used in this description of the invention merely for the sake of description specific embodiment And be not intended to limit the present invention.As description of the invention and it is used in the attached claims, unless on Other situations are hereafter clearly indicated, otherwise " one " of singular, "one" and "the" are intended to include plural form.
It will be further appreciated that the term "and/or" used in description of the invention and the appended claims is Refer to any combination and all possible combinations of one or more of associated item listed, and including these combinations.
Fig. 1 and Fig. 2 are please referred to, Fig. 1 is the application scenarios of invoice key message localization method provided in an embodiment of the present invention Schematic diagram.Fig. 2 is the schematic flow chart of invoice key message localization method provided in an embodiment of the present invention.Invoice key letter Localization method is ceased to be applied in server.Data interaction is carried out between server and terminal, obtains invoice figure to be positioned from terminal Classified and positioned as after, then by convolutional neural networks model, is shown with exporting invoice key message to terminal.
Fig. 2 is the flow diagram of invoice key message localization method provided in an embodiment of the present invention.As shown in Fig. 2, should Method includes the following steps S110 to S120.
S110, invoice image to be positioned is obtained.
In the present embodiment, invoice image to be positioned, which refers to, shoots the resulting photo of invoice by terminal.
S120, information extraction will be carried out in invoice image to be positioned input convolutional neural networks model, to obtain invoice pass Key information.
In the present embodiment, invoice key message refer to invoice codes on invoice, invoice number, the date of making out an invoice, the amount of money, The information of the texts such as the amount of tax to be paid, total amount.
In the present embodiment, above-mentioned convolutional neural networks model is by the invoice image with feature tag as instruction It is resulting to practice data training U-Net convolutional neural networks.
Convolutional neural networks model is widely used in target detection, and example is divided, in the Computer Vision Tasks such as object classification, Very good effect is achieved, shows its adaptability good for Computer Vision Task.For crucial point location, borrow The settling mode for image segmentation task of reflecting, the typical lane segmentation Pixel-level result handled using U-Net convolutional neural networks It can be found in Fig. 6.
It goes to extract characteristic layer using U-Net convolutional neural networks, due to the particularity of invoice image data, training data is not Street scene data are sufficient as, and effect is simultaneously bad.And in order to enhance anti-interference ability, U-Net convolutional neural networks are mentioned The feature taken is constrained, and goes to train U-Net convolutional neural networks using special loss function.
In one embodiment, referring to Fig. 3, above-mentioned step S120 may include step S121~S123.
S121, the invoice image for having feature tag is obtained, to obtain training data.
In the present embodiment, training data refers to the integrated data set of the invoice image with text feature label.
Specifically, a large amount of invoice image can be downloaded from website, and 14 layers of feature are carried out to invoice image and carry out label mark It after fixed, then is input in network, is trained.
S12, building U-Net convolutional neural networks.
In the present embodiment, U-Net convolutional neural networks are mainly made of two parts: constricted path and extensions path.It receives Contracting path is primarily used to capture the contextual information in picture, and the extensions path claimed in contrast is then in order to in picture The required part split carries out precise positioning.Many times the structure of deep learning needs a large amount of example and calculates money Source, but U-Net, which is based on FCN (full convolutional neural networks, Fully Convultional Neural Network), to be changed Into, and the data of some fewer samples can be trained using data enhancing.Constructing U-Net convolutional neural networks can To carry out convolution classification to training data, and the location information of the information such as text is exported, to form characteristic pattern.
The U-Net convolutional neural networks are the nets for generating the characteristic pattern of invoice image a quarter resolution ratio of multilayer Network.
By minimizing the characteristic pattern of prediction and the difference of actual feature tag and image, constantly training U-Net convolution Neural network, the training method are training method general in deep learning, keep predicted value and actual value very nearly the same.Wherein, The first eight layer of characteristic pattern of U-Net convolutional neural networks sorts out background, invoice codes, invoice number is made out an invoice day for classifying more Phase, the amount of money, the amount of tax to be paid, total amount.For each position on the first eight layer of characteristic pattern, correspond on every layer of characteristic pattern eight big The small value for being 0 to 1, this eight values and be 1, the value is maximum on the characteristic pattern of which layer, and representing the position has maximum probability category In which kind of.Each pixel can be found in this way corresponding to which type.After predicting background and classification in this way, for Non- background pixel, generates whether lean on proximal border in the 9th layer of characteristic pattern corresponding position prediction, and the tenth layer of characteristic pattern predicts each side Boundary position is left margin or right margin.Ten one to ten four layers of characteristic pattern then predicted boundary position and two nearest border vertices Positional shift, the position of the characteristic pattern in final each text box can predict a text box.
Using the invoice image with 14 layers of feature tag as training data, go to train U-Net convolutional neural networks, so that Entire U-Net convolutional neural networks learn how generation feature, and providing correct 14 layers of characteristic pattern indicates, and then can be therefrom extensive The information of multiple text box, visual characteristic pattern are as shown in Figure 7.
S123, U-Net convolutional neural networks are learnt based on deep learning frame using training data, to be rolled up Product neural network model.
In one embodiment, above-mentioned step S123 may include step S1231~S1233.
S1231, training data is inputted in U-Net convolutional neural networks, to obtain sample key message.
In the present embodiment, sample key message refer to training data by U-Net convolutional neural networks carry out classification and The text box information formed after positioning.
In one embodiment, above-mentioned step S1231 may include step S1231a~S1232b.
S1231a, process of convolution will be carried out in training data input U-Net convolutional neural networks, to obtain text box.
In the present embodiment, it is right after U-Net convolutional neural networks can first carry out the classification of background and invoice to training data Invoice itself carries out the positioning of text, to obtain text box.
Specifically, it is initially formed the characteristic pattern of invoice, in the pattern image by invoice at text box, under normal circumstances, from 9th layer of characteristic pattern of U-Net convolutional neural networks carries out process of convolution to invoice to the 14th layer of characteristic pattern, each to obtain The visual characteristic pattern of layer, then text box is formed by four endpoint locations of characteristic pattern.
S1231b, merge the text box, to form sample key message.
In one embodiment, there are multiple characteristic patterns on an invoice, therefore have multiple text boxes, this multiple text circle Fixed text information forms key message, and therefore, it is necessary to the text boxes that will acquire to merge, and forms sample key message. The position where sample key message can be finally determined according to the position on four vertex of text box, as shown in Figure 8.
S1232, penalty values are calculated according to sample key message.
In the present embodiment, penalty values are calculated using Loss function according to sample key message, penalty values expression passes through The sample key message of U-Net convolutional neural networks output and the difference value of the feature tag in training data, when penalty values are got over When big, the difference of the two is bigger, then current U-Net convolutional neural networks are formed by model and are not suitable for using working as penalty values When bigger, the difference of the two is smaller, and when penalty values approach a certain threshold value, then current U-Net convolutional neural networks institute The model of formation is suitable for using, the threshold value can according to the actual situation depending on.
S1233, U-Net convolutional neural networks are learnt based on deep learning frame according to penalty values, to be rolled up Product neural network model.
In the present embodiment, use tensorflow deep learning frame as the training of U-Net convolutional neural networks and Learning method.
In learning process, parameters that are extensive and adequately testing de-regulation network are carried out, the U- in paper is made Net convolutional neural networks are adapted to specific task and rate request.The efficiency of U-Net convolutional neural networks is higher, single Forward calculation only has about 1.28Gflops, and forward calculation being capable of a large amount of text detection tasks of real-time parallel processing.
Due to using full convolutional network, it is capable of handling the picture of arbitrary resolution.In addition, real actual use can be put into Bill detection algorithm need to face the fuzzy of picture, a series of problems, such as illumination is bad, physical deformation etc..By fine and wide General picture augmentation and generation, careful has handled this problem well, so that algorithm takes under reality scene, specific service test Obtain extraordinary effect.This method may operate on android equipment, iOS device and server, can quickly position invoice pass Key information text.
Above-mentioned invoice key message localization method, it is by obtaining invoice image to be positioned, invoice image to be positioned is defeated Enter to the convolutional neural networks model for taking U-Net convolutional neural networks as basic network, it is first right using convolutional neural networks model After invoice image progress to be positioned is sub-category, repositioning to the key message part of invoice image to be positioned, and export invoice pass Key information, realization quickly locates out invoice key message, and locating accuracy is high.
Fig. 9 is a kind of schematic block diagram of invoice key message positioning device 300 provided in an embodiment of the present invention.Such as Fig. 9 It is shown, correspond to the above invoice key message localization method, the present invention also provides a kind of invoice key message positioning devices 300. The invoice key message positioning device 300 includes the unit for executing above-mentioned invoice key message localization method, which can To be configured in server.
Specifically, referring to Fig. 9, the invoice key message positioning device 300 includes:
Data capture unit 301, for obtaining invoice image to be positioned;
Extraction unit 302 carries out information extraction for inputting invoice image to be positioned in convolutional neural networks model, with Obtain invoice key message.
In one embodiment, described device further include:
Model training unit, for training U-Net convolution as training data by the invoice image with feature tag Neural network, to obtain convolutional neural networks model.
In one embodiment, the model training unit includes:
Training data obtains subelement, for obtaining the invoice image for having feature tag, to obtain training data;
Network struction subelement, for constructing U-Net convolutional neural networks;
Learn subelement, for being based on deep learning frame to U-Net convolutional neural networks using training data It practises, to obtain convolutional neural networks model.
In one embodiment, the study subelement includes:
Sample key message forms module, for inputting training data in U-Net convolutional neural networks, to obtain sample Key message;
Penalty values computing module, for calculating penalty values according to sample key message;
Deep learning module, for being based on deep learning frame to U-Net convolutional neural networks according to penalty values It practises, to obtain convolutional neural networks model.
In one embodiment, the sample key message formation module includes:
Text box forms submodule, carries out process of convolution for inputting training data in U-Net convolutional neural networks, with Obtain text box;
Merge submodule, for merging the text box, to form sample key message.
It should be noted that it is apparent to those skilled in the art that, above-mentioned invoice key message positioning The specific implementation process of device 300 and each unit, can be with reference to the corresponding description in preceding method embodiment, for the side of description Just and succinctly, details are not described herein.
Above-mentioned invoice key message positioning device 300 can be implemented as a kind of form of computer program, the computer journey Sequence can be run in computer equipment as shown in Figure 10.
Referring to Fig. 10, Figure 10 is a kind of schematic block diagram of computer equipment provided by the embodiments of the present application.The calculating Machine equipment 500 can be server.
Refering to fig. 10, which includes processor 502, memory and the net connected by system bus 501 Network interface 505, wherein memory may include non-volatile memory medium 503 and built-in storage 504.
The non-volatile memory medium 503 can storage program area 5031 and computer program 5032.The computer program 5032 include program instruction, which is performed, and processor 502 may make to execute a kind of invoice key message positioning side Method.
The processor 502 is for providing calculating and control ability, to support the operation of entire computer equipment 500.
The built-in storage 504 provides environment for the operation of the computer program 5032 in non-volatile memory medium 503, should When computer program 5032 is executed by processor 502, processor 502 may make to execute a kind of invoice key message localization method.
The network interface 505 is used to carry out network communication with other equipment.It will be understood by those skilled in the art that in Figure 10 The structure shown, only the block diagram of part-structure relevant to application scheme, does not constitute and is applied to application scheme The restriction of computer equipment 500 thereon, specific computer equipment 500 may include more more or fewer than as shown in the figure Component perhaps combines certain components or with different component layouts.
Wherein, the processor 502 is for running computer program 5032 stored in memory, to realize following step It is rapid:
Obtain invoice image to be positioned;
Invoice image to be positioned is inputted in convolutional neural networks model and carries out information extraction, to obtain invoice key letter Breath;
Wherein, the convolutional neural networks model is by the invoice image with feature tag as training data training U-Net convolutional neural networks are resulting.
In one embodiment, processor 502 is realizing that the convolutional neural networks model is by with feature tag When invoice image is as step obtained by training data training U-Net convolutional neural networks, it is implemented as follows step:
The invoice image for having feature tag is obtained, to obtain training data;
Construct U-Net convolutional neural networks;
It is based on deep learning frame using training data to learn U-Net convolutional neural networks, to obtain convolution mind Through network model.
In one embodiment, processor 502 is based on deep learning frame to U-Net volumes in the realization utilization training data Product neural network is learnt, and when obtaining convolutional neural networks model step, is implemented as follows step:
Training data is inputted in U-Net convolutional neural networks, to obtain sample key message;
Penalty values are calculated according to sample key message;
It is based on deep learning frame according to penalty values to learn U-Net convolutional neural networks, to obtain convolutional Neural Network model.
Wherein, the U-Net convolutional neural networks are the characteristic patterns for generating the invoice image a quarter resolution ratio of multilayer Network.
In one embodiment, processor 502 realize it is described by training data input U-Net convolutional neural networks in, with When obtaining sample key message step, it is implemented as follows step:
Training data is inputted in U-Net convolutional neural networks and carries out process of convolution, to obtain text box;
Merge the text box, to form sample key message.
It should be appreciated that in the embodiment of the present application, processor 502 can be central processing unit (Central Processing Unit, CPU), which can also be other general processors, digital signal processor (Digital Signal Processor, DSP), specific integrated circuit (Application Specific Integrated Circuit, ASIC), ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic Device, discrete gate or transistor logic, discrete hardware components etc..Wherein, general processor can be microprocessor or Person's processor is also possible to any conventional processor etc..
Those of ordinary skill in the art will appreciate that be realize above-described embodiment method in all or part of the process, It is that relevant hardware can be instructed to complete by computer program.The computer program includes program instruction, computer journey Sequence can be stored in a storage medium, which is computer readable storage medium.The program instruction is by the department of computer science At least one processor in system executes, to realize the process step of the embodiment of the above method.
Therefore, the present invention also provides a kind of storage mediums.The storage medium can be computer readable storage medium.This is deposited Storage media is stored with computer program, and processor is made to execute following steps when wherein the computer program is executed by processor:
Obtain invoice image to be positioned;
Invoice image to be positioned is inputted in convolutional neural networks model and carries out information extraction, to obtain invoice key letter Breath;
Wherein, the convolutional neural networks model is by the invoice image with feature tag as training data training U-Net convolutional neural networks are resulting.
In one embodiment, the processor realizes the convolutional neural networks model executing the computer program It is tool when training step obtained by U-Net convolutional neural networks as training data as the invoice image with feature tag Body realizes following steps:
The invoice image for having feature tag is obtained, to obtain training data;
Construct U-Net convolutional neural networks;
It is based on deep learning frame using training data to learn U-Net convolutional neural networks, to obtain convolution mind Through network model.
In one embodiment, the processor is realized the utilization training data and is based in the execution computer program Deep learning frame learns U-Net convolutional neural networks, when obtaining convolutional neural networks model step, specific implementation Following steps:
Training data is inputted in U-Net convolutional neural networks, to obtain sample key message;
Penalty values are calculated according to sample key message;
It is based on deep learning frame according to penalty values to learn U-Net convolutional neural networks, to obtain convolutional Neural Network model.
Wherein, the U-Net convolutional neural networks are the characteristic patterns for generating the invoice image a quarter resolution ratio of multilayer Network.
In one embodiment, the processor is realized and described training data is inputted U- executing the computer program In Net convolutional neural networks, when obtaining sample key message step, it is implemented as follows step:
Training data is inputted in U-Net convolutional neural networks and carries out process of convolution, to obtain text box;
Merge the text box, to form sample key message.
The storage medium can be USB flash disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), magnetic disk Or the various computer readable storage mediums that can store program code such as CD.
Those of ordinary skill in the art may be aware that list described in conjunction with the examples disclosed in the embodiments of the present disclosure Member and algorithm steps, can be realized with electronic hardware, computer software, or a combination of the two, in order to clearly demonstrate hardware With the interchangeability of software, each exemplary composition and step are generally described according to function in the above description.This A little functions are implemented in hardware or software actually, the specific application and design constraint depending on technical solution.Specially Industry technical staff can use different methods to achieve the described function each specific application, but this realization is not It is considered as beyond the scope of this invention.
In several embodiments provided by the present invention, it should be understood that disclosed device and method can pass through it Its mode is realized.For example, the apparatus embodiments described above are merely exemplary.For example, the division of each unit, only Only a kind of logical function partition, there may be another division manner in actual implementation.Such as multiple units or components can be tied Another system is closed or is desirably integrated into, or some features can be ignored or not executed.
The steps in the embodiment of the present invention can be sequentially adjusted, merged and deleted according to actual needs.This hair Unit in bright embodiment device can be combined, divided and deleted according to actual needs.In addition, in each implementation of the present invention Each functional unit in example can integrate in one processing unit, is also possible to each unit and physically exists alone, can also be with It is that two or more units are integrated in one unit.
If the integrated unit is realized in the form of SFU software functional unit and when sold or used as an independent product, It can store in one storage medium.Based on this understanding, technical solution of the present invention is substantially in other words to existing skill The all or part of part or the technical solution that art contributes can be embodied in the form of software products, the meter Calculation machine software product is stored in a storage medium, including some instructions are used so that a computer equipment (can be a People's computer, terminal or network equipment etc.) it performs all or part of the steps of the method described in the various embodiments of the present invention.
The above description is merely a specific embodiment, but scope of protection of the present invention is not limited thereto, any Those familiar with the art in the technical scope disclosed by the present invention, can readily occur in various equivalent modifications or replace It changes, these modifications or substitutions should be covered by the protection scope of the present invention.Therefore, protection scope of the present invention should be with right It is required that protection scope subject to.

Claims (10)

1. invoice key message localization method characterized by comprising
Obtain invoice image to be positioned;
Invoice image to be positioned is inputted in convolutional neural networks model and carries out information extraction, to obtain invoice key message;
Wherein, the convolutional neural networks model is by the invoice image with feature tag as training data training U- Net convolutional neural networks are resulting.
2. invoice key message localization method according to claim 1, which is characterized in that the convolutional neural networks model It is resulting as training data training U-Net convolutional neural networks by the invoice image with feature tag, comprising:
The invoice image for having feature tag is obtained, to obtain training data;
Construct U-Net convolutional neural networks;
It is based on deep learning frame using training data to learn U-Net convolutional neural networks, to obtain convolutional Neural net Network model.
3. invoice key message localization method according to claim 2, which is characterized in that described to be based on using training data Deep learning frame learns U-Net convolutional neural networks, to obtain convolutional neural networks model, comprising:
Training data is inputted in U-Net convolutional neural networks, to obtain sample key message;
Penalty values are calculated according to sample key message;
It is based on deep learning frame according to penalty values to learn U-Net convolutional neural networks, to obtain convolutional neural networks Model.
4. invoice key message localization method according to claim 2, which is characterized in that the U-Net convolutional Neural net Network is the network for generating the characteristic pattern of invoice image a quarter resolution ratio of multilayer.
5. invoice key message localization method according to claim 3, which is characterized in that described that training data is inputted U- In Net convolutional neural networks, to obtain sample key message, comprising:
Training data is inputted in U-Net convolutional neural networks and carries out process of convolution, to obtain text box;
Merge the text box, to form sample key message.
6. invoice key message positioning device characterized by comprising
Data capture unit, for obtaining invoice image to be positioned;
Extraction unit carries out information extraction for inputting invoice image to be positioned in convolutional neural networks model, to be sent out Ticket key message.
7. invoice key message positioning device according to claim 6, which is characterized in that described device further include:
Model training unit, for training U-Net convolutional Neural as training data by the invoice image with feature tag Network, to obtain convolutional neural networks model.
8. invoice key message positioning device according to claim 7, which is characterized in that the model training unit packet It includes:
Training data obtains subelement, for obtaining the invoice image for having feature tag, to obtain training data;
Network struction subelement, for constructing U-Net convolutional neural networks;
Learn subelement, U-Net convolutional neural networks are learnt for being based on deep learning frame using training data, with Obtain convolutional neural networks model.
9. a kind of computer equipment, which is characterized in that the computer equipment includes memory and processor, on the memory It is stored with computer program, the processor is realized as described in any one of claims 1 to 5 when executing the computer program Method.
10. a kind of storage medium, which is characterized in that the storage medium is stored with computer program, the computer program quilt Processor can realize the method as described in any one of claims 1 to 5 when executing.
CN201910256914.7A 2019-04-01 2019-04-01 Invoice key information positioning method, invoice key information positioning device, computer equipment and storage medium Active CN110008956B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910256914.7A CN110008956B (en) 2019-04-01 2019-04-01 Invoice key information positioning method, invoice key information positioning device, computer equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910256914.7A CN110008956B (en) 2019-04-01 2019-04-01 Invoice key information positioning method, invoice key information positioning device, computer equipment and storage medium

Publications (2)

Publication Number Publication Date
CN110008956A true CN110008956A (en) 2019-07-12
CN110008956B CN110008956B (en) 2023-07-07

Family

ID=67169206

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910256914.7A Active CN110008956B (en) 2019-04-01 2019-04-01 Invoice key information positioning method, invoice key information positioning device, computer equipment and storage medium

Country Status (1)

Country Link
CN (1) CN110008956B (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110458162A (en) * 2019-07-25 2019-11-15 上海兑观信息科技技术有限公司 A kind of method of intelligent extraction pictograph information
CN110516541A (en) * 2019-07-19 2019-11-29 金蝶软件(中国)有限公司 Text positioning method, device, computer readable storage medium and computer equipment
CN110517186A (en) * 2019-07-30 2019-11-29 金蝶软件(中国)有限公司 Eliminate method, apparatus, storage medium and the computer equipment of invoice seal
CN110674815A (en) * 2019-09-29 2020-01-10 四川长虹电器股份有限公司 Invoice image distortion correction method based on deep learning key point detection
CN110738092A (en) * 2019-08-06 2020-01-31 深圳市华付信息技术有限公司 invoice text detection method
CN110751088A (en) * 2019-10-17 2020-02-04 深圳金蝶账无忧网络科技有限公司 Data processing method and related equipment
CN111652232A (en) * 2020-05-29 2020-09-11 泰康保险集团股份有限公司 Bill identification method and device, electronic equipment and computer readable storage medium
CN112069893A (en) * 2020-08-03 2020-12-11 中国铁道科学研究院集团有限公司电子计算技术研究所 Bill processing method and device, electronic equipment and storage medium
CN112115934A (en) * 2020-09-16 2020-12-22 四川长虹电器股份有限公司 Bill image text detection method based on deep learning example segmentation
CN112257712A (en) * 2020-10-29 2021-01-22 湖南星汉数智科技有限公司 Train ticket image rectification method and device, computer device and computer readable storage medium
CN112686307A (en) * 2020-12-30 2021-04-20 平安普惠企业管理有限公司 Method, device and storage medium for obtaining invoice based on artificial intelligence
CN116311297A (en) * 2023-04-12 2023-06-23 国网河北省电力有限公司 Electronic evidence image recognition and analysis method based on computer vision

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106845440A (en) * 2017-02-13 2017-06-13 山东万腾电子科技有限公司 A kind of augmented reality image processing method and system
CN107977665A (en) * 2017-12-15 2018-05-01 北京科摩仕捷科技有限公司 The recognition methods of key message and computing device in a kind of invoice
CN108256555A (en) * 2017-12-21 2018-07-06 北京达佳互联信息技术有限公司 Picture material recognition methods, device and terminal
CN108921163A (en) * 2018-06-08 2018-11-30 南京大学 A kind of packaging coding detection method based on deep learning
CN109214382A (en) * 2018-07-16 2019-01-15 顺丰科技有限公司 A kind of billing information recognizer, equipment and storage medium based on CRNN
CN109345540A (en) * 2018-09-15 2019-02-15 北京市商汤科技开发有限公司 A kind of image processing method, electronic equipment and storage medium
CN109345553A (en) * 2018-08-31 2019-02-15 厦门中控智慧信息技术有限公司 A kind of palm and its critical point detection method, apparatus and terminal device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106845440A (en) * 2017-02-13 2017-06-13 山东万腾电子科技有限公司 A kind of augmented reality image processing method and system
CN107977665A (en) * 2017-12-15 2018-05-01 北京科摩仕捷科技有限公司 The recognition methods of key message and computing device in a kind of invoice
CN108256555A (en) * 2017-12-21 2018-07-06 北京达佳互联信息技术有限公司 Picture material recognition methods, device and terminal
CN108921163A (en) * 2018-06-08 2018-11-30 南京大学 A kind of packaging coding detection method based on deep learning
CN109214382A (en) * 2018-07-16 2019-01-15 顺丰科技有限公司 A kind of billing information recognizer, equipment and storage medium based on CRNN
CN109345553A (en) * 2018-08-31 2019-02-15 厦门中控智慧信息技术有限公司 A kind of palm and its critical point detection method, apparatus and terminal device
CN109345540A (en) * 2018-09-15 2019-02-15 北京市商汤科技开发有限公司 A kind of image processing method, electronic equipment and storage medium

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110516541B (en) * 2019-07-19 2022-06-10 金蝶软件(中国)有限公司 Text positioning method and device, computer readable storage medium and computer equipment
CN110516541A (en) * 2019-07-19 2019-11-29 金蝶软件(中国)有限公司 Text positioning method, device, computer readable storage medium and computer equipment
CN110458162B (en) * 2019-07-25 2023-06-23 上海兑观信息科技技术有限公司 Method for intelligently extracting image text information
CN110458162A (en) * 2019-07-25 2019-11-15 上海兑观信息科技技术有限公司 A kind of method of intelligent extraction pictograph information
CN110517186A (en) * 2019-07-30 2019-11-29 金蝶软件(中国)有限公司 Eliminate method, apparatus, storage medium and the computer equipment of invoice seal
CN110738092A (en) * 2019-08-06 2020-01-31 深圳市华付信息技术有限公司 invoice text detection method
CN110738092B (en) * 2019-08-06 2024-04-02 深圳市华付信息技术有限公司 Invoice text detection method
CN110674815A (en) * 2019-09-29 2020-01-10 四川长虹电器股份有限公司 Invoice image distortion correction method based on deep learning key point detection
CN110751088A (en) * 2019-10-17 2020-02-04 深圳金蝶账无忧网络科技有限公司 Data processing method and related equipment
CN111652232A (en) * 2020-05-29 2020-09-11 泰康保险集团股份有限公司 Bill identification method and device, electronic equipment and computer readable storage medium
CN111652232B (en) * 2020-05-29 2023-08-22 泰康保险集团股份有限公司 Bill identification method and device, electronic equipment and computer readable storage medium
CN112069893A (en) * 2020-08-03 2020-12-11 中国铁道科学研究院集团有限公司电子计算技术研究所 Bill processing method and device, electronic equipment and storage medium
CN112115934A (en) * 2020-09-16 2020-12-22 四川长虹电器股份有限公司 Bill image text detection method based on deep learning example segmentation
CN112257712A (en) * 2020-10-29 2021-01-22 湖南星汉数智科技有限公司 Train ticket image rectification method and device, computer device and computer readable storage medium
CN112257712B (en) * 2020-10-29 2024-02-27 湖南星汉数智科技有限公司 Train ticket image alignment method and device, computer device and computer readable storage medium
CN112686307A (en) * 2020-12-30 2021-04-20 平安普惠企业管理有限公司 Method, device and storage medium for obtaining invoice based on artificial intelligence
CN116311297A (en) * 2023-04-12 2023-06-23 国网河北省电力有限公司 Electronic evidence image recognition and analysis method based on computer vision

Also Published As

Publication number Publication date
CN110008956B (en) 2023-07-07

Similar Documents

Publication Publication Date Title
CN110008956A (en) Invoice key message localization method, device, computer equipment and storage medium
Noh et al. Improving occlusion and hard negative handling for single-stage pedestrian detectors
CN107690657B (en) Trade company is found according to image
Li et al. Instance-level salient object segmentation
Li et al. Building block level urban land-use information retrieval based on Google Street View images
CN110135427B (en) Method, apparatus, device and medium for recognizing characters in image
CN107358242B (en) Target area color identification method and device and monitoring terminal
CN107977665A (en) The recognition methods of key message and computing device in a kind of invoice
CN110309706A (en) Face critical point detection method, apparatus, computer equipment and storage medium
CN110287960A (en) The detection recognition method of curve text in natural scene image
Hosny et al. Copy-move forgery detection of duplicated objects using accurate PCET moments and morphological operators
CN109934293A (en) Image-recognizing method, device, medium and obscure perception convolutional neural networks
CN107423278A (en) The recognition methods of essential elements of evaluation, apparatus and system
CN109934181A (en) Text recognition method, device, equipment and computer-readable medium
CN106503703A (en) System and method of the using terminal equipment to recognize credit card number and due date
CN108875731A (en) Target identification method, device, system and storage medium
CN110046617A (en) A kind of digital electric meter reading self-adaptive identification method based on deep learning
CN107657056A (en) Method and apparatus based on artificial intelligence displaying comment information
CN112418216A (en) Method for detecting characters in complex natural scene image
CN107506376A (en) Obtain the client of information point data in region
US20230137337A1 (en) Enhanced machine learning model for joint detection and multi person pose estimation
Ntavelis et al. AIM 2020 challenge on image extreme inpainting
CN110210478A (en) A kind of commodity outer packing character recognition method
CN103946865B (en) Method and apparatus for contributing to the text in detection image
US11809519B2 (en) Semantic input sampling for explanation (SISE) of convolutional neural networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 518000 Room 201, building A, No. 1, Qian Wan Road, Qianhai Shenzhen Hong Kong cooperation zone, Shenzhen, Guangdong (Shenzhen Qianhai business secretary Co., Ltd.)

Applicant after: Shenzhen Huafu Technology Co.,Ltd.

Address before: 518000 Room 201, building A, No. 1, Qian Wan Road, Qianhai Shenzhen Hong Kong cooperation zone, Shenzhen, Guangdong (Shenzhen Qianhai business secretary Co., Ltd.)

Applicant before: SHENZHEN HUAFU INFORMATION TECHNOLOGY Co.,Ltd.

GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20231214

Address after: 518000, 1702, Dashi Building, No. 28 Keji South 1st Road, Gaoxin District Community, Yuehai Street, Nanshan District, Shenzhen, Guangdong Province

Patentee after: Shenzhen Smart Shield Technology Co.,Ltd.

Address before: 518000 Room 201, building A, No. 1, Qian Wan Road, Qianhai Shenzhen Hong Kong cooperation zone, Shenzhen, Guangdong (Shenzhen Qianhai business secretary Co., Ltd.)

Patentee before: Shenzhen Huafu Technology Co.,Ltd.