WO2022145343A1 - Architecture for digitalizing documents using multi-model deep learning, and document image processing program - Google Patents
Architecture for digitalizing documents using multi-model deep learning, and document image processing program Download PDFInfo
- Publication number
- WO2022145343A1 WO2022145343A1 PCT/JP2021/047935 JP2021047935W WO2022145343A1 WO 2022145343 A1 WO2022145343 A1 WO 2022145343A1 JP 2021047935 W JP2021047935 W JP 2021047935W WO 2022145343 A1 WO2022145343 A1 WO 2022145343A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- character string
- document image
- layout
- unit
- document
- Prior art date
Links
- 238000013135 deep learning Methods 0.000 title description 2
- 238000000034 method Methods 0.000 claims abstract description 100
- 230000006870 function Effects 0.000 claims description 58
- 238000012937 correction Methods 0.000 claims description 56
- 238000007781 pre-processing Methods 0.000 claims description 29
- 238000012015 optical character recognition Methods 0.000 abstract description 6
- 238000003384 imaging method Methods 0.000 abstract description 2
- 230000008569 process Effects 0.000 description 89
- 238000012545 processing Methods 0.000 description 34
- 238000010586 diagram Methods 0.000 description 20
- 238000004891 communication Methods 0.000 description 14
- 238000005520 cutting process Methods 0.000 description 12
- 230000010365 information processing Effects 0.000 description 8
- OKTJSMMVPCPJKN-UHFFFAOYSA-N Carbon Chemical compound [C] OKTJSMMVPCPJKN-UHFFFAOYSA-N 0.000 description 5
- 229910052799 carbon Inorganic materials 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 4
- 230000008030 elimination Effects 0.000 description 3
- 238000003379 elimination reaction Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 238000012805 post-processing Methods 0.000 description 3
- 239000000284 extract Substances 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000009825 accumulation Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000007257 malfunction Effects 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/14—Image acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/19—Recognition using electronic means
- G06V30/192—Recognition using electronic means using simultaneous comparisons or correlations of the image signals with a plurality of references
- G06V30/194—References adjustable by an adaptive method, e.g. learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/40—Document-oriented image-based pattern recognition
- G06V30/41—Analysis of document content
- G06V30/412—Layout analysis of documents structured with printed lines or input boxes, e.g. business forms or tables
Definitions
- the present invention relates to an electronic document generator, an electronic document generation method, and an electronic document generation program, and more particularly to an electronic document generator that scans a paper document to generate an electronic document, an electronic document generation method, and an electronic document generation program. Is.
- the object of the electronic document generator, the electronic document generation method, and the electronic document generation program of the present disclosure is to convert a character string included in a document image into text data by a method different from the conventional optical character recognition. And.
- the electronic document generator is a character string that learns the correspondence between the document image acquisition unit that acquires the document image obtained by imaging the document and the character string included in the document image.
- the character string recognition unit that recognizes the character string included in the document image acquired by the document image acquisition unit and generates the text data related to the character string, and the text data as the text of the electronic medium. It has an output unit to output.
- a layout learning model in which a correspondence relationship between a plurality of elements included in a document image and identification information of each of the plurality of elements is learned is used. , Specify the range in each document image of the plurality of elements included in the document image acquired by the document image acquisition unit, recognize each type of the plurality of elements, and the document relating to each range of the plurality of elements.
- a layout recognition unit for acquiring position information in an image is further provided, and the character string recognition unit recognizes a character string included in a range specified by the layout recognition unit using a character string learning model and converts it into a character string.
- the text data may be generated, and the output unit may output each of the text data related to the plurality of elements as text in an electronic medium to each position information in the range related to the plurality of elements.
- the third aspect is that in the electronic document generator according to the second aspect, the type of the element may be any of a character string, a table, an image, a seal, or a handwriting.
- each cell in the table included in the element is cut out and the cell is cut out.
- the character string recognition unit further includes a cutout unit for acquiring position information in each document image of the above, and the character string recognition unit uses a character string learning model for a character string included in each of the cells cut out in the cutout unit. It may be recognized and the text data related to the character string may be generated.
- the fifth aspect is the document image including a plurality of elements in the electronic document generator according to the second to fourth aspects, and the element is given an annotation associated with the type corresponding to each of the elements. It also has a layout learning data generation unit that accumulates multiple document images with annotations and generates layout learning data, and the layout learning data is used for supervised learning of the layout learning model. May be good.
- a sixth aspect is that in the electronic document generator according to the fifth aspect, position information in each document image of a range related to a plurality of elements included in the document image is added to the document image together with annotations. May be good.
- a seventh aspect is a document of each type of a plurality of elements recognized by the layout recognition unit and a range of each of the plurality of elements based on the input in the electronic document generator according to the fifth or sixth aspect. At least one of the position information in the image may be modified, and a layout learning data correction unit for updating the layout learning data by adding the modified data may be further provided.
- the eighth aspect further includes a layout learning unit that relearns the layout learning model using the layout learning data updated by the layout learning data correction unit in the electronic document generation device according to the seventh aspect. It may be that.
- a ninth aspect is that the electronic document generator according to the second to eighth aspects further includes a character string learning data generation unit that generates character string learning data used for supervised learning of a character string learning model. May be good.
- the text data generated by the character string recognition unit is modified based on the input, and the modified text data is added to obtain the character string.
- a character string learning data correction unit for updating the learning data may be further provided.
- the eleventh aspect is the character string learning in which the character string learning model is relearned by using the character string learning data updated by the character string learning data correction unit in the electronic document generator according to the tenth aspect. It may be provided further.
- a twelfth aspect is the electronic document generator according to the second to eleventh aspects, wherein the character string recognition unit includes a plurality of character string learning models and is adapted to the language of the character string included in each of the plurality of elements.
- a character string learning model may be used.
- a thirteenth aspect further comprises a preprocessing unit that performs preprocessing on a document image acquired by a document image acquisition unit in the electronic document generation device according to the second to twelfth aspects, and the preprocessing unit is a background removing unit.
- the preprocessing unit is a background removing unit.
- a tilt correction unit and a shape adjustment unit are provided, the background removal unit removes the background of the document image acquired by the document image acquisition unit, and the tilt correction unit corrects the inclination of the document image acquired by the document image acquisition unit.
- the shape adjusting unit may adjust the overall shape and size of the document image acquired by the document image acquisition unit.
- a fourteenth aspect is the electronic document generator according to the second to thirteenth aspects, wherein the layout learning model includes a layout learning model for a contract, a layout learning model for an invoice, a layout learning model for a memorandum, and a delivery note. It may be either a layout learning model for a receipt or a layout learning model for a receipt.
- the computer used in the electronic document generation device includes a document image acquisition step of acquiring a document image in which a document is imaged, a document image, and a character string included in the document image.
- the character string recognition step that recognizes the character string included in the document image acquired in the document image acquisition step and generates the text data related to the character string.
- An output step that outputs the text data as text on an electronic medium.
- the electronic document generation program includes a document image acquisition function for acquiring a document image in which a document is imaged, and a document image and a character string included in the document image on a computer used in the electronic document generation device.
- a document image acquisition function for acquiring a document image in which a document is imaged
- a document image and a character string included in the document image on a computer used in the electronic document generation device With the character string recognition function that recognizes the character string included in the document image acquired by the document image acquisition function and generates the text data related to the character string using the character string learning model that learned the correspondence between , The output function that outputs text data as text on an electronic medium is demonstrated.
- the electronic document generator uses a document image acquisition unit that acquires a document image that is an image of a document, and a character string learning model that learns the correspondence between the document image and the character string included in the document image.
- a character string recognition unit that recognizes the character string included in the document image acquired by the document image acquisition unit and generates text data related to the character string, and an output unit that outputs the text data as text on an electronic medium. Since the character string included in the document image is recognized as characters using the machine-learned model, it is possible to improve the recognition efficiency of character recognition when converting the document image into text data.
- FIG. 1 is a diagram showing an outline of an electronic document generation system 100 including an electronic document generation device 10.
- the electronic document generation system 100 includes an electronic document generation device 10, a user terminal 12, a character string learning model 13, a layout learning model 14, a document image database 15, and the like.
- the electronic document generator 10, the user terminal 12, the character string learning model 13, the layout learning model 14, and the document image database 15 are connected to the information communication network 11, and each of them can communicate with each other.
- the electronic document generation system 100 uses the electronic document generation device 10 to recognize character strings included in a document image and generate text data.
- the electronic document generation device 10 recognizes the layout of the document image by using the layout learning model, and recognizes the character string included in the document image by using the character string learning model.
- the electronic document generator 10 is a kind of computer represented by, for example, a personal computer and is an information processing device.
- the electronic document generation device 10 also includes arithmetic processing devices and microcomputers included in various computers, and also includes devices and devices capable of realizing the functions according to the present disclosure by an application.
- the character string learning model 13 is a learning model that recognizes an image of a character string included in a document image, and is used for character recognition of the electronic document generation device 10.
- the storage location of the character string learning model 13 is arbitrary as long as it can be used by the electronic document generation device 10 via the information communication network 11, and is stored in an information processing device such as a personal computer, a server device, or a database.
- the character string learning model 13 represents an information processing device in which the character string learning model 13 is stored.
- the character string learning model 13 may be configured by an existing learning model, or may be independently configured as a learning model suitable for use of the electronic document generation device 10.
- the character string learning model 13 is provided with learning models suitable for various languages such as Japanese, English, and Chinese, and in FIG. 1, the first character string learning model, the second character string learning model, and the third character string are provided. It shall be described as a learning model.
- the character string learning model 13 is not limited to the one connected to the information communication network 11, but may be included in the electronic document generation device 10 and used under the direct control of the device 10. Further, the character string learning model 13 may be distributed and stored in a plurality of information processing devices connected to the information communication network 11.
- the layout learning model 14 learns the correspondence between a plurality of elements included in the document image and the identification information of each of the plurality of elements based on the layout learning data described later, and recognizes the layout of the document image. It is a model and is used for layout recognition of the electronic document generator 10. Similar to the character string learning model 13, the layout learning model 14 has an arbitrary storage location as long as it can be used by the electronic document generator 10 via the information communication network 11, and the information connected to the information communication network 11 is available. It is stored in the processing device. For convenience of explanation of the present embodiment, the layout learning model 14 represents an information processing device in which the layout learning model 14 is stored.
- the layout learning model 14 includes a layout learning model for contracts, a layout learning model for invoices, a layout learning model for memorandums, a layout learning model for delivery notes, a layout learning model for receipts, and the like.
- the layout learning model for contracts is a learning model that recognizes the layout of document images of contracts, and learns using layout learning data for contracts.
- the layout learning model for contracts learns what kind of information is in what position in the contract, and in particular, it is described in bullet points, often without a table, and there is a handwritten signature column. Learn about the layout unique to books.
- the layout learning data for the contract is generated based on, for example, 200 types of contract forms, and at least 3 or 4 contract document images per form, to which the annotation described later is added. ..
- the layout learning model for invoices is a learning model that recognizes the layout of document images of invoices, and learns using layout learning data for invoices.
- the layout learning model for invoices learns what information is in what position on the invoice, and in particular, the table often occupies a large area, and even if it is written in Japanese, it is alphanumerical. Learn about invoice-specific layouts, such as not a few words.
- the layout learning data for invoices is generated based on, for example, 200 types of invoice forms, and at least 3 or 4 invoice document images per form, to which the annotation described later is added. ..
- the layout learning model for the memorandum is a learning model that recognizes the layout of the document image of the memorandum, and learns using the layout learning data of the memorandum.
- the layout learning model for the memorandum learns what kind of information is in which position of the memorandum, and in particular, it learns the layout peculiar to the memorandum such as the fact that there is often no table and there is a handwritten signature line.
- the layout learning data for the memorandum is generated based on, for example, 200 types of memorandum forms to which the annotation described later is added, and at least three or four memorandum document images per form.
- the layout learning model for the delivery note is a learning model that recognizes the layout of the document image of the delivery note, and learns using the layout learning data for the delivery note.
- the layout learning model for the delivery note learns what kind of information is in which position on the delivery note, and in particular, the range occupied by the table is often large, and the product name and product number are often described. Learn about the layout peculiar to the delivery note.
- the layout learning data for the delivery note is generated based on, for example, 200 types of delivery note forms with the annotation described later, and at least 3 or 4 delivery note document images per form. ..
- the layout learning model for receipts is a learning model that recognizes the layout of document images of receipts, and learns using layout learning data for receipts.
- the layout learning model for receipts learns what kind of information is in what position on the receipt, and in particular, it often contains a handwritten column of the amount or a table with the amount. Learn about the layout specific to receipts.
- the layout learning data for receipts is generated based on, for example, 200 types of receipt forms and at least 3 or 4 receipt document images per form, which are annotated as described below. ..
- the layout learning model 14 is not limited to the use of the electronic document generation device 10 via the information communication network 11, but may be included in the electronic document generation device 10. Further, the layout learning model 14 may be distributed and stored in a plurality of information processing devices connected to the information communication network 11.
- the document image database 15 is a database that stores images of documents.
- the electronic document generation device 10 acquires the document image stored in the document image database 15 and generates the character string learning data used for learning the character string learning model and the layout learning data used for learning the layout learning model. ..
- the user terminal 12 is used for operating the electronic document generator 10.
- the electronic document is modified according to the modification input from the user of the user terminal 12, and the electronic document generator 10 accepts the modification and relearns at least one of the character string learning model 13 and the layout learning model 14. ..
- FIG. 2 is a block diagram showing a mechanical configuration of the electronic document generator 10.
- the electronic document generator 10 includes an input / output interface 20, a communication interface 21, a Read Only Memory (ROM) 22, a Random Access Memory (RAM) 23, a storage unit 24, a Central Processing Unit (CPU) 25, and a Graphics processing unit (GPU). It has 28 mag.
- the input / output interface 20 sends / receives data or the like to / from an external device of the electronic document generation device 10.
- the external device is an input device 26 and an output device 27 that input / output data or the like to the electronic document generation device 10.
- the input device 26 is a keyboard, a mouse, a scanner, and the like
- the output device 27 is a monitor, a printer, a speaker, and the like.
- the communication interface 21 has a function of inputting / outputting data of the electronic document generation device 10 when communicating with the outside via the information communication network 11.
- the storage unit 24 can be used as a storage device, and various applications required for the electronic document generation device 10 to operate, various data used by the applications, and the like are recorded.
- the GPU 28 is suitable for a lot of repetitive operations performed in executing machine learning and the like, and is used together with the CPU 25.
- the electronic document generation device 10 stores the electronic document generation program described later in the ROM 22 or the storage unit 24, and takes the electronic document generation program into the main memory composed of the RAM 23 and the like.
- the CPU 25 accesses the main memory in which the electronic document generation program is incorporated and executes the electronic document generation program.
- FIG. 3 is a diagram showing an outline of processing performed by the electronic document generator 10.
- the electronic document generator 10 performs the following processes I to III in this order.
- Process I performs preprocessing 55 including "background removal”, “tilt correction”, and “shape adjustment” of the document image.
- the pre-processing 55 refers to performing pre-processing for facilitating the execution (recognition) of character recognition using a learning model for an image including a character string, and is the recognition processing performed in processes II and III. The purpose is to improve recognition accuracy.
- the layout recognition process 56 is performed.
- the layout recognition process 56 first, "layout recognition" of the document image is performed.
- the layout recognition process 56 is a process of recognizing what kind of information is present at which position in the input image.
- Information here refers to character strings, tables, images, seals, handwriting, etc.
- the electronic document generator 10 recognizes the layout of the document image and the document image contains a table, the electronic document generator 10 performs "table recognition” and “cuts out the cell image” for the cells included in the table. ..
- the character string recognition process 57 is a process of converting an image including a character string into text data by using a character string learning model 13 that has learned the correspondence between the image and the character string included in the image.
- the character string recognition process 57 may include processes such as “arrangement of text data” and “noise reduction”.
- the image of the character string is converted into text data, and "text data arrangement” and “noise removal” are performed. "Arrangement of text data” means that when the image of the cut out character string contains a space, the space is recognized together with the character string, so that the text data is arranged together with the space.
- Noise removal means that when noise is contained in the image of the cut out character string, the noise is passively removed from the text data because it is not recognized by the electronic document generator 10.
- the noise referred to here refers to pixels that do not form characters and are included in the image of the cut out character string.
- FIG. 4 is a block diagram showing a functional configuration of the electronic document generator 10.
- the electronic document generation device 10 has a document image acquisition unit 31, a preprocessing unit 32, a background removal unit 32a, an inclination correction unit 32b, a shape adjustment unit 32c, and a layout recognition on the CPU 25.
- the document image acquisition unit 31 acquires a document image in which a document is imaged.
- the document image acquisition unit 31 may acquire a document image from the document image database 15.
- the document image acquisition unit 31 may obtain a document image from the scanner of the input device 26.
- FIG. 5 is a diagram illustrating input data and output data of the electronic document generation device 10, and FIG. 5A shows a document image acquired by the document image acquisition unit 31 as input data.
- the document image contains noise such as a stapler mark 50, a handwriting 51, a seal 52, and an image 53.
- noises interfere with or become unnecessary for information processing devices such as humans and personal computers to understand the contents of the document.
- Other examples of noise include holes made for filing and creases left on the paper. The creases can be perceived as lines and need to be removed so that they are not reflected in the electronic document.
- the electronic document generation device 10 converts the character string in the document image into text data and outputs the electronic document while maintaining the layout of the acquired document image (see FIG. 5 (b)).
- the electronic document generator 10 removes noise 54 by active processing of the Heiliki mark 50, the handwriting 51, the seal 52, and the image 53 recognized as noise, and other pixels in the document image not recognized as character strings and noise. Is removed by passive processing that does not remain in the electronic document.
- the table in the document image of FIG. 5B is output together with the text data as object data in the electronic document while maintaining the arrangement in the document image.
- the electronic document generator 10 can arbitrarily select the elements to be included in the electronic document to be output.
- the stapler mark 50, the handwritten 51, the seal 52, the image 53, and the like are removed in normal use, but the seal 52 and the image 53 can be included in an electronic document and output as image data.
- the preprocessing unit 32 (see FIG. 4) performs preprocessing 55 on the document image acquired by the document image acquisition unit 31.
- the preprocessing 55 is performed in order to improve the recognition accuracy of image recognition using the learning model by the layout recognition unit 33 and the character string recognition unit 35, which will be described later.
- the pretreatment unit 32 includes a background removal unit 32a, an inclination correction unit 32b, and a shape adjustment unit 32c.
- the background removing unit 32a removes the background of the document image acquired by the document image acquisition unit 31.
- FIG. 6 is a diagram illustrating background removal performed in the pretreatment 55.
- FIG. 6A shows a document image 58a before the background is removed
- FIG. 6B shows a document image 58b after the background is removed.
- the background removing unit 32a removes the background of the document image by changing the background color of the document image to white. Specifically, the background removing unit 32a detects the background color of the acquired document image and determines whether or not the background color is white. When it is determined that the background color is not white, the background removing unit 32a extracts information other than the background of the document image, makes the background color white, and then superimposes the extracted information.
- the background removing unit 32a by deleting the background, noise that causes a malfunction of image recognition by the layout recognition unit 33 and the character string recognition unit 35 can be removed, and the recognition accuracy can be improved.
- the tilt correction unit 32b (see FIG. 4) corrects the tilt of the document image acquired by the document image acquisition unit 31.
- the processing performed by the tilt correction unit 32b will be described with reference to FIG. 7.
- FIG. 7 is a diagram illustrating the inclination correction performed in the preprocessing 55.
- FIG. 7A shows the document image 59a before the tilt correction
- FIG. 7B shows the document image 59b after the tilt correction.
- the tilt correction unit 32b corrects the tilt of the character string when there is a tilted character string in the document image, and makes the character string parallel or perpendicular to the writing direction.
- the tilt correction unit 32b corrects the tilted character string so as to be parallel to the vertical writing direction, and when the document image is written horizontally, the tilted character string is written horizontally. Correct so that it is parallel to the direction.
- the tilt correction unit 32b extracts the character string of the document image and determines whether or not there is a tilted character string in the extracted character string. When it is determined that there is a tilted character string in the extracted character string, the tilt correction unit 32b detects the tilt angle of the tilted character string with respect to the writing direction, and tilts the tilted character string with respect to the tilted character string. Rotation processing is performed so that the angle becomes zero.
- the recognition accuracy of image recognition by the character string recognition unit 35 can be improved by correcting the tilt of the character string. Further, it is possible to reduce the layout recognition error by the layout recognition unit 33.
- the shape adjusting unit 32c (see FIG. 4) adjusts the overall shape and size of the document image acquired by the document image acquisition unit 31.
- the processing performed by the shape adjusting unit 32c will be described with reference to FIG.
- FIG. 8 is a diagram illustrating shape adjustment performed in the pretreatment.
- FIG. 8A shows a document image 60a before the shape adjustment
- FIG. 8B shows a document image 60b after the shape adjustment.
- the shape adjustment unit 32c adjusts the overall shape of the document image based on the overall shape of the actual document. conduct. Specifically, when the overall aspect ratio of the document image acquired by the document image acquisition unit 31 is different from the overall aspect ratio of the actual document, the overall aspect ratio of the document image is the overall aspect ratio of the actual document.
- the shape adjusting unit 32c adjusts so as to be equal to the ratio.
- the shape adjustment unit 32c performs the subsequent processing. Adjusts the size of the document image acquired by the document image acquisition unit 31 so that
- the layout recognition unit 33 by adjusting the shape and size of the document image acquired by the document image acquisition unit 31, the layout recognition unit 33 that is performed thereafter improves the recognition accuracy of the layout according to the actual document. Further, the recognition accuracy of image recognition by the character string recognition unit 35 can be improved.
- the layout recognition unit 33 uses a layout learning model 14 that has learned the correspondence between the plurality of elements included in the document image 61 and the identification information of each of the plurality of elements, and the document image acquisition unit 33.
- the range of each of the plurality of elements included in the document image 61 acquired in 31 is specified in the document image 61, each type of the plurality of elements is recognized, and the document image 61 relating to each range of the plurality of elements is recognized. Get the position information in.
- the type of the element may be any of a character string 48, a table 49, an image 53, a seal 52, or a handwriting 51.
- the type of the element is not limited to this, and stapler marks 50, punch hole marks, breakage (tear) marks, carbon stains for copying, and the like may be used.
- the type of element may be suitable for the type of document (for example, contract, invoice, memorandum, invoice, receipt, etc.). For example, if carbon for copying is attached to the back side of the receipt and the carbon is transferred to the front surface and becomes a stain, use the stain with the carbon for copying as the element type and actively stain with the carbon for copying. May be removed.
- type of document for example, contract, invoice, memorandum, invoice, receipt, etc.
- the layout learning model 14 is either a layout learning model for contracts, a layout learning model for invoices, a layout learning model for memorandums, a layout learning model for delivery notes, or a layout learning model for receipts. It may be that.
- the types of elements may be classified into necessary and unnecessary according to the type of document.
- the layout recognition unit 33 does not acquire the position information of the element and requires the recognized element. If it corresponds to, the position information of the element may be acquired.
- the layout recognition unit 33 may recognize only the necessary elements among the plurality of elements included in the document image 61 and acquire the position information of the elements.
- the layout recognition unit 33 may overlap the elements or the elements may be too far apart from each other. , Correct the range of each of the elements and the acquired position information based on the actual document.
- FIG. 9A and 9B are diagrams for explaining the correction process for eliminating the omission in the layout recognition process, FIG. 9A shows a state before the correction, and FIG. 9B shows a state after the correction.
- the layout recognition unit 33 When the layout recognition unit 33 recognizes the image 70 of the character string included in the document image acquired by the document image acquisition unit 31 as a character string, the layout recognition unit 33 determines whether or not there is a gap in the recognition range, and determines whether or not the recognition range is missing. If there is, perform correction processing to add the missing part.
- FIG. 9A shows how the layout recognition unit 33 recognizes the image 70 of the character string as a character string in the recognition range 72a.
- the recognition range 72a has a defect in the left end portion of the image 70 of the character string.
- the layout recognition unit 33 determines whether or not there is a black line within a predetermined range around the recognition range 72a, and if there is a black line, a correction is added to add the range 72b including the black line to the recognition range 72a. (See FIG. 9 (b)).
- the determination of presence / absence performed by the layout recognition unit 33 is not limited to the black line, and whether or not a line having the same color as the character or a line having a preset color is within a predetermined range around the recognition range 72a. May be determined. This is because the main purpose of the correction process for eliminating omissions performed in the layout recognition process is to improve the recognition accuracy of the character recognition process performed thereafter.
- the character string recognition unit 35 can normally perform character recognition.
- FIG. 10A and 10B are diagrams for explaining the correction process for eliminating the overlap performed in the layout recognition process, FIG. 10A shows a state before the correction, and FIG. 10B shows a state after the correction.
- the layout recognition unit 33 recognizes the image 73 of the character string included in the document image acquired by the document image acquisition unit 31 as a character string, the recognition range 75a overlaps with another element (for example, Table 74). It is determined whether or not the overlap occurs, and if an overlap occurs, a correction process for eliminating the overlap is performed.
- another element for example, Table 74
- FIG. 10A shows how the layout recognition unit 33 recognizes the image 73 of the character string as a character string in the recognition range 75a.
- the recognition range 75a overlaps the table 74 to the right of the image 73 of the character string beyond a blank (space).
- the layout recognition unit 33 determines whether or not there is a blank (space) of a predetermined size inside the recognition range 75a, and if there is the blank (space), the blank (space) and the blank.
- the recognition range 75a related to the portion on the right side of (space) is deleted to make the recognition range 75b (see FIG. 10B).
- the layout recognition unit 33 Since there is always a blank (space) of a predetermined size between an element and another element, the layout recognition unit 33 recognizes the blank (space) of a predetermined size inside the recognition range. We conclude that the range overlaps with other elements. According to the overlap elimination correction process performed in the layout recognition process, the layout recognition unit 33 can improve the layout recognition accuracy.
- FIG. 11A and 11B are diagrams for explaining the layout recognition performed in the layout recognition process 56, FIG. 11A shows the state of the document image 61 before the layout is recognized, and FIG. 11B shows the state of the document image 61 after the layout is recognized. The state of the document image 62 of the above is shown.
- the layout recognition unit 33 specifies the range of the elements (character string 48, table 49, seal 52, image 53) included in the document image 61 within the document image 61 by image recognition using the layout learning model 14.
- the range of the specified character string 48 is surrounded by a solid line, and the ranges of the specified table 49, the seal 52, and the image 53 are surrounded by a broken line.
- the boundaries of the elements need not be visible to humans as long as they can be recognized by the electronic document generator 10.
- the layout recognition unit 33 recognizes the type of the corresponding element by image recognition using the layout learning model 14 in the range in the specified document image 61, and relates to the document image 62 in the range together with the type of the element. Get location information.
- the position information may be represented by plane orthogonal coordinates with a predetermined point in the document image 62 as the origin.
- the layout learning model 14 is preset according to the type of the document image 61, and the layout recognition unit 33 recognizes the layout of the document image 61 using the preset layout learning model 14.
- the document image 61 acquired by the document image acquisition unit 31 is a contract
- image recognition is performed using the layout learning model 14 for the contract
- Image recognition is performed using 14, and if it is a memorandum, image recognition is performed using the layout learning model 14 for the memorandum, and if it is a delivery note, image recognition is performed using the layout learning model 14 for the delivery note. If it is a receipt, image recognition is performed using the layout learning model 14 for the receipt.
- the layout recognition unit 33 properly uses the layout learning model 14 according to the type of the document image 61 acquired by the document image acquisition unit 31, the recognition accuracy of the layout recognition of the document image 61 can be improved.
- the cutout unit 34 cuts out each of the cells in the table included in the element in the element whose type recognized by the layout recognition unit 33 corresponds to the table, and in each document image of the cell. Get the location information in.
- FIG. 12A and 12B are diagrams for explaining table recognition performed by the layout recognition process 56
- FIG. 12A shows Table 63 before being recognized by the layout recognition unit 33
- FIG. 12B shows the layout recognition unit 33.
- Table 64 after being recognized in.
- the line recognized as the vertical line 65 is represented as a one-dot chain line
- the line recognized as the horizontal line 66 is represented as a broken line.
- the layout recognition unit 33 recognizes the length and position of each of the vertical lines 65 and the horizontal lines 66 constituting the table 64.
- the layout recognition unit 33 recognizes all the cells included in the table 64 by recognizing the lengths and positions of all the vertical lines 65 and the horizontal lines 66 constituting the table 64. That is, the layout recognition unit 33 recognizes a quadrangle composed of two adjacent vertical lines 65 and two adjacent horizontal lines 66 as cells.
- the layout recognition unit 33 also recognizes the line types of the lines constituting Table 64.
- the recognized line type is reflected in the line object constituting the table included in the electronic document when the electronic document is reproduced based on the acquired document image. Therefore, for example, when the table line in the document image 62 is a broken line, the table line included in the electronic document reproduced based on the document image 62 is represented as a broken line object.
- the cutout unit 34 cuts out all the cells included in the table 64 grasped by the layout recognition unit 33 into an image for each cell alone. With reference to FIG. 13, cutting out of cell pixels by the cutting-out portion 34 will be described.
- FIG. 13 is a diagram illustrating cutting out of a cell image.
- the cell 67 cut out by the cutout portion 34 may include a plurality of character strings.
- the cutting unit 34 acquires the image of each cell and the position information of the cell in the table 64 for all the cells included in the table 64.
- the position information may be represented by plane orthogonal coordinates with a predetermined point in Table 64 as the origin, or may be represented by (rows, columns) in Table 64.
- the cutout unit 34 reproduces all the vertical lines and horizontal lines constituting the table recognized by the layout recognition unit 33, and generates the position information of all the cells.
- FIG. 14 is a diagram illustrating a character string in the cell image.
- the cutout unit 34 When the cut out cell 67 contains a character string of a plurality of lines, the cutout unit 34 further cuts out an image for each character string for all the character strings.
- the cell 67 shown in FIG. 14 contains two lines of character strings, and the cutout portion 34 cuts out an image 67a of the character string and an image 67b of the character string.
- the character string recognition unit 35 uses a character string learning model 13 that has learned the correspondence between the document image and the character string included in the document image, and the document image acquired by the document image acquisition unit 31. Characters are recognized for the character string included in, and text data related to the character string is generated.
- the character string recognition unit 35 may recognize the character string included in the range recognized by the layout recognition unit 33 using the character string learning model 13 and generate text data related to the character string.
- the character string recognition unit 35 may perform character recognition using the character string learning model 13 for the character strings included in each of the cells cut out by the cutout unit 34, and generate text data related to the character strings. ..
- the character string recognition unit 35 includes a plurality of character string learning models 13, and may use a character string learning model 13 adapted to the language of the character string included in each of the plurality of elements.
- the recognition accuracy can be improved by using a character string learning model suitable for recognizing an English character string.
- FIGS. 15 and 16 are diagrams for explaining the arrangement of text data performed in the character string recognition process 57
- FIG. 15A is an image 67a of a character string before character recognition is performed
- FIG. 15B is a character. It is a character string 68a after recognition, that is, text data 68a.
- FIG. 16A and 16B are diagrams for explaining noise removal performed by the character string recognition process 57
- FIG. 16A is an image 71a of a character string before character recognition is performed
- FIG. 16B is a character recognition. It is the character string 71b after it is performed, that is, the text data 71b.
- the image 67a of the character string shown in FIG. 15A has a handwritten check mark in addition to the character string of one line.
- the character string contains a space between words.
- the character string recognition unit 35 recognizes the entire image 67a of the character string by using the character string learning model 13 and generates text data.
- the character string recognition unit 35 recognizes characters for two words "L / C NO:”, “ILC18H000219", and a blank space between the two words in the image 67a of the character string, and 2
- the text data corresponding to the tokens and the text data corresponding to the blank space between the two tokens are generated (68a: see FIG. 15B). Therefore, since the character string recognition unit 35 also recognizes the space between words and phrases and converts them into text data, the two words and phrases can be arranged separately as in the image 67a.
- the character string recognition unit 35 recognizes the character of the image 67a of the character string, the handwriting check mark is not recognized and is not included in the text data. Therefore, the handwriting check mark is deleted from the output electronic document (the handwriting check mark is deleted). 68a: FIG. 15 (b). Therefore, noise such as handwriting check marks that are not the target of character recognition by the character string recognition unit 35 is passively removed from the electronic document.
- the character string recognition unit 35 recognizes the entire image 71a of the character string using the character string learning model 13 and generates text data.
- the character string recognition unit 35 recognizes characters for the entire image 71a of the character string, and generates the text data corresponding to the character string for the character string "autiated to act on behalf of the" (71b: FIG. 16B). reference).
- the noise contained in the image 71a of the character string is not the target of character recognition by the character string recognition unit 35, it is passively removed from the electronic document (71b: see FIG. 16B).
- the character string learning model 13 has FIGS. 15 (a) and 15 (b). ) And the data associated with FIG. 16 (a) and FIG. 16 (b) as teacher data, by learning a large number of characters from the image using such deep learning. Recognition can be realized.
- the character string recognition unit 35 acquires attribute data such as the size and typeface of the character included in the character string when recognizing the character string included in the images 67a and 71a by using the character string learning model 13. You may.
- the attribute data of this character is reflected as the attribute data of the text data output by the output unit 36 described later.
- the output unit 36 (see FIG. 4) outputs text data as text on an electronic medium.
- the output unit 36 may output each of the text data related to the plurality of elements as text on an electronic medium to each position information in the range related to the plurality of elements.
- the electronic medium is not limited to data electronically stored in a recording medium, but also includes data itself that can be handled by an information processing device such as a personal computer, not in a state of being stored in the recording medium.
- the position information of the element may be represented by plane orthogonal coordinates with a predetermined point in the document image 62 as the origin.
- the output unit 36 since the text data related to the plurality of elements is output based on the position information related to the element, noise is removed while maintaining the layout of the acquired document image 61, and the inside of the document image 61 is used. It is possible to convert the character string of the above into text data and output an electronic document.
- the output unit 36 may reflect the character attribute data acquired by the character string recognition unit 35 in the text data and output it to an electronic document.
- the electronic document generator 10 can reproduce attribute data such as character size and typeface included in the document image 61 as attribute data of text data included in the electronic document to be output. can.
- the layout learning data generation unit 40 (see FIG. 4) is a document image including a plurality of elements, and annotations associated with the types corresponding to each of the elements are attached to the elements, and the annotations are added. Data for layout learning is generated by accumulating a plurality of document images.
- the layout learning data is used for supervised learning of the layout learning model 14.
- the document image stored in the layout learning data may be given position information in each document image of the range related to the plurality of elements included in the document image together with the annotation.
- FIGS. 17 to 23 are diagrams showing an example of layout learning data to which annotations are added.
- the layout learning data generation unit 40 acquires a document image from the document image database 15, annotates the document image, and generates layout learning data.
- the user can manually generate the layout learning data without using the layout learning data generation unit 40.
- the document image acquired from the document image database 15 can be annotated by using the user terminal 12.
- layout learning data used for learning the layout learning model 14 for invoices will be described.
- Annotation symbols are added to each element so that the electronic document generator 10 can identify and classify character strings, tables, images, stamps, outer frames, and noises as elements included in the document image.
- Annotation symbol 76 of the character string is given to the element related to the character string, the character string is surrounded by a rectangular frame line, and the tag of "Text" is attached to the frame line as a mark.
- the portion surrounded by the rectangular frame line is learned by the layout learning model 14 as the range occupied by the elements related to the character string in the document image.
- Annotation symbol 77 of the table is given to the element related to the table, a rectangular frame line is superimposed on the outer frame of the table, and the tag of "Border Table" is attached to the frame line as a mark.
- the portion surrounded by the rectangular frame line is learned by the layout learning model 14 as the range occupied by the elements related to the table in the document image.
- An annotation symbol 78 of the image is attached to the element related to the image, a frame line indicating the annotation symbol is superimposed on the boundary line of the image, and the tag of "Image" is attached to the frame as a mark.
- Images shall include logos, marks, photographs, illustrations and the like.
- the portion surrounded by the frame line is learned by the layout learning model 14 as the range occupied by the elements related to the image in the document image.
- Annotation symbol 79 of the seal is given to the element related to the seal, a frame line indicating the annotation symbol is superimposed on the boundary line of the seal, and the tag of "Hun” is attached to the frame line as a mark.
- the portion covered with the frame line is learned by the layout learning model 14 as the range occupied by the element related to the seal in the document image.
- Annotation symbol 80 of the outer frame is given to the element related to the outer frame, the frame line is superimposed on the boundary line of the outer frame, and the tag of "Border" is attached to the frame line as a mark.
- the layout learning model 14 learns about the length and position of the four line segments constituting the frame line.
- a noise annotation symbol 81 is added to the element related to noise, the noise is surrounded by a rectangular frame, and the tag of "Noise" is attached to the frame as a mark.
- the portion covered with the frame is learned by the layout learning model 14 as the range occupied by the element related to the noise in the document image.
- the layout learning data used for learning table recognition will be described with reference to FIG.
- the alternate long and short dash line which is the vertical line annotation symbol 83, is superimposed on all the vertical lines constituting the table, and the horizontal line annotation symbol is applied to all the horizontal lines constituting the table. Overlay the dashed line 84.
- the layout learning model 14 can learn about the size of the table, the range occupied by the table, the position, and the information of all the cells included in the table.
- the cell information is the number of cells contained in the table and the position of each cell in the table, and the position in the table is represented by (row, column) of the table.
- FIG. 20 is layout learning data in which a character string is included in each cell.
- FIG. 21 is layout learning data for recognizing a table relating to a cell containing a one-line character string, a cell containing a two-line character string, and a cell containing a three-line character string.
- a character string annotation symbol 76 is added to each of the character strings without being affected by the number of rows of the character string contained in one cell, and the character string is bounded by a rectangular border. Enclose with and attach the tag of "Text" to the frame line as a mark.
- the layout learning model 14 learns the range of the annotation symbol 76 of the character string and the position of the character string in the table.
- the electronic document generator 10 can reproduce the table by outputting the text data related to the character string to the electronic document together with the object data related to all the vertical lines and the horizontal lines constituting the table.
- An annotation symbol 76 for the character string is added to each of the character strings included in the cells of the table of FIG. 22, the character string is surrounded by a rectangular frame line, and the tag "Text" is attached to the frame line as a mark.
- the layout learning model 14 learns about the range of the annotation symbol 76 of the character string and the position information of the character string in the document.
- the electronic document generator 10 can reproduce a table in an electronic document by placing text data related to a character string at a position in the document.
- the electronic document generator 10 can reproduce the table in the electronic document only by outputting the text data without reproducing the vertical lines and the horizontal lines constituting the table in the electronic document.
- the layout learning model 14 can learn the range and position of the seal by the element related to the character string and the blank located at the bottom of the character string without using the element related to the seal. can.
- the layout learning data correction unit 41 (see FIG. 4) has the position information in the document image of each type of the plurality of elements acquired by the layout recognition unit 33 and the range of each of the plurality of elements based on the input. At least one of the above is modified, and the layout learning data is updated by adding this modified data.
- the document image 61 before the image is recognized by the layout recognition unit 33 There may be a discrepancy between the document image 61 before the image is recognized by the layout recognition unit 33 and the document image 62 after the image is recognized by the layout recognition unit 33.
- a part of the character string may not be recognized, an element to be recognized as an image may be recognized as a seal, or the position of the table may be misaligned.
- the document image 62 after the image is recognized by the layout recognition unit 33 is modified so as to match the document image 61 before the image is recognized by the layout recognition unit 33, and the corrected data is used.
- the layout learning data is updated.
- the layout learning unit 42 (see FIG. 4) relearns the layout learning model 14 using the layout learning data updated by the layout learning data correction unit 41. By re-learning the layout learning model 14, the recognition accuracy of the layout of the document image can be improved.
- the character string learning data generation unit 43 (see FIG. 4) generates character string learning data used for supervised learning of the character string learning model 13.
- the character string learning data correction unit 44 (see FIG. 4) corrects the text data generated by the character string recognition unit 35 based on the input, and adds the corrected text data for character string learning. Update the data.
- the character string learning unit 45 (see FIG. 4) relearns the character string learning model 13 using the character string learning data updated by the character string learning data correction unit 44.
- the character string learning data generation unit 43 acquires a document image from the document image database 15, annotates the document image, and generates character string learning data.
- the user can manually generate the character string learning data without using the character string learning data generation unit 43.
- the document image acquired from the document image database 15 can be annotated by using the user terminal 12.
- FIG. 21 is a diagram showing an example of character string learning data to which annotations are added.
- FIG. 24 is an output screen of the character string learning data generation unit 43, which is displayed on the user terminal 12 or the output device 27 of the electronic document generation device 10.
- the character string learning data generation unit 43 assigns text data corresponding to each of the character strings to the character strings included in the document image acquired from the document image database 15 as the comment 85 of the text data.
- the annotation may be added as the text data annotation 85 instead of the text data.
- the character string learning data generation unit 43 When the character string included in the document image contains a blank, the character string learning data generation unit 43 generates the character string learning data so that the text data corresponding to the character string also contains a blank.
- FIG. 25 is a flowchart of an electronic document generation program.
- the electronic document generation method is executed by the CPU 25 of the electronic document generation device 10 based on the electronic document generation program.
- the electronic document generation program realizes various functions such as a document image acquisition function, a preprocessing function, a layout recognition function, a cutting function, a character recognition function, and an output function for the CPU 25 of the electronic document generation device 10. These functions are executed in the order shown in FIG. 25, but the order may be changed as appropriate. Since each function overlaps with the description of the electronic document generation device 10 described above, a detailed description thereof will be omitted.
- the document image acquisition function acquires a document image obtained by converting a document into an image (S31: document image acquisition step).
- the format of the document image includes, for example, PDF, JPG, GIF, and the like, and other data formats that the electronic document generator 10 can process as an image may be included.
- the pre-processing function performs pre-processing on the document image acquired by the document image acquisition function (S32: pre-processing step).
- the pre-processing function has a background removal function, a tilt correction function, and a shape adjustment function.
- the background removal function removes the background of the document image acquired by the document image acquisition function
- the tilt correction function removes the background of the document image acquired by the document image acquisition function.
- the tilt of the image is corrected
- the shape adjustment function adjusts the overall shape and size of the document image acquired by the document image acquisition function.
- the layout recognition function uses a layout learning model 14 that learns the correspondence between a plurality of elements included in the document image and the identification information of each of the plurality of elements to obtain a document image acquired by the document image acquisition function.
- the range in each document image of the plurality of elements included is specified, each type of the plurality of elements is recognized, and the position information in the document image relating to each range of the plurality of elements is acquired (S33:).
- Layout recognition step ).
- the types of elements may be classified into necessary and unnecessary according to the type of document.
- the layout recognition function does not acquire the position information of the element and recognizes it. If the specified element corresponds to a necessary one, the position information of the element may be acquired.
- the layout recognition function may recognize only the necessary elements among the plurality of elements included in the document image 61 and acquire the position information of the elements.
- the layout recognition function recognizes each type of element, acquires the position information of the document image related to each range of the element, and then if the elements overlap or the elements are too far apart, the layout recognition function recognizes each type of the element. Based on the actual document, the range of each of the elements and the acquired position information are corrected.
- the layout recognition function recognizes the length and position of each of the vertical and horizontal lines that make up the table.
- the layout recognition function grasps all the cells included in the table by grasping the lengths and positions of all the vertical lines and horizontal lines constituting the table. That is, the layout recognition function recognizes a quadrangle composed of two adjacent vertical lines and two adjacent horizontal lines as cells.
- the layout recognition function also recognizes the line types of the lines that make up the table.
- the recognized line type is reflected in the line object constituting the table included in the electronic document when the electronic document is reproduced based on the acquired document image.
- the table line in the document image is a dashed line
- the table line contained in the electronic document reproduced based on the document image is represented as a dashed object.
- the cutout function cuts out each of the cells in the table included in the element in the element whose type recognized by the layout recognition function corresponds to the table, and acquires the position information in each document image of the cell ( S34: Cutting step).
- the cutout function reproduces all the vertical and horizontal lines constituting the table recognized by the layout recognition function, and generates the position information of all the cells.
- the cell cut out by the cutout function may contain multiple character strings.
- the cutout function further cuts out an image for each character string for all the character strings.
- the image of the character string recognized by the layout recognition function and the image of the character string cut out by the cutout function are sent to the character recognition function line by line.
- the character recognition function uses a character string learning model that learns the correspondence between the document image and the character string included in the document image, and recognizes the character string included in the document image acquired by the document image acquisition function. , Generates text data related to the character string (S35: character recognition step).
- the output function outputs the text data as text on an electronic medium (S36: output step).
- the output function outputs text data based on the position information of the character string acquired by the layout recognition function and the position information in the document image of the cell acquired by the cutout portion, and reproduces the text as text on an electronic medium.
- FIGS. 26 to 28 are flowcharts of an embodiment relating to an electronic document generation program.
- the flowcharts shown in FIGS. 26 to 28 show a flowchart of one electronic document generation program by combining them.
- step S102 the document image acquisition unit 31 acquires a document image or PDF from the document image database 15.
- step S103 it is determined whether or not the data acquired by the document image acquisition unit 31 is PDF. If it is not a PDF (No: S103), that is, if the data acquired by the document image acquisition unit 31 is a document image, the process proceeds to step S106.
- step S104 the PDF is converted into a document image, and then the document image is acquired (S105).
- the preprocessing unit 32 performs preprocessing on the acquired document image.
- the pretreatment unit 32 includes a background removal unit 32a, an inclination correction unit 32b, and a shape adjustment unit 32c.
- the background removing unit 32a removes the background of the acquired document image.
- the tilt correction unit 32b corrects the tilt and corrects the tilt of the character string.
- the shape adjusting unit 32c adjusts the overall shape and size of the acquired document image.
- step S107 the layout recognition unit 33 acquires a document image that has undergone preprocessing performed by the preprocessing unit 32.
- the acquired document image after preprocessing is sent to the document image cutting process of step S115, step S120, and step S136 described later.
- the layout recognition unit 33 performs layout recognition of the document image, specifies the range of a plurality of elements included in the document image for each element, and acquires the type and position information for each element. do.
- the types of elements are character strings, tables, images, seals, and handwriting.
- the layout recognition unit 33 adjusts the position information of the minimum boundary box of the acquired element.
- the minimum boundary box means the rectangle surrounding the element and having the smallest area, and means the range occupied by the element.
- the layout recognition unit 33 collates the document image with the acquired element, and if there is a discrepancy between the document image and the position information of the acquired element, the layout recognition unit 33 adjusts the position information of the minimum boundary box of the acquired element. ..
- the layout recognition unit 33 acquires the layout information after the adjustment process of the minimum boundary box performed in step S110.
- the layout information includes element types and position information.
- the layout recognition unit 33 refers to the layout information of the internally stored element sent by the process of step S130 described later, and determines whether or not other elements remain in the document image. do.
- step S130 When the layout information of all the elements is included in the layout information of the internally stored elements sent by the process of step S130, the layout recognition unit 33 has other elements remaining in the document image. It is determined that there is no such thing (No: S112), the process proceeds to step S131, the loop termination processing of step S112 to step S130 is performed, and the process proceeds to step S132.
- the layout recognition unit 33 has other elements in the document image. It is determined that it remains (Yes: S112), and the process proceeds to step S113.
- step S113 the layout recognition unit 33 determines whether or not the element remaining in the document image is a table. If no table remains in the document image (No: S113), layout information other than the table is sent to step S130 described later.
- step S113 If the table remains in the document image (Yes: S113), the process proceeds to step S114. Since the document image is related to the receipt, it often includes a table. Therefore, if it is determined that the document image does not include the table, the layout recognition unit 33 may interrupt the process and confirm whether or not the electronic document relates to a receipt.
- step S114 the layout recognition unit 33 acquires the size and position information of all the vertical lines and horizontal lines constituting the table in the document image. If the size and position information of all the vertical lines and the horizontal lines constituting the table are acquired, the size and position of the cells can be acquired for all the cells included in the table.
- step S115 the cutout unit 34 cuts out a table image from the preprocessed document image acquired by the process of step S107.
- step S116 the cutout unit 34 acquires an image of the table cut out in step S115.
- step S117 and step S118 the cutting unit 34 performs a process of extracting cells from the image of the table acquired in step S116 (step S117), and acquires cell information (step S118).
- the cell information is the row, column, and coordinates corresponding to the cell position information in the table.
- the cell information acquired in step S118 is sent to step S127, which will be described later.
- step S119 the cutting unit 34 refers to the layout information of the internally stored table sent by the process of step S127, and determines whether or not other cells remain in the table.
- step S127 When the layout information of all the cells is included in the layout information of the internally stored cells sent by the process of step S127, the cutout unit 34 has no other cells left in the table. (No: S119), the process proceeds to step S128, the loop end processing of step S119 is performed, and the process proceeds to step S130.
- step S127 when the layout information of all the cells is not included in the layout information of the internally stored cells sent by the process of step S127, the cutout portion 34 has other cells remaining in the table. (Yes: S119), and the process proceeds to step S120.
- step S120 the cutting unit 34 performs a process of cutting out a cell image from the preprocessed document image acquired by the process of step S107.
- step S121 the cutting unit 34 acquires an image of the cell cut out by the process of step S120.
- step S122 the character string recognition unit 35 performs a character string recognition process on the image of the cell acquired by the process of step S121.
- step S123 the character string recognition unit 35 acquires the position information of the character string for which the character string recognition process has been performed.
- step S124 the character string recognition unit 35 adjusts the position information of the minimum boundary box of the character string acquired by the process of step S123.
- the character string recognition unit 35 collates the document image with the position information of the acquired character string, and if there is a discrepancy between the document image and the position information of the acquired character string, the minimum boundary box of the acquired character string. Adjust the position information of.
- step S125 the character string recognition unit 35 acquires the position information after the adjustment process of the position information of the minimum boundary box of the character string carried out in step S124.
- step S126 and step S127 the character string recognition unit 35 merges the cell information acquired by the process of step S118 and the position information of the adjusted character string acquired by the process of step S125 (step S127).
- step S126 Internally stored in the internal storage device as table layout information (step S126).
- the internal storage device refers to either or both of the RAM 23 and the storage unit 24 shown in FIG.
- steps S119 to S127 is performed for all cells included in the table. After the processing of steps S119 to S127 is performed on the last cell included in the table, the loop termination processing of step S128 is performed, and the character string recognition unit 35 shifts to step S130.
- step S129 and step S130 the output unit 36 merges the layout information of the table acquired by the process of step S126 and the layout information other than the table acquired by the process of step S113 (step S130), and all the elements. Is internally stored in the internal storage device as the layout information of (step S129).
- steps S112 to S130 are performed for all the elements included in the document image. After the processing of steps S112 to S130 is performed for the last element included in the document image, the loop end processing of step S131 is performed, and the character string recognition unit 35 shifts to step S132.
- step S132 the character string recognition unit 35 determines whether or not other elements remain in the document image.
- the character string recognition unit 35 refers to the layout information of the internally stored elements sent by the process of step S140 described later, and determines whether or not other elements remain in the document image.
- step S140 When the layout information of all the elements is included in the layout information of the internally stored elements sent by the process of step S140, the character string recognition unit 35 has another degree in the document image. It is determined that there is no remaining (No: S132), the process proceeds to step S141, the loop termination process of step S132 to step S140 is performed, and the process proceeds to step S142.
- step S140 when the layout information of all the elements is not included in the layout information of the internally stored elements sent by the process of step S140, the character string recognition unit 35 has another element in the document image. Is determined to remain (Yes: S132), and the process proceeds to step S133.
- step S133 the character string recognition unit 35 determines whether or not the element remaining in the document image is a character string.
- the process proceeds to step S135.
- step S133 When the character string recognition unit 35 determines that the element remaining in the document image is not a character string (No: S133), the loop continuation process of shifting to step S132 is performed (step S134). In step S135, the character string recognition unit 35 acquires the position information of the character string.
- step S136 and step S137 the character string recognition unit 35 cuts out an image of the character string from the preprocessed document image acquired by the process of step S107 (step S136), and acquires the image of the character string.
- step S138 and step S139 the character string recognition unit 35 performs a character string recognition process on the image of the character string acquired by the process of step S137 (step S138), and the text data predicted by the character string recognition process. Is generated (step S139).
- step S140 the character string recognition unit 35 merges the position information of the character string acquired by the process of step S135 and the text data generated by the process of step S139 to generate the layout information of the element.
- the layout information of the generated element is sent to step S129.
- step S129 the layout information of the sent elements is internally stored in the internal storage device.
- the internal storage device refers to either or both of the RAM 23 and the storage unit 24 shown in FIG.
- step S132 to step S140 is performed until it is determined by the processing of step S132 that the layout information of all the elements is included in the layout information of the elements sent by the processing of step S140.
- step S141 in response to the determination by the process of step S132 that the layout information of all the elements is included in the layout information of the elements sent by the process of step S140, steps S132 to S140 The process of ending the loop up to is performed, and the process proceeds to step S142.
- step S142 the electronic document generator 10 performs post-processing.
- the text data, images, and position information of all the elements are output to JSON (Javascript objectionation) and converted to TSV (Tab-Separated Values).
- JSON Javascript objectionation
- TSV Tab-Separated Values
- the output unit 36 has a simple text file as a final form, an HTML (HyperText Markup Language), a file format editable by commercially available character editing software, and a file format in which the information of all the elements that have undergone post-processing can be edited. Editable is output as an electronic document such as PDF.
- HTML HyperText Markup Language
- the electronic document generator 10 recognizes the layout of the document image using the layout learning model 14, and then performs character recognition of the document image using the character string learning model 13. That is, since the electronic document generation device 10 identifies the types of a plurality of elements included in the document image and performs character recognition suitable for the types of elements, the recognition accuracy of character recognition can be improved.
- the electronic document generator 10 uses the character string learning model 13 to characterize a document image, as compared with character recognition in character units, which has been performed by conventional OCR text recognition technology. Since the recognition is performed for each character string, the recognition efficiency at the time of character recognition can be improved.
- the character recognition is performed for each character string instead of the character recognition for each character, so that noise existing on the characters and the like are generated.
- Character recognition can be performed by suppressing the influence of the above, and the recognition accuracy of character recognition can be improved as compared with character recognition performed in units of one character.
- a character that is erroneously recognized by character recognition using the conventional OCR text recognition technique can be correctly recognized by character recognition using the character string learning model 13.
- the character when a seal is superimposed on a character, the character may be erroneously recognized by the conventional OCR text recognition technique, but can be correctly recognized by character recognition using the character string learning model 13. ..
- the electronic document generator 10 recognizes the elements whose types correspond to the table for each character string included in the image related to the cell alone, and therefore the characters included in the table. It is possible to improve the recognition accuracy of character recognition in a column.
- the character string learning model 13 and the layout learning model 14 are learned from the annotated character string learning data and the layout learning data, so that the layout recognition unit 33 and the character string recognition unit 35 are learned. It is possible to improve the recognition accuracy of.
- the present disclosure is not limited to the electronic document generator 10 according to the above-described embodiment, and is carried out by various other modifications or applications as long as it does not deviate from the gist of the present disclosure described in the claims. It is possible.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Artificial Intelligence (AREA)
- Databases & Information Systems (AREA)
- Character Input (AREA)
- Character Discrimination (AREA)
- Image Analysis (AREA)
- Electrically Operated Instructional Devices (AREA)
Abstract
Description
記憶部24は、記憶装置として利用でき、電子文書生成装置10が動作する上で必要となる各種アプリケーション及び当該アプリケーションによって利用される各種データなどが記録される。GPU28は、機械学習などを実行する上で行われる繰り返し演算を多用する場合に適しており、CPU25とともに用いる。 The communication interface 21 has a function of inputting / outputting data of the electronic
The
電子文書生成装置10は、次に述べる処理I~IIIをこの順に行う。 Next, with reference to FIG. 3, the outline of the processing performed by the
The
前処理55とは、文字列を含む画像に対して、学習モデルを用いた文字認識を実行(認識)しやすくするための事前の処理を行うことをいい、処理II、IIIで行う認識処理の認識精度を向上させることを目的とする。 Process I performs preprocessing 55 including "background removal", "tilt correction", and "shape adjustment" of the document image.
The pre-processing 55 refers to performing pre-processing for facilitating the execution (recognition) of character recognition using a learning model for an image including a character string, and is the recognition processing performed in processes II and III. The purpose is to improve recognition accuracy.
レイアウト認識処理56では、先ず文書画像の「レイアウト認識」を行う。レイアウト認識処理56とは、入力された画像内で、どの位置に、どのような情報があるのかを認識する処理である。 In the process II, the
In the
文字列認識処理57とは、文字列を含む画像を、画像と画像に含まれる文字列との対応関係を学習した文字列学習モデル13を用いて、テキストデータに変換する処理のことである。文字列認識処理57は、「テキストデータの配置」及び「ノイズ除去」などの処理を含むものとしてもよい。 In the process III, the character
The character
「テキストデータの配置」とは、切出した文字列の画像にスペースが含まれている場合には文字列とともにスペースも一緒に認識されるので、テキストデータはスペースとともに配置されることを指す。 In the character
"Arrangement of text data" means that when the image of the cut out character string contains a space, the space is recognized together with the character string, so that the text data is arranged together with the space.
文書画像取得部31は、文書画像データベース15から文書画像を取得してもよい。或いは、文書画像取得部31は、入力装置26のスキャナーから文書画像を入手してもよい。 The document image acquisition unit 31 (see FIG. 4) acquires a document image in which a document is imaged.
The document
図5は電子文書生成装置10の入力データと出力データとを説明する図であり、図5(a)は入力データとして文書画像取得部31に取得された文書画像を示す。当該文書画像には、ホチキス跡50、手書き51、印章52、及び画像53などのノイズが存在する。 With reference to FIG. 5, the document image acquired by the document
FIG. 5 is a diagram illustrating input data and output data of the electronic
前処理55は、後述するレイアウト認識部33及び文字列認識部35による、学習モデルを用いる画像認識の認識精度を向上させるために行われる。 The preprocessing unit 32 (see FIG. 4) performs preprocessing 55 on the document image acquired by the document
The
背景除去部32a(図4参照)は、文書画像取得部31が取得した文書画像の背景を除去する。 The
The background removing unit 32a (see FIG. 4) removes the background of the document image acquired by the document
図7を参照して、傾き補正部32bにより行われる処理について説明する。図7は、前処理55で行う傾き補正を説明する図である。図7(a)は傾き補正される前の文書画像59aを示し、図7(b)は傾き補正された後の文書画像59bを示す。 The tilt correction unit 32b (see FIG. 4) corrects the tilt of the document image acquired by the document
The processing performed by the tilt correction unit 32b will be described with reference to FIG. 7. FIG. 7 is a diagram illustrating the inclination correction performed in the
図8を参照して、形状調整部32cにより行われる処理について説明する。図8は、前処理で行う形状調整を説明する図である。図8(a)は形状調整される前の文書画像60aを示し、図8(b)は形状調整された後の文書画像60bを示す。 The shape adjusting unit 32c (see FIG. 4) adjusts the overall shape and size of the document image acquired by the document
The processing performed by the shape adjusting unit 32c will be described with reference to FIG. FIG. 8 is a diagram illustrating shape adjustment performed in the pretreatment. FIG. 8A shows a
この場合、レイアウト認識部33は、文書画像61に含まれる複数の要素のうち、認識した要素が不要なものに該当する場合は当該要素の位置情報は取得されず、認識した要素が必要なものに該当する場合は当該要素の位置情報を取得することとしてもよい。または、レイアウト認識部33は、文書画像61に含まれる複数の要素のうち、必要な要素のみを認識し、当該要素の位置情報を取得することとしてもよい。 The types of elements may be classified into necessary and unnecessary according to the type of document.
In this case, if the recognized element corresponds to an unnecessary element among the plurality of elements included in the
図13を参照して、切出部34によるセル画素の切り出しについて説明する。図13は、セル画像の切り出しを説明する図である。切出部34により切り出されたセル67は、複数の文字列を含む場合もある。 The
With reference to FIG. 13, cutting out of cell pixels by the cutting-out
文字列認識部35は、英語で書かれた文書画像を文字認識する場合に、英語の文字列の認識に適した文字列学習モデルを用いることで、認識精度を向上させることができる。 The character
When the character
図15は、文字列認識処理57で行うテキストデータの配置を説明する図であり、図15(a)は文字認識が行われる前の文字列の画像67aであり、図15(b)は文字認識が行われた後の文字列68a、すなわちテキストデータ68aである。 The character recognition performed by the character
15A and 15B are diagrams for explaining the arrangement of text data performed in the character
従って、文字列認識部35は、字句と字句の間にあるスペースについても認識してテキストデータに変換するので、画像67aと同様に2つの字句を離して配置することができる。 The character
Therefore, since the character
出力部36は、複数の要素に係る範囲の各々の位置情報に、複数の要素に係るテキストデータの各々を電子媒体のテキストとして出力してもよい。 The output unit 36 (see FIG. 4) outputs text data as text on an electronic medium.
The
要素の位置情報は、文書画像62内の所定点を原点とした平面直交座標によって表されてもよい。 The electronic medium is not limited to data electronically stored in a recording medium, but also includes data itself that can be handled by an information processing device such as a personal computer, not in a state of being stored in the recording medium.
The position information of the element may be represented by plane orthogonal coordinates with a predetermined point in the
レイアウト学習用データに蓄積される文書画像に、アノテーションとともに文書画像に含まれる複数の要素に係る範囲の各々の文書画像内における位置情報が付与されてもよい。 The layout learning data is used for supervised learning of the
The document image stored in the layout learning data may be given position information in each document image of the range related to the plurality of elements included in the document image together with the annotation.
レイアウト学習モデル14は再学習されることで、文書画像のレイアウトの認識精度を向上させることができる。 The layout learning unit 42 (see FIG. 4) relearns the
By re-learning the
文字列学習用データ修正部44(図4参照)は、入力に基づいて、文字列認識部35により生成されたテキストデータが修正され、この修正されたテキストデータを追加することで文字列学習用データを更新する。 The character string learning data generation unit 43 (see FIG. 4) generates character string learning data used for supervised learning of the character
The character string learning data correction unit 44 (see FIG. 4) corrects the text data generated by the character
文書画像の形式は、一例としてPDF、JPG、及びGIFなどがあり、この他電子文書生成装置10が画像として処理できるデータ形式のものは含み得る。 The document image acquisition function acquires a document image obtained by converting a document into an image (S31: document image acquisition step).
The format of the document image includes, for example, PDF, JPG, GIF, and the like, and other data formats that the
前処理機能は背景除去機能、傾き補正機能、及び形状調整機能を備え、背景除去機能は文書画像取得機能が取得した文書画像の背景を除去し、傾き補正機能は文書画像取得機能が取得した文書画像の傾きを補正し、形状調整機能は文書画像取得機能が取得した文書画像の全体の形状及び大きさを調整する。 The pre-processing function performs pre-processing on the document image acquired by the document image acquisition function (S32: pre-processing step).
The pre-processing function has a background removal function, a tilt correction function, and a shape adjustment function. The background removal function removes the background of the document image acquired by the document image acquisition function, and the tilt correction function removes the background of the document image acquired by the document image acquisition function. The tilt of the image is corrected, and the shape adjustment function adjusts the overall shape and size of the document image acquired by the document image acquisition function.
この場合、レイアウト認識機能は、文書画像取得機能に取得された文書画像に含まれる複数の要素のうち、認識した要素が不要なものに該当する場合は当該要素の位置情報は取得されず、認識した要素が必要なものに該当する場合は当該要素の位置情報を取得することとしてもよい。または、レイアウト認識機能は、文書画像61に含まれる複数の要素のうち、必要な要素のみを認識し、当該要素の位置情報を取得することとしてもよい。 The types of elements may be classified into necessary and unnecessary according to the type of document.
In this case, if the recognized element corresponds to an unnecessary element among a plurality of elements included in the document image acquired by the document image acquisition function, the layout recognition function does not acquire the position information of the element and recognizes it. If the specified element corresponds to a necessary one, the position information of the element may be acquired. Alternatively, the layout recognition function may recognize only the necessary elements among the plurality of elements included in the
切出機能は、レイアウト認識機能により認識された表を構成する全ての縦線及び横線を再生し、全てのセルの位置情報を生成する。 The cutout function cuts out each of the cells in the table included in the element in the element whose type recognized by the layout recognition function corresponds to the table, and acquires the position information in each document image of the cell ( S34: Cutting step).
The cutout function reproduces all the vertical and horizontal lines constituting the table recognized by the layout recognition function, and generates the position information of all the cells.
出力機能は、レイアウト認識機能により取得された文字列の位置情報、及び切出部により取得されたセルの文書画像内における位置情報に基づいてテキストデータを出力し、電子媒体のテキストとして再生する。 The output function outputs the text data as text on an electronic medium (S36: output step).
The output function outputs text data based on the position information of the character string acquired by the layout recognition function and the position information in the document image of the cell acquired by the cutout portion, and reproduces the text as text on an electronic medium.
ステップS103において、文書画像取得部31が取得したデータがPDFか否かの判定を行う。PDFではない場合(No:S103)、即ち、文書画像取得部31が取得したデータが文書画像であった場合、ステップS106に移行する。 In step S102, the document
In step S103, it is determined whether or not the data acquired by the document
取得された前処理後の文書画像は、後述のステップS115、ステップS120、及びステップS136の文書画像切り出し処理に送られる。 In step S107, the
The acquired document image after preprocessing is sent to the document image cutting process of step S115, step S120, and step S136 described later.
要素の種類は、文字列、表、画像、印章、手書きである。 In step S108 and step S109, the
The types of elements are character strings, tables, images, seals, and handwriting.
最小境界ボックスとは、要素を囲う矩形のうち面積が最小のものをいい、当該要素が占める範囲を意味する。レイアウト認識部33は、文書画像と取得した要素とを照合し、文書画像と取得した要素の位置情報との間に齟齬があった場合は取得した要素の最小境界ボックスの位置情報の調整を行う。 In step S110, the
The minimum boundary box means the rectangle surrounding the element and having the smallest area, and means the range occupied by the element. The
文書画像に表が残っていない場合(No:S113)、後述のステップS130へ表以外のレイアウト情報を送る。 In step S113, the
If no table remains in the document image (No: S113), layout information other than the table is sent to step S130 described later.
ステップS116において、切出部34は、ステップS115にて切り出された表の画像を取得する。 In step S115, the
In step S116, the
ステップS118にて取得されたセルの情報は、後述するステップS127に送られる。 The cell information is the row, column, and coordinates corresponding to the cell position information in the table.
The cell information acquired in step S118 is sent to step S127, which will be described later.
ステップS121において、切出部34は、ステップS120の処理により切り出されたセルの画像を取得する。 In step S120, the cutting
In step S121, the cutting
ステップS123において、文字列認識部35は、文字列認識の処理が行われた文字列の位置情報を取得する。 In step S122, the character
In step S123, the character
文字列認識部35は、文書画像と取得した文字列の位置情報とを照合し、文書画像と取得した文字列の位置情報との間に齟齬があった場合は取得した文字列の最小境界ボックスの位置情報の調整を行う。 In step S124, the character
The character
文字列認識部35は、後述のステップS140の処理により送られてきた内部記憶された要素のレイアウト情報を参照して、文書画像の中に他の要素が残っているか否かを判定する。 In step S132, the character
The character
文字列認識部35が、文書画像に残っている要素が文字列であると判定した場合(Yes:S133)、ステップS135に移行する。 In step S133, the character
When the character
ステップS135において、文字列認識部35は、文字列の位置情報を取得する。 When the character
In step S135, the character
なお、上記した各機能部の処理は、電子文書生成装置10のCPU25により実行される処理である。 In step S142, the
The processing of each of the above-mentioned functional units is a processing executed by the
11 情報通信ネットワーク
12 ユーザ端末
13 文字列学習モデル
14 レイアウト学習モデル
15 文書画像データベース
20 入出力インターフェース
21 通信インターフェース
22 ROM
23 RAM
24 記憶部
25 CPU
26 入力装置
27 出力装置
28 GPU
31 文書画像取得部
32 前処理部
32a 背景除去部
32b 傾き補正部
32c 形状調整部
33 レイアウト認識部
34 切出部
35 文字列認識部
36 出力部
40 レイアウト学習用データ生成部
41 レイアウト学習用データ修正部
42 レイアウト学習部
43 文字列学習用データ生成部
44 文字列学習用データ修正部
45 文字列学習部
47 傾き補正前の文書画像
48 文字列
49 表
50 ホッチキス跡
51 手書き
52 印章
53 画像
54 ノイズ除去
55 前処理
56 レイアウト認識処理
57 文字列認識処理
58a、59a、60a 文書画像
58b、59b、60b 文書画像
61、62 文書画像
63、64 表
65 縦線
66 横線
67 セル画像
69、70、73 文字列の画像
71a 文字列の画像
71b テキストデータ
72 認識範囲
73 表
75 認識範囲
76 文字列の注釈記号
77 表の注釈記号
78 画像の注釈記号
79 印章の注釈記号
80 外枠の注釈記号
81 ノイズの注釈記号
82 手書きの注釈記号
83 縦線の注釈記号
84 横線の注釈記号
85 テキストデータの注釈
100 電子文書生成システム
S31 文書画像取得ステップ
S32 前処理ステップ
S33 レイアウト認識ステップ
S34 切出ステップ
S35 文字認識ステップ
S36 出力ステップ 10
23 RAM
24
26
31 Document
Claims (16)
- 文書を画像化した文書画像を取得する文書画像取得部と、
文書画像と当該文書画像に含まれる文字列との対応関係を学習した文字列学習モデルを用いて、
前記文書画像取得部に取得された前記文書画像に含まれる文字列を文字認識し、当該文字列に係るテキストデータを生成する文字列認識部と、
前記テキストデータを電子媒体のテキストとして出力する出力部と、
を備えることを特徴とする電子文書生成装置。 A document image acquisition unit that acquires a document image that is an image of a document,
Using a character string learning model that learned the correspondence between the document image and the character string included in the document image,
A character string recognition unit that recognizes a character string included in the document image acquired by the document image acquisition unit and generates text data related to the character string, and a character string recognition unit.
An output unit that outputs the text data as text on an electronic medium,
An electronic document generator, characterized in that it comprises. - 文書画像に含まれる複数の要素と、当該複数の要素の各々の識別情報との対応関係を学習したレイアウト学習モデルを用いて、
前記文書画像取得部に取得された前記文書画像に含まれる複数の要素の各々の前記文書画像内における範囲を特定し、前記複数の要素の各々の種類を認識し、前記複数の要素の各々の前記範囲に係る前記文書画像内における位置情報を取得するレイアウト認識部をさらに備え、
前記文字列認識部は、前記レイアウト認識部により認識された前記範囲に含まれる文字列について、前記文字列学習モデルを用いて文字認識し、前記文字列に係るテキストデータを生成し、
前記出力部は、前記複数の要素に係る前記範囲の各々の前記位置情報に、前記複数の要素に係る前記テキストデータの各々を電子媒体のテキストとして出力する、
ことを特徴とする請求項1に記載の電子文書生成装置。 Using a layout learning model that learned the correspondence between a plurality of elements included in a document image and the identification information of each of the plurality of elements,
The range of each of the plurality of elements included in the document image acquired by the document image acquisition unit within the document image is specified, each type of the plurality of elements is recognized, and each of the plurality of elements is recognized. Further, a layout recognition unit for acquiring position information in the document image related to the range is provided.
The character string recognition unit recognizes a character string included in the range recognized by the layout recognition unit using the character string learning model, and generates text data related to the character string.
The output unit outputs each of the text data related to the plurality of elements to the position information in each of the ranges related to the plurality of elements as text on an electronic medium.
The electronic document generator according to claim 1. - 前記要素の前記種類は、文字列、表、画像、印章、又は手書きのいずれかである、
ことを特徴とする請求項2に記載の電子文書生成装置。 The type of the element is either a string, a table, an image, a seal, or a handwriting.
The electronic document generator according to claim 2. - 前記レイアウト認識部により認識された前記種類が前記表に該当する前記要素において、当該要素に含まれる前記表の中のセルの各々を切り出し、前記セルの各々の前記文書画像内における位置情報を取得する切出部をさらに備え、
前記文字列認識部は、前記切出部に切り出された前記セルの各々に含まれる文字列について、前記文字列学習モデルを用いて文字認識を行い、前記文字列に係るテキストデータを生成する、
ことを特徴とする請求項3に記載の電子文書生成装置。 In the element whose type corresponds to the table recognized by the layout recognition unit, each of the cells in the table included in the element is cut out, and the position information of each of the cells in the document image is acquired. With more cutouts
The character string recognition unit recognizes a character string included in each of the cells cut out in the cutout unit using the character string learning model, and generates text data related to the character string.
The electronic document generator according to claim 3. - 複数の前記要素を含む文書画像であって、当該要素に当該要素の各々に該当する前記種類に関連付けられたアノテーションが付与されており、
前記アノテーションが付与された複数の前記文書画像を蓄積してレイアウト学習用データを生成するレイアウト学習用データ生成部をさらに備え、
前記レイアウト学習用データは前記レイアウト学習モデルの教師有り学習に用いられる、
ことを特徴とする請求項2ないし4のいずれか1項に記載の電子文書生成装置。 A document image containing a plurality of the elements, and the elements are annotated with the annotations associated with the types corresponding to each of the elements.
It further includes a layout learning data generation unit that accumulates a plurality of the document images to which the annotation is attached and generates layout learning data.
The layout learning data is used for supervised learning of the layout learning model.
The electronic document generator according to any one of claims 2 to 4, wherein the electronic document generator is characterized by the above. - 前記文書画像に、前記アノテーションとともに前記文書画像に含まれる前記複数の要素に係る範囲の各々の前記文書画像内における位置情報が付与されることを特徴とする請求項5に記載の電子文書生成装置。 The electronic document generator according to claim 5, wherein the document image is provided with position information in the document image for each of the ranges related to the plurality of elements included in the document image together with the annotation. ..
- 入力に基づいて、前記レイアウト認識部により認識された前記複数の要素の各々の種類、及び前記複数の要素の各々の範囲の前記文書画像内における位置情報の少なくともいずれかが修正され、この修正されたデータを追加することで前記レイアウト学習用データを更新するレイアウト学習用データ修正部をさらに備える、
ことを特徴とする請求項5または6に記載の電子文書生成装置。 Based on the input, at least one of the types of the plurality of elements recognized by the layout recognition unit and the position information in the document image of each range of the plurality of elements is corrected and corrected. It further includes a layout learning data correction unit that updates the layout learning data by adding the data.
The electronic document generator according to claim 5 or 6. - 前記レイアウト学習用データ修正部により更新された前記レイアウト学習用データを用いて、前記レイアウト学習モデルの再学習を行うレイアウト学習部をさらに備える、
ことを特徴とする請求項7に記載の電子文書生成装置。 A layout learning unit for re-learning the layout learning model using the layout learning data updated by the layout learning data correction unit is further provided.
The electronic document generator according to claim 7. - 前記文字列学習モデルの教師有り学習に用いる文字列学習用データを生成する文字列学習用データ生成部をさらに備える、
ことを特徴とする請求項2ないし8のいずれか1項に記載の電子文書生成装置。 It further includes a character string learning data generation unit that generates character string learning data used for supervised learning of the character string learning model.
The electronic document generator according to any one of claims 2 to 8. - 入力に基づいて、前記文字列認識部により生成されたテキストデータが修正され、この修正されたテキストデータを追加することで前記文字列学習用データを更新する文字列学習用データ修正部をさらに備える、
ことを特徴とする請求項9に記載の電子文書生成装置。 Based on the input, the text data generated by the character string recognition unit is modified, and the character string learning data correction unit that updates the character string learning data by adding the modified text data is further provided. ,
The electronic document generator according to claim 9. - 前記文字列学習用データ修正部により更新された前記文字列学習用データを用いて、前記文字列学習モデルの再学習を行う文字列学習部をさらに備える、
ことを特徴とする請求項10に記載の電子文書生成装置。 A character string learning unit for re-learning the character string learning model using the character string learning data updated by the character string learning data correction unit is further provided.
The electronic document generator according to claim 10. - 前記文字列認識部は、複数の前記文字列学習モデルを備え、前記複数の要素の各々に含まれる文字列の言語に適応した前記文字列学習モデルを用いる、
ことを特徴とする請求項2ないし11のいずれか1項に記載の電子文書生成装置。 The character string recognition unit includes a plurality of the character string learning models, and uses the character string learning model adapted to the language of the character string included in each of the plurality of elements.
The electronic document generator according to any one of claims 2 to 11. - 前記文書画像取得部が取得した文書画像について前処理を行う前処理部をさらに備え、
前記前処理部は、背景除去部、傾き補正部、及び形状調整部を備え、
前記背景除去部は、前記文書画像取得部が取得した前記文書画像の背景を除去し、
前記傾き補正部は、前記文書画像取得部が取得した前記文書画像の傾きを補正し、
前記形状調整部は、前記文書画像取得部が取得した前記文書画像の全体の形状及び大きさを調整する
ことを特徴とする請求項2ないし12のいずれか1項に記載の電子文書生成装置。 A pre-processing unit that performs pre-processing on the document image acquired by the document image acquisition unit is further provided.
The pretreatment unit includes a background removal unit, a tilt correction unit, and a shape adjustment unit.
The background removing unit removes the background of the document image acquired by the document image acquisition unit.
The tilt correction unit corrects the tilt of the document image acquired by the document image acquisition unit.
The electronic document generation device according to any one of claims 2 to 12, wherein the shape adjusting unit adjusts the overall shape and size of the document image acquired by the document image acquisition unit. - 前記レイアウト学習モデルは、契約書用のレイアウト学習モデル、請求書用のレイアウト学習モデル、覚書用のレイアウト学習モデル、納品書用のレイアウト学習モデル、又は領収書用のレイアウト学習モデルのいずれかであることを特徴とする請求項2ないし13のいずれか1項に記載の電子文書生成装置。 The layout learning model is either a layout learning model for contracts, a layout learning model for invoices, a layout learning model for memorandums, a layout learning model for invoices, or a layout learning model for receipts. The electronic document generator according to any one of claims 2 to 13, wherein the electronic document generator is characterized by the above.
- 電子文書生成装置に用いられるコンピュータが、
文書を画像化した文書画像を取得する文書画像取得ステップと、
文書画像と当該文書画像に含まれる文字列との対応関係を学習した文字列学習モデルを用いて、
前記文書画像取得ステップにて取得された前記文書画像に含まれる文字列を文字認識し、当該文字列に係るテキストデータを生成する文字列認識ステップと、
前記テキストデータを電子媒体のテキストとして出力する出力ステップと、
を実行することを特徴とする電子文書生成方法。 The computer used for the electronic document generator
A document image acquisition step to acquire a document image that is an image of a document,
Using a character string learning model that learned the correspondence between the document image and the character string included in the document image,
A character string recognition step that recognizes a character string included in the document image acquired in the document image acquisition step and generates text data related to the character string, and a character string recognition step.
An output step for outputting the text data as text on an electronic medium, and
An electronic document generation method characterized by performing. - 電子文書生成装置に用いられるコンピュータに、
文書を画像化した文書画像を取得する文書画像取得機能と、
文書画像と当該文書画像に含まれる文字列との対応関係を学習した文字列学習モデルを用いて、
前記文書画像取得機能にて取得された前記文書画像に含まれる文字列を文字認識し、当該文字列に係るテキストデータを生成する文字列認識機能と、
前記テキストデータを電子媒体のテキストとして出力する出力機能と、
を発揮させることを特徴とする電子文書生成プログラム。 For computers used in electronic document generators
A document image acquisition function that acquires a document image that is an image of a document,
Using a character string learning model that learned the correspondence between the document image and the character string included in the document image,
A character string recognition function that recognizes a character string included in the document image acquired by the document image acquisition function and generates text data related to the character string, and a character string recognition function.
An output function that outputs the text data as text on an electronic medium,
An electronic document generation program characterized by demonstrating.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU2021412659A AU2021412659A1 (en) | 2020-12-28 | 2021-12-23 | Architecture for digitalizing documents using multi-model deep learning, and document image processing program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2020-219612 | 2020-12-28 | ||
JP2020219612A JP7150809B2 (en) | 2020-12-28 | 2020-12-28 | Document digitization architecture by multi-model deep learning, document image processing program |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022145343A1 true WO2022145343A1 (en) | 2022-07-07 |
Family
ID=82259389
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2021/047935 WO2022145343A1 (en) | 2020-12-28 | 2021-12-23 | Architecture for digitalizing documents using multi-model deep learning, and document image processing program |
Country Status (3)
Country | Link |
---|---|
JP (2) | JP7150809B2 (en) |
AU (1) | AU2021412659A1 (en) |
WO (1) | WO2022145343A1 (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004178010A (en) * | 2002-11-22 | 2004-06-24 | Toshiba Corp | Document processor, its method, and program |
JP2020046860A (en) * | 2018-09-18 | 2020-03-26 | 株式会社三菱Ufj銀行 | Form reading apparatus |
JP2020184109A (en) * | 2019-04-26 | 2020-11-12 | Arithmer株式会社 | Learning model generation device, character recognition device, learning model generation method, character recognition method, and program |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH045779A (en) * | 1990-04-24 | 1992-01-09 | Oki Electric Ind Co Ltd | Character recognizing device |
JP3162158B2 (en) * | 1992-02-20 | 2001-04-25 | 株式会社リコー | Image reading device |
JPH07160809A (en) * | 1993-12-08 | 1995-06-23 | Nec Corp | Ocr device |
JPH08263588A (en) * | 1995-03-28 | 1996-10-11 | Fuji Xerox Co Ltd | Character recognition device |
JP3940491B2 (en) * | 1998-02-27 | 2007-07-04 | 株式会社東芝 | Document processing apparatus and document processing method |
-
2020
- 2020-12-28 JP JP2020219612A patent/JP7150809B2/en active Active
-
2021
- 2021-12-23 WO PCT/JP2021/047935 patent/WO2022145343A1/en active Application Filing
- 2021-12-23 AU AU2021412659A patent/AU2021412659A1/en active Pending
-
2022
- 2022-08-30 JP JP2022137023A patent/JP2022169754A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004178010A (en) * | 2002-11-22 | 2004-06-24 | Toshiba Corp | Document processor, its method, and program |
JP2020046860A (en) * | 2018-09-18 | 2020-03-26 | 株式会社三菱Ufj銀行 | Form reading apparatus |
JP2020184109A (en) * | 2019-04-26 | 2020-11-12 | Arithmer株式会社 | Learning model generation device, character recognition device, learning model generation method, character recognition method, and program |
Also Published As
Publication number | Publication date |
---|---|
JP2022169754A (en) | 2022-11-09 |
JP7150809B2 (en) | 2022-10-11 |
AU2021412659A9 (en) | 2024-02-08 |
JP2022104411A (en) | 2022-07-08 |
AU2021412659A1 (en) | 2023-07-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8508756B2 (en) | Image forming apparatus having capability for recognition and extraction of annotations and additionally written portions | |
JP5121599B2 (en) | Image processing apparatus, image processing method, program thereof, and storage medium | |
CN101600032B (en) | Information processing apparatus, method of processing information, control program, and recording medium | |
JP4854491B2 (en) | Image processing apparatus and control method thereof | |
US20080100624A1 (en) | Image processing apparatus and method, and program | |
US20090055159A1 (en) | Translation apparatus, translation method and computer readable medium | |
CN107633055B (en) | Method for converting picture into HTML document | |
US9460089B1 (en) | Flow rendering of annotation characters | |
US20090210215A1 (en) | Document image processing device and document image processing program | |
US20200104586A1 (en) | Method and system for manual editing of character recognition results | |
US7929772B2 (en) | Method for generating typographical line | |
US8600175B2 (en) | Apparatus and method of processing image including character string | |
US7983485B2 (en) | System and method for identifying symbols for processing images | |
JP7035656B2 (en) | Information processing equipment and programs | |
WO2022145343A1 (en) | Architecture for digitalizing documents using multi-model deep learning, and document image processing program | |
JP5020698B2 (en) | Image processing apparatus, image processing method, and image processing program | |
JP2022092119A (en) | Image processing apparatus, image processing method, and program | |
JP2008181174A (en) | Method for creating drawing original for patent application or utility model registration application | |
JP2018151699A (en) | Information processing device and program | |
JP4998176B2 (en) | Translation apparatus and program | |
JP2009223363A (en) | Document processor and document processing program | |
US11170211B2 (en) | Information processing apparatus for extracting portions filled with characters from completed document without user intervention and non-transitory computer readable medium | |
US9411547B1 (en) | Compensation for print shift in standardized forms to facilitate extraction of data therefrom | |
JP4310176B2 (en) | Image processing apparatus, image processing method, and program | |
JP2009205209A (en) | Document image processor and document image processing program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21915203 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021412659 Country of ref document: AU Date of ref document: 20211223 Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11202304908R Country of ref document: SG |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21915203 Country of ref document: EP Kind code of ref document: A1 |