CN114119410B - Method and device for correcting cells in distorted tabular image - Google Patents

Method and device for correcting cells in distorted tabular image Download PDF

Info

Publication number
CN114119410B
CN114119410B CN202111401956.9A CN202111401956A CN114119410B CN 114119410 B CN114119410 B CN 114119410B CN 202111401956 A CN202111401956 A CN 202111401956A CN 114119410 B CN114119410 B CN 114119410B
Authority
CN
China
Prior art keywords
offset
distortion
distorted
cell
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111401956.9A
Other languages
Chinese (zh)
Other versions
CN114119410A (en
Inventor
孙鹏程
王潇茵
杜婉茹
刘萱
李瑞群
郑俊康
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Aerospace Hongkang Intelligent Technology Beijing Co ltd
Original Assignee
Aerospace Hongkang Intelligent Technology Beijing Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aerospace Hongkang Intelligent Technology Beijing Co ltd filed Critical Aerospace Hongkang Intelligent Technology Beijing Co ltd
Priority to CN202111401956.9A priority Critical patent/CN114119410B/en
Publication of CN114119410A publication Critical patent/CN114119410A/en
Application granted granted Critical
Publication of CN114119410B publication Critical patent/CN114119410B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • G06T5/80
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning

Abstract

The present disclosure relates to a method and apparatus for correcting cells in a distorted table image, the method for correcting cells in a distorted table image comprising: acquiring a distortion table image; carrying out image segmentation on the distortion table image to obtain each cell in the distortion table image; determining distorted cells in the respective cells; carrying out target detection on the distorted table image to obtain a distorted unit cell detection frame; calculating the intersection and parallel ratio of the distortion unit cell and the corresponding distortion unit cell detection frame to determine the distortion degree of the distortion unit cell; determining a frame line for correcting the distortion cell and a correction frame for correcting contents within the frame line based on the distortion degree; affine transformation is performed on the frame lines of the distortion unit cells and the contents in the frame lines based on the correction frames. According to the present disclosure, the frame lines of the cells in the distorted form image and the contents within the frame lines can be corrected.

Description

Method and device for correcting cells in distorted tabular image
Technical Field
The present disclosure relates to the field of image processing technologies, and in particular, to a method and an apparatus for correcting cells in a distorted table image.
Background
Currently, the most common sources of form images are: the document is obtained by scanning the document with a scanner or by shooting the document with a camera of the electronic apparatus or the like. However, in actual practice, it is inevitable that the scanned or photographed form image is distorted. When it is desired to perform character recognition (OCR) on content in a spreadsheet image, the accuracy of recognition is greatly affected if there is distortion in the image. Therefore, it is necessary to correct the form image before performing character recognition on the form image to facilitate the character recognition.
Disclosure of Invention
The present disclosure provides a method and apparatus for correcting cells in a distorted form image, an electronic device, and a computer-readable storage medium to solve at least the problems of the related art described above.
According to a first aspect of embodiments of the present disclosure, there is provided a method of correcting a cell in a distorted table image, comprising: acquiring a distortion table image; carrying out image segmentation on the distortion table image to obtain each cell in the distortion table image; determining distorted cells in the respective cells; carrying out target detection on the distorted table image to obtain a distorted unit cell detection frame; calculating the intersection and parallel ratio of the distortion unit cell and the corresponding distortion unit cell detection frame to determine the distortion degree of the distortion unit cell; determining a frame line for correcting the distortion cell and a correction frame for correcting contents within the frame line based on the distortion degree; affine transformation is performed on the frame lines of the distortion unit cells and the contents in the frame lines based on the correction frames.
Optionally, the determining a correction frame for correcting the outline of the distorted cell and the content in the outline based on the distortion degree includes: carrying out approximate fitting on the frame lines of the distorted cells to obtain the minimum circumscribed rectangle of the distorted cells; and determining a middle area correction frame of the minimum bounding rectangle and the distorted cell detection frame as the correction frame based on the distortion degree.
Optionally, the diagonal coordinates [ x1, y1, x2, y2] of the middle region correction box are calculated by: [ x1, y1, x2, y2] ([ xa1, ya1, xa2, ya2] + [ xb1, yb1, xb2, yb2 ]). times.D/2, where [ xa1, ya1, xa2, ya2] are diagonal coordinates of the minimum bounding rectangle, [ xb1, yb1, xb2, yb2] are diagonal coordinates of the distorted cell detection box, D is the distortion degree, and D is 0.5. ltoreq. D.ltoreq.1.
Optionally, the determining a correction frame for correcting the outline of the distorted cell and the content in the outline based on the distortion degree includes: determining a distorted position of the distorted cell; and carrying out concave-convex correction on the distortion position based on the distortion degree to obtain the correction frame.
Alternatively, the coordinates of the pixel point of the distortion position after the irregularity correction are [ xd ± offset _ x, yd ± offset _ y ], where [ xd, yd ] is the coordinates of the pixel point of the distortion position before the irregularity correction, offset _ x is the correction amount for xd, offset _ y is the correction amount for yd, offset _ x is 0, offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)) in the case where the distortion position is located only on the lateral frame line of the distortion cell, and when yd is corrected in the manner of yd + offset _ y, if yd + offset _ y > yp, offset _ y is 0, when yd is corrected in the manner of yd-offset _ y, if yd-offset _ y < y, offset _ y is 0, offset _ y is located only on the longitudinal frame line of the distortion cell, offset _ y is 0, offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), and when xd + offset _ x is used to correct xd, if xd + offset _ x > xp, offset _ x is 0, when xd is used to correct xd, if xd-offset _ x < xp, offset _ x is 0, and when the distortion position is located on both the lateral and vertical frame lines, offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)), and when xd + offset _ x is used to correct xd, if xd + offset _ x > xd, offset _ x is 0, when xd + offset _ x is used to correct xd, if xd-offset _ x < xp, offset _ x is 0, if yd + offset _ y > yp, if yd + offset _ y is corrected, offset _ y is 0, and if yd-offset _ y < yp, offset _ y is 0, where H is the height of the distorted cell detection frame, W is the width of the distorted cell detection frame, D is the distortion degree, 0.5 or less D is less than 1, xp is the abscissa of the undistorted position on the same vertical frame line as [ xd, yd ], and yp is the ordinate of the undistorted position on the same horizontal frame line as [ xd, yd ], when yd is corrected by yd-offset _ y, if yd-offset _ y is smaller than yp.
Optionally, the determining a distorted cell of the respective cells includes: respectively extracting the features of each cell; determining the distortion cell from the extracted features.
Optionally, the image segmentation is performed on the distorted table image through a trained image segmentation model, so as to obtain each cell in the distorted table image.
Optionally, target detection is performed on the distorted table image through a trained target detection model, so as to obtain a distorted cell detection frame.
Optionally, before the image segmentation and the target detection are performed on the distorted table image, the method further includes: converting the distortion table image into a gray image to obtain a gray image of the distortion table image; and carrying out binarization processing on the distorted table image gray-scale map to obtain a binarized distorted table image.
According to a second aspect of the embodiments of the present disclosure, there is provided an apparatus for correcting a cell in a distorted form image, including: an image acquisition unit configured to acquire a distortion table image; an image segmentation unit configured to perform image segmentation on the distortion table image to obtain each cell in the distortion table image; a determination unit configured to determine a distorted cell among the respective cells; the target detection unit is configured to perform target detection on the distorted table image to obtain a distorted cell detection frame; a calculation unit configured to calculate an intersection ratio of the distortion cell to the corresponding distortion cell detection frame to determine a distortion degree of the distortion cell; a correction frame determination unit configured to determine a frame line for correcting the distortion cell and a correction frame for correcting contents within the frame line based on the degree of distortion; an affine transformation unit configured to perform affine transformation on a frame line of the distortion cell and contents within the frame line based on the correction frame.
Optionally, the correction frame determination unit is configured to: carrying out approximate fitting on the frame lines of the distorted cells to obtain the minimum circumscribed rectangle of the distorted cells; and determining a middle area correction frame of the minimum bounding rectangle and the distorted cell detection frame as the correction frame based on the distortion degree.
Optionally, the diagonal coordinates [ x1, y1, x2, y2] of the middle region correction box are calculated by: [ x1, y1, x2, y2] ([ xa1, ya1, xa2, ya2] + [ xb1, yb1, xb2, yb2 ]). times.D/2, where [ xa1, ya1, xa2, ya2] are diagonal coordinates of the minimum bounding rectangle, [ xb1, yb1, xb2, yb2] are diagonal coordinates of the distorted cell detection box, D is the distortion degree, and D is 0.5. ltoreq. D.ltoreq.1.
Optionally, the correction frame determination unit is configured to: determining a distorted position of the distorted cell; and carrying out concave-convex correction on the distortion position based on the distortion degree to obtain the correction frame.
Alternatively, the coordinates of the pixel point of the distortion position after the irregularity correction are [ xd ± offset _ x, yd ± offset _ y ], where [ xd, yd ] is the coordinates of the pixel point of the distortion position before the irregularity correction, offset _ x is the correction amount for xd, offset _ y is the correction amount for yd, offset _ x is 0, offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)) in the case where the distortion position is located only on the lateral frame line of the distortion cell, and when yd is corrected in the manner of yd + offset _ y, if yd + offset _ y > yp, offset _ y is 0, when yd is corrected in the manner of yd-offset _ y, if yd-offset _ y < y, offset _ y is 0, offset _ y is located only on the longitudinal frame line of the distortion cell, offset _ y is 0, offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), and when xd + offset _ x is used to correct xd, if xd + offset _ x > xp, offset _ x is 0, when xd is used to correct xd, if xd-offset _ x < xp, offset _ x is 0, and when the distortion position is located on both the lateral and vertical frame lines, offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)), and when xd + offset _ x is used to correct xd, if xd + offset _ x > xd, offset _ x is 0, when xd + offset _ x is used to correct xd, if xd-offset _ x < xp, offset _ x is 0, if yd + offset _ y > yp, if yd + offset _ y is corrected, offset _ y is 0, and if yd-offset _ y < yp, offset _ y is 0, where H is the height of the distorted cell detection frame, W is the width of the distorted cell detection frame, D is the distortion degree, 0.5 or less D is less than 1, xp is the abscissa of the undistorted position on the same vertical frame line as [ xd, yd ], and yp is the ordinate of the undistorted position on the same horizontal frame line as [ xd, yd ], when yd is corrected by yd-offset _ y, if yd-offset _ y is smaller than yp.
Optionally, the determining unit is configured to: respectively extracting the features of each cell; determining the distortion cell from the extracted features.
Optionally, the image segmentation unit is configured to: and carrying out image segmentation on the distorted table image through a trained image segmentation model to obtain each cell in the distorted table image.
Optionally, the target detection unit is configured to: and carrying out target detection on the distorted table image through a trained target detection model to obtain a distorted unit cell detection frame.
Optionally, the image acquisition unit is further configured to: after the distortion table image is obtained, converting the distortion table image into a gray image to obtain a gray image of the distortion table image; and carrying out binarization processing on the distorted table image gray-scale map to obtain a binarized distorted table image.
According to a third aspect of the embodiments of the present disclosure, there is provided an electronic apparatus including: a processor; a memory for storing computer executable instructions; wherein the computer executable instructions, when executed by the processor, cause the processor to implement the method as described above.
According to a fourth aspect of embodiments of the present disclosure, there is provided a computer-readable storage medium, wherein instructions, when executed by a processor, cause the processor to perform the method as described above.
The technical scheme provided by the embodiment of the disclosure at least brings the following beneficial effects:
according to the method and the device for correcting the cells in the distorted table image, the frame lines of the cells in the distorted table image and the content in the frame lines can be corrected so as to be beneficial to character recognition.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and, together with the description, serve to explain the principles of the disclosure and are not to be construed as limiting the disclosure.
Fig. 1 is an implementation scenario diagram illustrating a method of correcting cells in a distorted form image according to an exemplary embodiment of the present disclosure.
Fig. 2 is a flowchart illustrating a method of correcting cells in a distorted form image according to an exemplary embodiment of the present disclosure.
Fig. 3 is a schematic diagram for explaining the cross-over ratio.
Fig. 4 is a block diagram illustrating an apparatus for correcting a cell in a distorted table image according to an exemplary embodiment of the present disclosure.
Fig. 5 is a block diagram illustrating an electronic device according to an exemplary embodiment of the present disclosure.
Detailed Description
In order to make the technical solutions of the present disclosure better understood by those of ordinary skill in the art, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the accompanying drawings.
It should be noted that the terms "first," "second," and the like in the description and claims of the present disclosure and in the above-described drawings are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the disclosure described herein are capable of operation in sequences other than those illustrated or otherwise described herein. The embodiments described in the following examples do not represent all embodiments consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.
In this case, the expression "at least one of the items" in the present disclosure means a case where three types of parallel expressions "any one of the items", "a combination of any plural ones of the items", and "the entirety of the items" are included. For example, "include at least one of a and B" includes the following three cases in parallel: (1) comprises A; (2) comprises B; (3) including a and B. For another example, "at least one of the first step and the second step is performed", which means that the following three cases are juxtaposed: (1) executing the step one; (2) executing the step two; (3) and executing the step one and the step two.
At present, some distortion inevitably exists in a form image acquired by scanning or photographing. This greatly affects the accuracy of subsequent text recognition of the form image.
Therefore, the present disclosure provides a method and an apparatus for correcting cells in a distorted table image, which can correct the frame lines of the cells in the distorted table image and the content (characters, etc.) in the frame lines, thereby improving the accuracy of subsequent character recognition.
Fig. 1 is an implementation scenario diagram illustrating a method of correcting cells in a distorted form image according to an exemplary embodiment of the present disclosure. As shown in fig. 1, the implementation scenario includes a server 100, user terminals 110 and 120. The user terminals 110 and 120 and the server 100 may be used to implement the method for correcting cells in the distorted form image of the present disclosure, and the correction may be performed by the server 100 after the distorted form image is acquired by the user terminals 110 and 120. At this time, the server 100 can perform bidirectional communication with the user terminals 110 and 120. The user terminals are not limited to the number and types shown in the figures, and include mobile terminals such as smart phones and tablet computers with camera functions, personal computers, and the like, and may also include any other electronic devices capable of shooting or scanning. The server 100 may be a single server, a server cluster composed of several servers, a cloud computing platform, or a virtualization center.
Next, a method and apparatus for correcting cells in a distorted form image according to an exemplary embodiment of the present disclosure will be described with reference to fig. 2 to 4.
Fig. 2 is a flowchart illustrating a method of correcting cells in a distorted form image according to an exemplary embodiment of the present disclosure.
As shown in fig. 2, in step S210, a distortion table image is acquired. Here, the distorted table image is an image in which distortion occurs in a normal table image without any distortion. The distortion table image may be acquired by various methods in the related art. Also, the distorted form image may be an image that is distorted due to an improper operation when a document with a form is scanned or photographed. The distortion may be a case where a part of the form image is deformed without being in a normal position.
Then, in step S220, the distortion table image acquired in step S210 is subjected to image segmentation to obtain each cell in the distortion table image. Here, each cell in the distortion table image can be obtained by various conventional image segmentation methods. According to an exemplary embodiment of the disclosure, the distorted table image may be subjected to image segmentation by a trained image segmentation model, resulting in each cell in the distorted table image. Here, the trained image segmentation model is obtained by training in advance based on the target of image segmentation. This makes it possible to accurately obtain each cell in the distortion table image. Further, the distortion table image may be preprocessed as described below prior to image segmentation of the distortion table image to facilitate image segmentation.
According to an exemplary embodiment of the present disclosure, before step S220, the method may further include: converting the distorted table image into a gray image to obtain a gray image of the distorted table image; and then, carrying out binarization processing on the distorted table image gray-scale map to obtain a binarized distorted table image. That is, in the case where the distorted table image is not a gray map, it is converted into a gray map, and then the gray values of the pixel points in the gray map are set to 0 or 255 (for thresholding, 255 is set when less than the threshold value and 0 is set when greater than the threshold value), thereby obtaining a binarized (and inverted) distorted table image. Here, the distortion table image may be converted into a grayscale image by various existing methods. The binarization processing may be performed by various conventional methods. For example, a binarized image that can reflect the entire image and local features can be obtained by appropriate threshold selection. By performing such preprocessing on the distorted table image, subsequent image segmentation and target detection can be facilitated.
Then, in step S230, a distortion cell among the respective cells obtained by image division is determined. Here, the distorted cell is a cell with distortion. The cells with distortion in each cell can be determined by various existing methods.
According to an exemplary embodiment of the present disclosure, step S230 may include: feature extraction is performed on each cell, and then a distorted cell is determined according to the extracted features. For example, whether a cell is a distorted cell may be determined based on a feature that can indicate that the cell is distorted, such as a frame line feature of the cell. The feature may be extracted by various existing methods, or may be extracted by a predetermined feature extraction model.
Then, in step S240, target detection is performed on the distortion table image acquired in step S210, resulting in a distortion cell detection frame. Here, the target of the target detection is a distorted cell. Similarly, the distorted table image may be preprocessed before the target detection is performed on the distorted table image, so as to facilitate the target detection. The specific method of pretreatment is as described above. Here, the distorted cell detection frame can be detected by various existing object detection methods. According to the exemplary embodiment of the disclosure, the distorted table image can be subjected to target detection through a trained target detection model, so that a distorted cell detection frame is obtained. Here, the trained object detection model is obtained by training in advance according to an object to be detected. The target detection model can adopt a Yolov3 network, and the Yolov3 network comprises: darkent53 (feature extractor model), FPN (feature pyramid network), Yolo Head function, etc. In the present invention, the size of the distortion cell detection frame is taken as the size of the cell of the distortion cell in the normal case for determining the distortion degree. The order of step S240 and step S220 may be changed or may be executed simultaneously, and this does not affect the execution of the present invention.
Then, in step S250, the intersection ratio of the distortion cell obtained by image division and the corresponding distortion cell detection frame is calculated to determine the distortion degree of the distortion cell. Here, the Intersection over ratio (IoU) is the ratio of the Intersection area of the distortion cell and the distortion cell detection box to the phase area (as shown in fig. 3), and is used to measure the similarity between the two. The intersection ratio has a value range between 0 and 1, wherein 0 means that the two do not have overlapped pixels, and 1 means that the two are equal. In the present invention, the intersection ratio and the distortion degree are linearly corresponding, and the larger the intersection ratio is, the smaller the distortion degree is, and the smaller the intersection ratio is, the larger the distortion degree is. Here, the distortion degree of the distortion cell may be determined based on the cross-over ratio by various existing methods.
Then, in step S260, based on the distortion degree, a frame line for correcting the distortion cell and a correction frame for correcting the content within the frame line are determined.
According to an exemplary embodiment of the present disclosure, step S260 may include: and performing approximate fitting on the frame lines of the distorted cells to obtain the minimum circumscribed rectangle of the distorted cells. Then, based on the distortion degree, a middle area correction frame of the minimum bounding rectangle and the distortion cell detection frame is determined as a correction frame. That is, the middle area between the minimum bounding rectangle and the distorted cell detection frame is taken as the correction frame. Here, the minimum bounding rectangle may be obtained by an existing approximation fitting method.
According to an exemplary embodiment of the present disclosure, diagonal coordinates [ x1, y1, x2, y2] of the middle region correction box of the minimum bounding rectangle and the distorted cell detection box may be calculated by:
[x1,y1,x2,y2]=([xa1,ya1,xa2,ya2]+[xb1,yb1,xb2,yb2])×D/2,
wherein [ xa1, ya1, xa2 and ya2] are diagonal coordinates of the minimum circumscribed rectangle, [ xb1, yb1, xb2 and yb2] are diagonal coordinates of the distorted cell detection box, D is the distortion degree, and D is more than or equal to 0.5 and less than or equal to 1. This makes it possible to obtain a frame line for correcting the distortion cell and a correction frame for correcting the content in the frame line.
Further, according to an exemplary embodiment of the present disclosure, step S260 may include: the distorted position of the distorted cell is determined. Then, based on the degree of distortion, the distortion position is corrected for unevenness, and a correction frame is obtained. Here, the distortion position of the distortion cell may be determined by various existing methods. Then, the correction frame can be obtained by performing convex correction on the distorted position such as a concave and concave correction on the distorted position such as a convex based on the degree of distortion.
According to an exemplary embodiment of the present disclosure, coordinates of a pixel point of a distortion position after concave-convex correction may be set as [ xd ± offset _ x, yd ± offset _ y ], where [ xd, yd ] is coordinates of a pixel point of a distortion position before concave-convex correction, offset _ x is a correction amount for xd, and offset _ y is a correction amount for yd. At this time, when the distortion position is located only on the horizontal frame line of the distortion cell, only the ordinate is corrected, offset _ x is 0, and offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)), and when yd is corrected in the manner of yd + offset _ y (that is, when the distortion position is downward distorted), if yd + offset _ y > yp, offset _ y is 0, and when yd is corrected in the manner of yd-offset _ y (that is, when the distortion position is upward distorted), if yd-offset _ y < yp, offset _ y is 0.
In addition, when the distortion position is located only on the vertical frame line of the distortion cell, only the abscissa is corrected, offset _ y is 0, and offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), and when xd is corrected in the manner of xd + offset _ x (that is, when the distortion position is distorted leftward), if xd + offset _ x > xp, offset _ x is 0, and when xd is corrected in the manner of xd-offset _ x (that is, when the distortion position is distorted rightward), if xd-offset _ x < xp, offset _ x is 0.
Further, in the case where the distortion position is located on both the horizontal frame line and the vertical frame line (i.e., the intersection of the horizontal frame line and the vertical frame line), the abscissa and the ordinate are corrected, respectively, with offset _ x being 1/D × 100 × sin (2 × pi × xd/(2 × W)), offset _ y being 1/D × 100 × sin (2 × pi × yd/(2 × H)), and when xd + offset _ x is used to correct xd (i.e., the distortion position is distorted to the left), if xd + offset _ x > xp, offset _ x is 0, when xd is corrected in the manner of xd-offset _ x (i.e., the distortion position is distorted to the right), if xd-offset _ x < xp, offset _ x is 0, when yd is corrected in the manner of yd + offset _ x, i.e., the distortion position is distorted to the right, when yd-offset _ x < xp, i.e., the distortion position is corrected to the lower direction, if yd + offset _ y > yp, then offset _ y is 0, and when yd is corrected in the manner of yd-offset _ y (i.e., when the distortion position is upward distortion), if yd-offset _ y < yp, then offset _ y is 0. H is the height of the distorted cell detection frame, W is the width of the distorted cell detection frame, D is the distortion degree, and D is more than or equal to 0.5 and less than or equal to 1. Here, xp is the abscissa of the undistorted position on the same vertical frame line as [ xd, yd ]. In addition, a distorted position and an undistorted position exist on the vertical frame line where distortion occurs, and when the distorted position is corrected, excessive correction can be restricted according to the abscissa of the undistorted position on the vertical frame line. yp is the ordinate of the undistorted position on the same transverse frame line as [ xd, yd ]. Similarly, a distorted position and an undistorted position exist on the distorted horizontal frame line, and when the distorted position is corrected, the excessive correction can be restricted according to the ordinate of the undistorted position on the horizontal frame line.
This makes it possible to obtain a frame line for correcting the distortion cell and a correction frame for correcting the content in the frame line.
Then, in step S270, affine transformation is performed on the frame lines of the distortion cells and the contents inside the frame lines based on the correction frame obtained as described above. Thereby, the frame line of the distorted cell and the content inside the frame line are corrected.
According to the method for correcting the cells in the distorted table image, the frame lines and the contents in the frame lines of the cells in the distorted table image can be corrected to be beneficial to character recognition.
Fig. 4 is a block diagram illustrating an apparatus for correcting a cell in a distorted table image according to an exemplary embodiment of the present disclosure.
As shown in fig. 4, the apparatus 400 for correcting cells in a distorted form image includes: an image acquisition unit 410, an image segmentation unit 420, a determination unit 430, an object detection unit 440, a calculation unit 450, a correction frame determination unit 460, and an affine transformation unit 470.
Wherein the image acquisition unit 410 is configured to acquire a distortion table image.
According to an exemplary embodiment of the present disclosure, the image acquisition unit 410 may be further configured to: after the distorted table image is obtained, preprocessing is carried out on the distorted table image, the distorted table image is converted into a gray-scale image, and the gray-scale image of the distorted table image is obtained. And then carrying out binarization processing on the distorted table image gray-scale map to obtain a binarized distorted table image. Thereby, subsequent image segmentation and object detection is facilitated.
The image segmentation unit 420 is configured to perform image segmentation on the distortion table image obtained by the image acquisition unit 410, resulting in individual cells in the distortion table image.
According to an exemplary embodiment of the present disclosure, the image segmentation unit 420 may be configured to: and carrying out image segmentation on the distorted table image through the trained image segmentation model to obtain each cell in the distorted table image.
The determination unit 430 is configured to determine a distorted cell among the respective cells into which the image division unit 420 performs the image division.
According to an exemplary embodiment of the present disclosure, the determining unit 430 may be configured to perform feature extraction on the respective cells, and then determine a distorted cell according to the extracted features.
The target detection unit 440 is configured to perform target detection on the distorted table image obtained by the image obtaining unit 410, resulting in a distorted cell detection box.
According to an exemplary embodiment of the present disclosure, the target detection unit 440 may be configured to: and carrying out target detection on the distorted table image through the trained target detection model to obtain a distorted unit cell detection frame.
The calculation unit 450 is configured to calculate an intersection ratio of the distortion cell and the corresponding distortion cell detection frame to determine a distortion degree of the distortion cell.
The correction frame determination unit 460 is configured to determine a correction frame for correcting the outline of the distortion cell and the content within the outline, based on the degree of distortion.
According to an exemplary embodiment of the present disclosure, the correction frame determination unit 460 may be configured to: and performing approximate fitting on the frame lines of the distorted cells to obtain the minimum circumscribed rectangle of the distorted cells. Then, based on the distortion degree, a middle area correction frame of the minimum bounding rectangle and the distortion cell detection frame is determined as a correction frame.
According to an exemplary embodiment of the present disclosure, the diagonal coordinates [ x1, y1, x2, y2] of the middle region correction box are calculated by: [ x1, y1, x2, y2] ([ xa1, ya1, xa2, ya2] + [ xb1, yb1, xb2, yb2 ]). times.D/2, where [ xa1, ya1, xa2, ya2] are the diagonal coordinates of the minimum bounding rectangle, [ xb1, yb1, xb2, yb2] are the diagonal coordinates of the distorted cell detection box, D is the distortion degree, and D is 0.5. ltoreq. D.ltoreq.1. This makes it possible to obtain a frame line for correcting the distortion cell and a correction frame for correcting the content in the frame line.
Further, according to an exemplary embodiment of the present disclosure, the correction frame determination unit 460 may be configured to: the distorted position of the distorted cell is determined. Then, based on the degree of distortion, the distortion position is corrected for unevenness, and a correction frame is obtained.
According to the exemplary embodiment of the present disclosure, coordinates of a pixel point of a distortion position after concave-convex correction are [ xd ± offset _ x, yd ± offset _ y ], where [ xd, yd ] is coordinates of a pixel point of a distortion position before concave-convex correction, offset _ x is a correction amount for xd, offset _ y is a correction amount for yd, and if the distortion position is located only on a lateral frame line of a distortion cell, offset _ x is 0, offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)), and, when yd is corrected in a manner of yd + offset _ y, if yd + offset _ y > yp, offset _ y is 0, and if yd-offset _ y is corrected in a manner of yd-offset _ y, offset _ y is 0, if yd-offset _ y < yp, offset _ y is located only on a longitudinal frame line, offset _ y is 0, offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), and when xd is corrected in the manner of xd + offset _ x, if xd + offset _ x > xp, offset _ x is 0, when xd is corrected in the manner of xd-offset _ x, if xd-offset _ x < xp, offset _ x is 0, when the distortion position is both on the lateral frame line and on the vertical frame line, offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)), and when xd is corrected in the manner of xd + offset _ x, if xd + offset _ x > x 100 × sin (2 × pi × yd/(2 × H)), if x + offset _ x > 0, if xd + offset _ x > x _ x, offset _ x is 0, when xd is corrected in the manner of xd + offset _ x, when yd is corrected by using yd + offset _ y, if yd + offset _ y > yp, offset _ y is 0, when yd is corrected by using yd-offset _ y, if yd-offset _ y < yp, offset _ y is 0, wherein H is the height of the distorted cell detection frame, W is the width of the distorted cell detection frame, D is the distortion degree, 0.5 is equal to or less than 1, xp is the abscissa of the undistorted position on the same vertical frame line as [ xd, yd ], and yp is the ordinate of the undistorted position on the same horizontal frame line as [ xd, yd ].
This makes it possible to obtain a frame line for correcting the distortion cell and a correction frame for correcting the content in the frame line.
The affine transformation unit 470 is configured to affine-transform the frame line of the distortion cell and the contents within the frame line based on the above-described correction frame. Thereby, the frame line of the distorted cell and the content inside the frame line are corrected.
According to the device for correcting the cells in the distorted table image, the frame lines and the contents in the frame lines of the cells in the distorted table image can be corrected to be beneficial to character recognition.
It should be understood that the apparatus for correcting cells in a distorted form image according to an exemplary embodiment of the present disclosure may perform the method described above with reference to fig. 2, and thus, in order to avoid redundancy, detailed description thereof is omitted.
Fig. 5 shows a block diagram of an electronic device according to an exemplary embodiment of the present disclosure.
The electronic device 500 includes at least one memory 510 having stored therein a set of computer-executable instructions that, when executed by the at least one processor, cause the at least one processor 520 to perform a method of correcting cells in a distorted form image according to an exemplary embodiment of the disclosure.
By way of example, the electronic device 500 may be a PC computer, tablet device, personal digital assistant, smartphone, or other device capable of executing the set of instructions described above. Here, the electronic device 500 need not be a single electronic device, but can be any collection of devices or circuits that can execute the above instructions (or sets of instructions) individually or in combination. The electronic device 500 may also be part of an integrated control system or system manager, or may be configured as a portable electronic device that interfaces with local or remote (e.g., via wireless transmission).
In the electronic device 500, the processor 520 may include a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), a programmable logic device, a special purpose processor system, a microcontroller, or a microprocessor. By way of example, and not limitation, processor 520 may also include an analog processor, a digital processor, a microprocessor, a multi-core processor, a processor array, a network processor, or the like.
Processor 520 may execute instructions or code stored in memory, where memory 510 may also store data. The instructions and data may also be transmitted or received over a network via a network interface device, which may employ any known transmission protocol.
Memory 510 may be integrated with processor 520, e.g., with RAM or flash memory disposed within an integrated circuit microprocessor or the like. Further, memory 510 may comprise a stand-alone device, such as an external disk drive, storage array, or any other storage device usable by a database system. The memory 510 and the processor 520 may be operatively coupled or may communicate with each other, such as through I/O ports, network connections, etc., so that the processor 520 can read files stored in the memory 510.
In addition, the electronic device 500 may also include a video display (such as a liquid crystal display) and a user interaction interface (such as a keyboard, mouse, touch input device, etc.). All components of the electronic device may be connected to each other via a bus and/or a network.
According to an exemplary embodiment of the present disclosure, there may also be provided a computer-readable storage medium, wherein instructions in the computer-readable storage medium, when executed by at least one processor, cause the at least one processor to perform the method of correcting a cell in a distorted form image of the present disclosure. Examples of the computer-readable storage medium herein include: read-only memory (ROM), random-access programmable read-only memory (PROM), electrically erasable programmable read-only memory (EEPROM), random-access memory (RAM), dynamic random-access memory (DRAM), static random-access memory (SRAM), flash memory, non-volatile memory, CD-ROM, CD-R, CD + R, CD-RW, CD + RW, DVD-ROM, DVD-R, DVD + R, DVD-RW, DVD + RW, DVD-RAM, BD-ROM, BD-R, BD-R LTH, BD-RE, Blu-ray or compact disc memory, Hard Disk Drive (HDD), solid-state drive (SSD), card-type memory (such as a multimedia card, a Secure Digital (SD) card or a extreme digital (XD) card), magnetic tape, a floppy disk, a magneto-optical data storage device, an optical data storage device, a hard disk, a magnetic tape, a magneto-optical data storage device, a hard disk, a magnetic tape, a magnetic data storage device, a magnetic tape, a magnetic data storage device, a magnetic tape, a magnetic data storage device, a magnetic tape, a magnetic data storage device, a magnetic tape, a magnetic data storage device, A solid state disk, and any other device configured to store and provide a computer program and any associated data, data files, and data structures to a processor or computer in a non-transitory manner such that the processor or computer can execute the computer program. The computer program in the computer-readable storage medium described above can be run in an environment deployed in a computer apparatus, such as a client, a host, a proxy device, a server, and the like, and further, in one example, the computer program and any associated data, data files, and data structures are distributed across a networked computer system such that the computer program and any associated data, data files, and data structures are stored, accessed, and executed in a distributed fashion by one or more processors or computers.
According to an exemplary embodiment of the present disclosure, a computer program product is provided, comprising computer instructions which, when executed by a processor, implement the method of correcting cells in a distorted form image of an exemplary embodiment of the present disclosure.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This disclosure is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.

Claims (20)

1. A method of correcting cells in a distorted tabular image, comprising:
acquiring a distortion table image;
carrying out image segmentation on the distortion table image to obtain each cell in the distortion table image;
determining distorted cells in the respective cells;
carrying out target detection on the distorted table image to obtain a distorted unit cell detection frame;
calculating the intersection and parallel ratio of the distortion unit cell and the corresponding distortion unit cell detection frame to determine the distortion degree of the distortion unit cell;
determining a frame line for correcting the distortion cell and a correction frame for correcting contents within the frame line based on the distortion degree;
affine transformation is performed on the frame lines of the distortion unit cells and the contents in the frame lines based on the correction frames.
2. The method of claim 1, wherein determining a correction box for correcting a box line of the distorted cell and contents within the box line based on the distortion degree comprises:
carrying out approximate fitting on the frame lines of the distorted cells to obtain the minimum circumscribed rectangle of the distorted cells;
and determining a middle area correction frame of the minimum bounding rectangle and the distorted cell detection frame as the correction frame based on the distortion degree.
3. The method of claim 2, wherein the diagonal coordinates [ x1, y1, x2, y2] of the middle region correction box are calculated by:
[x1,y1,x2,y2]=([xa1,ya1,xa2,ya2]+[xb1,yb1,xb2,yb2])×D/2,
wherein [ xa1, ya1, xa2 and ya2] are diagonal coordinates of the minimum circumscribed rectangle, [ xb1, yb1, xb2 and yb2] are diagonal coordinates of the distorted cell detection box, D is the distortion degree, and D is more than or equal to 0.5 and less than or equal to 1.
4. The method of claim 1, wherein determining a correction box for correcting a box line of the distorted cell and contents within the box line based on the distortion degree comprises:
determining a distorted position of the distorted cell;
and carrying out concave-convex correction on the distortion position based on the distortion degree to obtain the correction frame.
5. The method according to claim 4, wherein coordinates of a pixel point of the distorted position after the irregularity correction are [ xd ± offset _ x, yd ± offset _ y ], where [ xd, yd ] is coordinates of a pixel point of the distorted position before the irregularity correction, offset _ x is a correction amount for xd, and offset _ y is a correction amount for yd,
in the case where the distortion position is located only on the lateral frame line of the distortion cell, offset _ x is 0, and offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)), and when yd is corrected in the manner of yd + offset _ y, if yd + offset _ y > yp, offset _ y is 0, and when yd is corrected in the manner of yd-offset _ y, if yd-offset _ y < yp, offset _ y is 0,
in the case where the distortion position is located only on the vertical frame line of the distortion cell, offset _ y is 0, and offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), and when xd + offset _ x is used for correction of xd, if xd + offset _ x > xp, offset _ x is 0, and when xd is used for correction of xd in the xd-offset _ x manner, if xd-offset _ x < xp, offset _ x is 0,
in the case where the distortion position is located on both the lateral frame line and the longitudinal frame line, offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)), and when xd is corrected in the manner of xd + offset _ x, if xd + offset _ x > xp, offset _ x is 0, and when xd is corrected in the manner of xd-offset _ x, if xd-offset _ x < xp, offset _ x is 0,
when yd is corrected in the manner of yd + offset _ y, if yd + offset _ y > yp, then offset _ y is 0, when yd is corrected in the manner of yd-offset _ y, if yd-offset _ y < yp, then offset _ y is 0,
h is the height of the distorted cell detection frame, W is the width of the distorted cell detection frame, D is the distortion degree, D is more than or equal to 0.5 and less than or equal to 1, xp is the abscissa of the undistorted position on the same longitudinal frame line with [ xd, yd ], yp is the ordinate of the undistorted position on the same transverse frame line with [ xd, yd ].
6. The method of claim 1, wherein determining the distorted cell of the individual cells comprises:
respectively extracting the features of each cell;
determining the distortion cell from the extracted features.
7. The method of claim 1,
and carrying out image segmentation on the distorted table image through a trained image segmentation model to obtain each cell in the distorted table image.
8. The method of claim 1,
and carrying out target detection on the distorted table image through a trained target detection model to obtain a distorted unit cell detection frame.
9. The method of claim 1, wherein prior to image segmentation and target detection of the distortion table image, further comprising:
converting the distortion table image into a gray image to obtain a gray image of the distortion table image;
and carrying out binarization processing on the distorted table image gray-scale map to obtain a binarized distorted table image.
10. An apparatus for correcting cells in a distorted tabular image, comprising:
an image acquisition unit configured to acquire a distortion table image;
an image segmentation unit configured to perform image segmentation on the distortion table image to obtain each cell in the distortion table image;
a determination unit configured to determine a distorted cell among the respective cells;
the target detection unit is configured to perform target detection on the distorted table image to obtain a distorted cell detection frame;
a calculation unit configured to calculate an intersection ratio of the distortion cell to the corresponding distortion cell detection frame to determine a distortion degree of the distortion cell;
a correction frame determination unit configured to determine a frame line for correcting the distortion cell and a correction frame for correcting contents within the frame line based on the degree of distortion;
an affine transformation unit configured to perform affine transformation on a frame line of the distortion cell and contents within the frame line based on the correction frame.
11. The apparatus according to claim 10, wherein the correction frame determination unit is configured to:
carrying out approximate fitting on the frame lines of the distorted cells to obtain the minimum circumscribed rectangle of the distorted cells;
and determining a middle area correction frame of the minimum bounding rectangle and the distorted cell detection frame as the correction frame based on the distortion degree.
12. The apparatus of claim 11, wherein the diagonal coordinates [ x1, y1, x2, y2] of the middle region correction box are calculated by:
[x1,y1,x2,y2]=([xa1,ya1,xa2,ya2]+[xb1,yb1,xb2,yb2])×D/2,
wherein [ xa1, ya1, xa2 and ya2] are diagonal coordinates of the minimum circumscribed rectangle, [ xb1, yb1, xb2 and yb2] are diagonal coordinates of the distorted cell detection box, D is the distortion degree, and D is more than or equal to 0.5 and less than or equal to 1.
13. The apparatus according to claim 10, wherein the correction frame determination unit is configured to:
determining a distorted position of the distorted cell;
and carrying out concave-convex correction on the distortion position based on the distortion degree to obtain the correction frame.
14. The apparatus according to claim 13, wherein coordinates of a pixel point of the distorted position after the irregularity correction are [ xd ± offset _ x, yd ± offset _ y ], where [ xd, yd ] is coordinates of a pixel point of the distorted position before the irregularity correction, offset _ x is a correction amount for xd, and offset _ y is a correction amount for yd,
in the case where the distortion position is located only on the lateral frame line of the distortion cell, offset _ x is 0, and offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)), and when yd is corrected in the manner of yd + offset _ y, if yd + offset _ y > yp, offset _ y is 0, and when yd is corrected in the manner of yd-offset _ y, if yd-offset _ y < yp, offset _ y is 0,
in the case where the distortion position is located only on the vertical frame line of the distortion cell, offset _ y is 0, and offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), and when xd + offset _ x is used for correction of xd, if xd + offset _ x > xp, offset _ x is 0, and when xd is used for correction of xd in the xd-offset _ x manner, if xd-offset _ x < xp, offset _ x is 0,
in the case where the distortion position is located on both the lateral frame line and the longitudinal frame line, offset _ x is 1/D × 100 × sin (2 × pi × xd/(2 × W)), offset _ y is 1/D × 100 × sin (2 × pi × yd/(2 × H)), and when xd is corrected in the manner of xd + offset _ x, if xd + offset _ x > xp, offset _ x is 0, and when xd is corrected in the manner of xd-offset _ x, if xd-offset _ x < xp, offset _ x is 0,
when yd is corrected in the manner of yd + offset _ y, if yd + offset _ y > yp, then offset _ y is 0, when yd is corrected in the manner of yd-offset _ y, if yd-offset _ y < yp, then offset _ y is 0,
h is the height of the distorted cell detection frame, W is the width of the distorted cell detection frame, D is the distortion degree, D is more than or equal to 0.5 and less than or equal to 1, xp is the abscissa of the undistorted position on the same longitudinal frame line with [ xd, yd ], yp is the ordinate of the undistorted position on the same transverse frame line with [ xd, yd ].
15. The apparatus of claim 10, wherein the determining unit is configured to:
respectively extracting the features of each cell;
determining the distortion cell from the extracted features.
16. The apparatus of claim 10, wherein the image segmentation unit is configured to:
and carrying out image segmentation on the distorted table image through a trained image segmentation model to obtain each cell in the distorted table image.
17. The apparatus of claim 10, wherein the target detection unit is configured to:
and carrying out target detection on the distorted table image through a trained target detection model to obtain a distorted unit cell detection frame.
18. The apparatus of claim 10, wherein the image acquisition unit is further configured to: after the distortion table image is acquired,
converting the distortion table image into a gray image to obtain a gray image of the distortion table image;
and carrying out binarization processing on the distorted table image gray-scale map to obtain a binarized distorted table image.
19. An electronic device, comprising:
a processor;
a memory for storing computer executable instructions;
wherein the computer executable instructions, when executed by the processor, cause the processor to implement the method of any one of claims 1 to 9.
20. A computer-readable storage medium, wherein instructions in the computer-readable storage medium, when executed by a processor, cause the processor to perform the method of any of claims 1-9.
CN202111401956.9A 2021-11-19 2021-11-19 Method and device for correcting cells in distorted tabular image Active CN114119410B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111401956.9A CN114119410B (en) 2021-11-19 2021-11-19 Method and device for correcting cells in distorted tabular image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111401956.9A CN114119410B (en) 2021-11-19 2021-11-19 Method and device for correcting cells in distorted tabular image

Publications (2)

Publication Number Publication Date
CN114119410A CN114119410A (en) 2022-03-01
CN114119410B true CN114119410B (en) 2022-04-22

Family

ID=80371612

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111401956.9A Active CN114119410B (en) 2021-11-19 2021-11-19 Method and device for correcting cells in distorted tabular image

Country Status (1)

Country Link
CN (1) CN114119410B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115082939B (en) * 2022-05-12 2023-11-17 吉林省吉林祥云信息技术有限公司 System and method for correcting distortion table in image based on arc differentiation
CN115937010B (en) * 2022-08-17 2023-10-27 北京字跳网络技术有限公司 Image processing method, device, equipment and medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108510445A (en) * 2018-03-30 2018-09-07 长沙全度影像科技有限公司 A kind of Panorama Mosaic method
CN110163198A (en) * 2018-09-27 2019-08-23 腾讯科技(深圳)有限公司 A kind of Table recognition method for reconstructing, device and storage medium
CN111931864A (en) * 2020-09-17 2020-11-13 南京甄视智能科技有限公司 Method and system for multiple optimization of target detector based on vertex distance and cross-over ratio
CN111989692A (en) * 2019-09-30 2020-11-24 北京市商汤科技开发有限公司 Form recognition method, form extraction method and related device
CN112733855A (en) * 2020-12-30 2021-04-30 科大讯飞股份有限公司 Table structuring method, table recovery equipment and device with storage function

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107249096B (en) * 2016-06-14 2021-02-26 杭州海康威视数字技术股份有限公司 Panoramic camera and shooting method thereof

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108510445A (en) * 2018-03-30 2018-09-07 长沙全度影像科技有限公司 A kind of Panorama Mosaic method
CN110163198A (en) * 2018-09-27 2019-08-23 腾讯科技(深圳)有限公司 A kind of Table recognition method for reconstructing, device and storage medium
CN111989692A (en) * 2019-09-30 2020-11-24 北京市商汤科技开发有限公司 Form recognition method, form extraction method and related device
CN111931864A (en) * 2020-09-17 2020-11-13 南京甄视智能科技有限公司 Method and system for multiple optimization of target detector based on vertex distance and cross-over ratio
CN112733855A (en) * 2020-12-30 2021-04-30 科大讯飞股份有限公司 Table structuring method, table recovery equipment and device with storage function

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
Fully automatic stitching and distortion correction of transmission electron microscope images;Verena Kaynig 等;《Journal of Structural Biology》;20100505;163-173 *
RDCFace: Radial Distortion Correction for Face Recognition;He Zhao 等;《2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)》;20200805;7718-7727 *
一种新的基于规则点阵的图像畸变校正方法;李欢 等;《仪器仪表学报》;20040831;第25卷(第4期);581-582 *
基于卷积网络的受电弓图像目标检测与矫正方法研究;庞鸿宇 等;《电气化铁道》;20211015;1-6 *
基于改进的Faster R-CNN的通用目标检测框架;马佳良 等;《计算机应用》;20210910;第21卷(第9期);2712-2719 *

Also Published As

Publication number Publication date
CN114119410A (en) 2022-03-01

Similar Documents

Publication Publication Date Title
CN114119410B (en) Method and device for correcting cells in distorted tabular image
US10867171B1 (en) Systems and methods for machine learning based content extraction from document images
CN108416902B (en) Real-time object identification method and device based on difference identification
US11816710B2 (en) Identifying key-value pairs in documents
AU2017232186A1 (en) Fast and robust image alignment for burst mode
US9082181B2 (en) Image processing method and apparatus
CN110874618B (en) OCR template learning method and device based on small sample, electronic equipment and medium
US11710210B1 (en) Machine-learning for enhanced machine reading of non-ideal capture conditions
WO2020060632A1 (en) Converting an image into a structured table
WO2020125062A1 (en) Image fusion method and related device
RU2581786C1 (en) Determination of image transformations to increase quality of optical character recognition
CN107305682B (en) Method and device for splicing images
CN113436222A (en) Image processing method, image processing apparatus, electronic device, and storage medium
WO2021168703A1 (en) Character processing and identifying methods, storage medium, and terminal device
CN112419207A (en) Image correction method, device and system
US10354161B2 (en) Detecting font size in a digital image
CN112052702A (en) Method and device for identifying two-dimensional code
CN114359932B (en) Text detection method, text recognition method and device
CN115223166A (en) Picture pre-labeling method, picture labeling method and device, and electronic equipment
CN112102203A (en) Image correction method, device and equipment
CN108304840B (en) Image data processing method and device
CN115797291A (en) Circuit terminal identification method and device, computer equipment and storage medium
CN113780294B (en) Text character segmentation method and device
CN111476090A (en) Watermark identification method and device
CN114648751A (en) Method, device, terminal and storage medium for processing video subtitles

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant