WO2022183907A1 - Procédé et appareil de traitement d'image, dispositif de reconnaissance de facture intelligent et support de stockage - Google Patents

Procédé et appareil de traitement d'image, dispositif de reconnaissance de facture intelligent et support de stockage Download PDF

Info

Publication number
WO2022183907A1
WO2022183907A1 PCT/CN2022/076400 CN2022076400W WO2022183907A1 WO 2022183907 A1 WO2022183907 A1 WO 2022183907A1 CN 2022076400 W CN2022076400 W CN 2022076400W WO 2022183907 A1 WO2022183907 A1 WO 2022183907A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
point
seal
unfolding
line
Prior art date
Application number
PCT/CN2022/076400
Other languages
English (en)
Chinese (zh)
Inventor
徐青松
李青
Original Assignee
杭州睿胜软件有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 杭州睿胜软件有限公司 filed Critical 杭州睿胜软件有限公司
Publication of WO2022183907A1 publication Critical patent/WO2022183907A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • G06V10/245Aligning, centring, orientation detection or correction of the image by locating a pattern; Special marks for positioning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/148Segmentation of character regions
    • G06V30/153Segmentation of character regions using recognition of characters or words
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/09Recognition of logos
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Definitions

  • Embodiments of the present disclosure relate to an image processing method, an image processing apparatus, an intelligent invoice recognition device, and a non-transitory computer-readable storage medium.
  • irregularly arranged characters are arcs, curved surfaces or have a perspective effect
  • the recognition of irregularly arranged characters is not accurate. Irregularly arranged text recognition has always been a technical difficulty in the field of text recognition.
  • At least one embodiment of the present disclosure provides an image processing method, including: acquiring an input image, wherein the input image includes an input seal, and the input seal includes a first object; identifying the input seal in the input image, To obtain a seal image, wherein, the seal image includes an intermediate seal corresponding to the input seal; feature extraction processing is performed on the seal image to obtain a feature point image; the seal image and the feature point image are processed.
  • the input seal is laterally expanded along the expansion line to obtain an expanded seal image; region recognition processing is performed on the expanded seal image to determine the first intermediate object region in the expanded seal image, wherein, The region corresponding to the first intermediate object region in the input image is the first object region, and the first object is located in the first object region; object recognition processing is performed on the first intermediate object region to obtain The identification obtains the first identification result.
  • the pixels corresponding to the middle seal in the seal image have a first pixel value
  • the seal image except for the pixels corresponding to the middle seal has a first pixel value
  • the pixels of have a second pixel value, and the first pixel value and the second pixel value are different.
  • identifying the input seal in the input image to obtain a seal image includes: using an image segmentation model to identify the input image to obtain The initial seal pixels corresponding to the input seal; the initial seal pixels are blurred to obtain a seal pixel mask area; according to the seal pixel mask area, determine the corresponding input seal in the input image. pixel; set the pixel value of the pixel corresponding to the input seal in the input image to the first pixel value and set the pixel value of the pixel other than the pixel corresponding to the input seal in the input image to the desired value. the second pixel value to obtain the stamp image.
  • the unfolding line includes a first annular unfolding line
  • the first annular unfolding line is an edge line of the input stamp.
  • the feature point image is processed to obtain the first unfolding point, the second unfolding point and the unfolding line, including: processing the seal image and the feature point image based on the algorithm of OpenCV to obtain the initial second unfolding point and an initial first annular unfolding line, wherein the initial second unfolding point and the initial first annular unfolding line are located in the seal image; processing to determine a characteristic object area corresponding to the first object in the seal image, determine an opening area in the seal image based on the characteristic object area, obtain any point in the opening area, and based on the
  • the arbitrary point and the initial first annular expansion line are used to determine the initial first expansion point, wherein the initial first expansion point is located in the seal image, and the arbitrary point, the initial first expansion point and all the The line segment between any two points in the initial second expansion point does not overlap with the feature object area, and
  • determining the initial first unfolding point based on the any point and the initial first annular unfolding line includes: obtaining the initial first unfolding point based on the any point A point corresponding to the any point on the initial first annular unfolding line is used as the initial first unfolding point.
  • the unfolding line includes a first annular unfolding line and a second annular unfolding line
  • the first annular unfolding line is an edge line of the input stamp
  • the second annular development line is located in an area surrounded by the first annular development line
  • the first object area is located in the area surrounded by the first annular development line and the second annular development line
  • the seal image and the feature point image are processed to obtain the first unfolding point, the second unfolding point and the unfolding line, including: using an algorithm based on OpenCV to analyze the seal image and the feature point image.
  • the extraction model processes the seal image and the feature point image to determine a feature object area corresponding to the first object in the seal image, and determines an opening area in the seal image based on the feature object area , obtain any point in the opening area, and determine the initial first expansion point and the initial second expansion point based on the arbitrary point, the initial first annular expansion line and the initial second annular expansion line, wherein , the initial first unfolding point and the initial second unfolding point are located in the stamp image, and any two points among the any point, the initial first unfolding point and the initial second unfolding point are located
  • the connecting line segment between is not overlapped with the feature object area; the initial first unfolding point, the initial second unfolding point, the initial first circular unfolding line and the initial second circular unfolding line are changed from The stamp image is mapped to the input image to obtain the first expansion point
  • an initial first development point and an initial first development point are determined.
  • Two expansion points including: based on the arbitrary point, acquiring a point on the initial first annular expansion line corresponding to the arbitrary point as the initial first expansion point; based on the arbitrary point, acquiring the initial first expansion point A point corresponding to the any point on the two-ring unfolding line is used as the initial second unfolding point.
  • the initial first unfolding point, the initial second unfolding point, and the any point are located on the same straight line.
  • the any point is a center point of the opening area.
  • the first annular expansion line is expanded into a straight line in the expanded seal image.
  • the first object is text
  • the shape of the first object area is an arc
  • the shape of the input stamp is a circle, and the second expansion point is the center of the circle; or, the shape of the input stamp is an ellipse, and the second expansion point is the midpoint of the line connecting the two foci of the ellipse.
  • the image processing method provided by an embodiment of the present disclosure further includes: determining a center point of the first intermediate object area; an intermediate object region is mapped back to the input image to determine the first object region, and a center point of the first intermediate object region is mapped back to the input image to determine a center point of the first object region; determining The center point of the input seal; the correction angle used to correct the input image is determined by the center point of the first object area and the center point of the input seal; the input image is corrected based on the correction angle Correction is performed to obtain a corrected input image.
  • the input stamp further includes a second object
  • the image processing method further includes: performing region recognition processing on the corrected input image to determine The second intermediate object area, wherein the area corresponding to the second intermediate object area in the input image is the second object area, and the second object is located in the second object area; The object area is subjected to object recognition processing to obtain a second recognition result.
  • acquiring an input image includes: acquiring an original image, wherein the original image includes an original seal; Determining a seal area, marking the seal area through a seal labeling frame, and slicing the seal labeling frame to obtain an intermediate input image, wherein the original seal is located in the seal area, and the seal labeling frame includes In the seal area, the intermediate input image includes the original seal, and the intermediate input image is processed to remove interference pixels in the intermediate input image to obtain the input image, wherein the interference pixels include Pixels of interfering objects in the intermediate input image that do not belong to the original seal, and the input seal corresponds to the original seal.
  • At least one embodiment of the present disclosure further provides an image processing apparatus, including: a memory for non-transitory storage of computer-readable instructions; and a processor for executing the computer-readable instructions, the computer-readable instructions being executed by The processor executes the image processing method according to any one of the above embodiments when running.
  • At least one embodiment of the present disclosure further provides an intelligent invoice recognition device, including: an image acquisition component for acquiring an invoice image of a paper invoice; a memory for storing the invoice image and computer-readable instructions; a processor for using upon reading the invoice image and determining the input image based on the invoice image, and executing the computer readable instructions, the computer readable instructions being executed by the processor to execute the above described embodiments image processing method.
  • At least one embodiment of the present disclosure further provides a non-transitory computer-readable storage medium for non-transitory storage of computer-readable instructions, which, when executed by a computer, can execute any of the foregoing embodiments. image processing method.
  • FIG. 1 is a schematic flowchart of an image processing method provided by some embodiments of the present disclosure
  • FIG. 2A is a schematic diagram of an original image provided by some embodiments of the present disclosure.
  • FIG. 2B is a schematic diagram of an intermediate input image determined based on the original image shown in FIG. 2A;
  • FIG. 2C is a schematic diagram of an input image obtained by identifying the intermediate input image shown in FIG. 2B;
  • FIG. 2D is a schematic diagram of a seal image obtained by recognizing the input image shown in FIG. 2C;
  • FIG. 2E is a schematic diagram of a feature point image obtained by performing feature extraction processing on the seal image shown in FIG. 2D;
  • 2F is a schematic diagram of another input image provided by some embodiments of the present disclosure.
  • FIG. 2G is another schematic diagram of a seal image obtained by recognizing the input image shown in FIG. 2C;
  • Fig. 2H is the schematic diagram of the expanded seal image obtained by expanding the input seal in Fig. 2C;
  • Fig. 2I is the schematic diagram of the first intermediate object region obtained by region recognition processing to the unfolded seal image in Fig. 2H;
  • 2J is a schematic diagram of a first recognition result obtained by performing object recognition processing on the first intermediate object region in FIG. 2I;
  • 3A is a schematic diagram of another original image provided by some embodiments of the present disclosure.
  • 3B is a schematic diagram of an intermediate input image determined based on the original image shown in FIG. 3A;
  • 3C is a schematic diagram of an input image obtained by identifying the intermediate input image shown in FIG. 3B;
  • 3D is a schematic diagram of a seal image obtained by recognizing the input image shown in FIG. 3C;
  • 3E is a schematic diagram of a feature point image obtained by performing feature extraction processing on the seal image shown in FIG. 3D;
  • 3F is a schematic diagram of an expanded seal image obtained by expanding the input seal in FIG. 3C;
  • 3G is a schematic diagram of a first intermediate object region obtained by performing region identification processing on the expanded seal image in FIG. 3F;
  • 3H is a schematic diagram of a first recognition result obtained by performing object recognition processing on the first intermediate object region in FIG. 3G;
  • FIG. 4 is a schematic block diagram of an image processing apparatus according to some embodiments of the present disclosure.
  • FIG. 5 is a schematic block diagram of an intelligent invoice recognition device according to some embodiments of the present disclosure.
  • FIG. 6 is a schematic diagram of a storage medium provided by some embodiments of the present disclosure.
  • seals such as official seals or invoice seals have irregularly arranged characters, for example, arc characters.
  • arc characters irregularly arranged characters
  • the recognition of these arc characters is not accurate.
  • the seal is tilted when stamping, it will also cause the regularly arranged characters in the image corresponding to the seal, such as horizontally arranged characters or vertically arranged characters, will also be slanted or reversed, making it impossible to judge the seal. forward direction, resulting in inaccurate recognition.
  • At least one embodiment of the present disclosure provides an image processing method, an image processing apparatus, an intelligent invoice recognition device, and a non-transitory computer-readable storage medium.
  • the image processing method includes: acquiring an input image, wherein the input image includes an input seal, and the input seal includes a first object; identifying the input seal in the input image to obtain a seal image, wherein the seal image includes an intermediate seal corresponding to the input seal; Perform feature extraction processing on the seal image to obtain the feature point image; process the seal image and the feature point image to obtain the first unfolding point, the second unfolding point and the unfolding line; take the difference between the first unfolding point and the second unfolding point.
  • the connecting line between them is used as the unfolding reference line and the first unfolding point is the unfolding starting point, and the input stamp is horizontally unfolded along the unfolding line to obtain the unfolding seal image;
  • the first intermediate object area wherein the area corresponding to the first intermediate object area in the input image is the first object area, and the first object is located in the first object area; object recognition processing is performed on the first intermediate object area, so as to obtain The first recognition result.
  • the image processing method can well realize the recognition of irregularly arranged objects (eg, characters, etc.) in the input image, improve the accuracy of identifying the irregularly arranged objects, and obtain accurate recognition results.
  • irregularly arranged objects eg, characters, etc.
  • “irregularly arranged objects” may mean that multiple objects (eg, characters) are not arranged in a row or column, that is, multiple objects are not arranged along the same straight line , for example, the centers of multiple objects are arranged along a curve (eg, a wavy line) or a polyline, etc.
  • the image processing method provided by the embodiment of the present disclosure can be applied to the image processing apparatus provided by the embodiment of the present disclosure, and the image processing apparatus can be configured on an electronic device.
  • the electronic device may be a personal computer, a mobile terminal, etc.
  • the mobile terminal may be a hardware device such as a mobile phone and a tablet computer.
  • FIG. 1 is a schematic flowchart of an image processing method provided by some embodiments of the present disclosure
  • FIG. 2A is a schematic diagram of an original image provided by some embodiments of the present disclosure
  • FIG. 2B is a determination based on the original image shown in FIG. 2A
  • the intermediate input image shown in FIG. 2C is an input image obtained by recognizing the intermediate input image shown in FIG. 2B
  • 3A is a schematic diagram of another original image provided by some embodiments of the present disclosure
  • FIG. 3B is an intermediate input image determined based on the original image shown in FIG. 3A
  • FIG. 3C is obtained by identifying the intermediate input image shown in FIG. 3B the input image.
  • step S10 of the image processing method provided by the embodiment of the present disclosure an input image is acquired.
  • the input image includes an input seal
  • the input seal may be various types of seals such as contract-specific seals, invoice-specific seals, and the like.
  • the input image can be any image that includes a seal, for example, as shown in FIG. 2C, in some embodiments, the input image can be an image including a company seal, as shown in FIG. 3C, in other embodiments, the input image can be For including the image of the special stamp for the invoice.
  • the input stamp may be a regular shape stamp such as a circular stamp, an oval stamp, a polygon stamp (for example, a rectangular stamp), or an irregular shape stamp.
  • the input image shown in FIG. 2C includes a circular seal
  • the input image shown in FIG. 3C includes an oval seal.
  • the input image may also be a document image or the like.
  • the input seal includes a first object
  • the first object may be a character
  • the character may be a number, a Chinese character (Chinese characters, Chinese words, etc.), foreign characters (eg, foreign letters, foreign words, etc., such as English, Japanese, Korean, German, etc.), special characters (eg, percent sign "%"), punctuation marks, etc.
  • the characters may also include graphics (eg, circles, rectangles, etc.), and the like.
  • the first object may be text. As shown in FIG. 2C and FIG. 3C , the first object may include a plurality of characters arranged irregularly, and the centers of the plurality of characters are arranged in a curve, for example , arranged in an arc.
  • the first object includes "Hangzhou Ruisheng Software Co., Ltd.”, and the centers of the characters in “Hangzhou Ruisheng Software Co., Ltd.” are arranged on an arc line; as shown in Fig. 3C, the first object An object includes “Hangzhou Ruisheng Software Co., Ltd.”, and the centers of the characters in “Hangzhou Ruisheng Software Co., Ltd.” are arranged on an elliptical arc.
  • step S10 includes: acquiring an original image; processing the original image through a seal area recognition model to determine the seal area, marking the seal area through the seal annotation frame, and slicing the seal annotation frame to Obtain an intermediate input image; process the intermediate input image to remove interfering pixels in the intermediate input image to obtain an input image.
  • both the original image and the intermediate input image include the original stamp, the original stamp is located within the stamp area, and the stamp callout box includes the stamp area.
  • the area of the original seal can be marked in the original image, and then the area corresponding to the original seal can be cut from the original image to obtain a separate intermediate input image, so that in subsequent operations, the cut intermediate input image can be directly obtained. to be processed.
  • the original image may be an image including a company seal
  • the intermediate input image shown in FIG. 2B can be obtained by processing the original image shown in FIG. 2A
  • the original image may be an image including a special seal for invoices
  • the intermediate input image shown in FIG. 3B can be obtained by processing the original image shown in FIG. 3A .
  • the stamp callout box can be a rectangular box, so that the intermediate input image can have a rectangular shape.
  • the dimensions of the stamp callout box can be the same as the dimensions of the intermediate input image.
  • the embodiments of the present disclosure are not limited to this, and the size of the seal annotation frame and the size of the intermediate input image may also be different.
  • the size of the seal annotation frame is larger than the size of the intermediate input image, that is, the intermediate input image is located in the seal annotation frame. inside the box.
  • seal marking frame may also be a diamond frame, an oval frame, a circular frame, and the like.
  • the seal area recognition model can be implemented using machine learning technology, and the seal area recognition model is a pre-trained model.
  • the seal region recognition model can be implemented by neural networks such as deep convolutional neural network (CNN) or deep residual network (Resnet).
  • CNN deep convolutional neural network
  • Resnet deep residual network
  • the size of the intermediate input image can be set by the user according to the actual situation.
  • the original image may be an image captured by a digital camera or a mobile phone, and the original image may be a grayscale image or a color image.
  • the original image may be an image directly collected by an image collection device, or may be an image obtained after preprocessing the directly collected image.
  • the image processing method provided by the embodiments of the present disclosure may further include an operation of preprocessing the original image. Preprocessing can eliminate irrelevant information or noise information in the original image, so as to better process the original image.
  • the preprocessing may include, for example, scaling, cropping, gamma correction, image enhancement, or noise reduction filtering on the original image.
  • acquiring the input image includes: acquiring the original image, processing the original image through the seal region recognition model to determine the intermediate input image; processing the intermediate input image to remove interference in the intermediate input image pixels to get the input image.
  • the area of the stamp annotation frame can be marked in the original image, and this area is the intermediate input image, so that in subsequent operations, the marked area can be directly processed, for example, to remove interference pixels. That is, the stamp callout box in the original image can not be cut.
  • the input stamp in the input image corresponds to the original stamp in the intermediate input image.
  • the size, shape, etc. of the input stamp and the original stamp are the same, except that the input stamp is located in the input image, and the original stamp is located in the original image and the intermediate input image.
  • some pixels of the original stamp may be removed, resulting in the original stamp and the input stamp being merged with each other.
  • the objects included in the original seal and the input image are the same, e.g. the original seal included the text "Hangzhou Ruisheng Software Co., Ltd.” and the input image also included the text "Hangzhou Ruisheng Software Co., Ltd.”.
  • interfering pixels include pixels of interfering objects in the intermediate input image that do not belong to the original stamp.
  • the interfering objects may include horizontal lines covered by the original seal or characters or numbers on the date of the seal, and may also include other characters, numbers or graphics that do not overlap with the original seal.
  • the interfering objects may be printed words, symbols, graphics, etc.
  • the intermediate input image includes horizontal lines and commas that do not belong to the original seal, and the horizontal lines and commas are Interfering objects, the pixel corresponding to the horizontal line and comma is the interference pixel.
  • the interfering objects can be handwritten words, symbols, graphics, etc.
  • the input image includes numbers and points that do not belong to the original seal (ie, handwritten 2021.1.7), the numbers and The point is the interference object, and the pixel corresponding to the number and the point is the interference pixel.
  • processing the intermediate input image to remove interfering pixels in the intermediate input image to obtain the input image may include: using an image segmentation model (such as U-Net model, Mask-RCNN model, etc.)
  • the input image is identified to obtain the initial interference pixels of the interference object; the initial interference pixels are blurred to obtain the interference pixel mask area; the interference pixels corresponding to the interference object are determined according to the interference pixel mask area; the interference pixels in the intermediate input image are removed.
  • Interfering pixels corresponding to interfering objects are obtained to obtain the input image.
  • the process of identifying and removing the interfering pixels may be performed based on the difference between the pixel value of the pixel corresponding to the interfering object and the pixel value of the pixel corresponding to the original seal.
  • Gaussian blurring can be performed on the initial interference pixels through the GaussianBlur function of Gaussian filtering based on OpenCV to expand the area corresponding to the initial interference pixels, thereby obtaining the interference pixel mask area.
  • the interference pixel corresponding to the interference object can be determined.
  • the interfering pixels corresponding to the interfering objects in the intermediate input image can be removed by the inpaint function based on OpenCV, so as to obtain an image from which the interfering objects are removed, that is, the input image is obtained.
  • FIG. 2D is a schematic diagram of a seal image obtained by recognizing the input image shown in FIG. 2C
  • FIG. 3D is a schematic diagram of a seal image obtained by recognizing the input image shown in FIG. 3C .
  • step S11 the input seal in the input image is recognized to obtain a seal image.
  • the seal image includes an intermediate seal corresponding to the input seal.
  • the size and shape of the input stamp and the intermediate stamp are the same.
  • the objects included in the input stamp and the objects included in the intermediate stamp and their relative positional relationships are also the same.
  • the difference between the input stamp and the intermediate stamp is the same. where: the input stamp is located in the input image, and the intermediate stamp is located in the stamp image.
  • step S11 includes: using an image segmentation model (such as U-Net model, Mask-RCNN model, etc.) to identify the input image to obtain initial seal pixels corresponding to the input seal; Blur processing to obtain the seal pixel mask area; determine the pixel corresponding to the input seal in the input image according to the seal pixel mask area; set the pixel value of the pixel corresponding to the input seal in the input image to the first pixel value and set the input
  • the pixel values of the pixels other than the pixels corresponding to the input seal in the image are the second pixel values, so as to obtain the seal image.
  • the seal image can be a black and white image with obvious black and white contrast, and the black and white image has less noise interference, which can effectively improve the recognition of the content in the seal image.
  • the pixels corresponding to the middle seal in the seal image have the first pixel value
  • the pixels in the seal image except the pixels corresponding to the middle seal have the second pixel value
  • the first pixel value and the second pixel value are different.
  • both the first pixel value and the second pixel value may be grayscale values
  • the first pixel value may be 255
  • the second pixel value may be 0.
  • both the image segmentation model for recognizing input images and the image segmentation model for recognizing intermediate input images can be implemented using machine learning technology (eg, deep learning technology), and both are pre-trained models.
  • the image segmentation model for recognizing the input image and the image segmentation model for recognizing the intermediate input image can be two different models, but both adopt the U-Net model structure.
  • FIG. 2E is a feature point image obtained by performing feature extraction processing on the seal image shown in FIG. 2D
  • FIG. 3E is a feature point image obtained by performing feature extraction processing on the seal image shown in FIG. 3D .
  • step S12 feature extraction processing is performed on the seal image to obtain feature point images.
  • step S12 feature extraction processing may be performed on the seal image through a pre-trained feature extraction model to obtain feature point images.
  • Feature extraction models can also be implemented based on machine learning techniques.
  • feature extraction processing is performed on the seal image shown in FIG. 2D to obtain the feature point image shown in FIG. 2E
  • feature extraction processing is performed on the seal image shown in FIG. 3D to obtain the feature point image shown in FIG. 3E .
  • the first object includes "Hangzhou Ruisheng Software Co., Ltd.”
  • the feature point image shown in FIG. 2E includes 11 features.
  • the 11 feature points correspond to each character in "Hangzhou Ruisheng Software Co., Ltd.” and the center point of the middle seal. For each character, the feature point corresponding to the character is located in the center of the region corresponding to the character.
  • each of the original image, the intermediate input image, the input image and the seal image includes the first object "Hangzhou Ruisheng Software Co., Ltd.”.
  • the image segmentation model is established by processing the input image or intermediate input image into a black and white image and labeling the sample, and then putting it into the U-net model for training; the feature extraction model is also by using the seal image as a sample. After labeling, it is established by training the neural network model.
  • FIG. 2F is a schematic diagram of another input image provided by an embodiment of the present disclosure.
  • step S13 the stamp image and the feature point image are processed to obtain the first development point, the second development point and the development line.
  • the unfolding line includes a first annular unfolding line
  • the first annular unfolding line is the edge line of the input stamp.
  • the edge line of the input stamp may be the circle shown in FIG. 2C .
  • the edge line of the input stamp may be an elliptical circle as shown in FIG. 3C .
  • step S13 includes: processing the seal image and the feature point image based on the algorithm of OpenCV to obtain the initial second expansion point and the initial first annular expansion line; processing the seal image and the feature point image through the expansion point extraction model, Determine the characteristic object area corresponding to the first object in the seal image, determine the opening area in the seal image based on the characteristic object area, and obtain any point in the opening area; an expansion point; the initial first expansion point, the initial second expansion point and the initial first circular expansion line are mapped from the stamp image to the input image to obtain the first expansion point, the second expansion point and the first circular expansion line.
  • the initial first unfolding point, the initial second unfolding point, and the initial first annular unfolding line are all located in the stamp image.
  • the connecting line segment between any two points among any point, the initial first unfolding point and the initial second unfolding point does not overlap with the feature object area, so that it can be ensured that when the input stamp is horizontally unfolded, the first The object is split into two parts.
  • the area 100 shown in FIG. 2D is the characteristic object area.
  • the first object "Hangzhou Ruisheng Software Co., Ltd.” is located in the characteristic object area.
  • the region corresponding to the characteristic object region 100 in the input image is the first object region 200 , and in the input image, the first object is located in the first object region 200 .
  • the shape of the characteristic object area 100 may be an arc, and the shape of the first object area 200 may also be an arc.
  • the initial first annular development line may be the edge line of the middle seal shown in FIG. 2D
  • the edge line of the middle seal may be the white circle shown in FIG. 2D .
  • an annular area may be determined based on the characteristic object area 100 , for example, an annular area, the annular area includes the characteristic object area 100 , and the part of the annular area that does not belong to the characteristic object area 100 is is the opening area 110 , and the arbitrary point B is a point in the opening area 110 .
  • the seal image and the feature point image may be processed based on the Hough gradient circle finding algorithm using OpenCV to obtain the initial second expansion point and the initial first annular expansion line.
  • Hough gradient circle finding algorithm of OpenCV For the specific implementation process of the Hough gradient circle finding algorithm of OpenCV, reference may be made to relevant descriptions in the prior art, and details are not described here. It should be noted that, in the embodiments of the present disclosure, other methods may also be used to obtain the initial second deployment point and the initial first annular deployment line. No restrictions apply.
  • the expansion point extraction model may be implemented based on machine learning, and the expansion point extraction model may be a neural network model.
  • determining the initial first expansion point based on any point and the initial first annular expansion line includes: based on any point, acquiring a point corresponding to any point on the initial first annular expansion line as the initial first expansion point.
  • the initial second unfolding point may be the center point of the middle seal.
  • the shape of the middle seal is a circle
  • the initial second unfolding point A1 is the center of the circle (ie The center point of the middle seal)
  • the initial first unfolding point C1 may be the intersection between the extension line connecting the initial second unfolding point A1 and any point B1 and the initial first annular unfolding line.
  • the second unfolding point may be the center point of the input stamp, eg, in some embodiments 2F
  • the initial first expansion point C1 is mapped to the first expansion point C2
  • the initial second expansion point A1 is mapped to the second expansion point A2
  • any point B1 is mapped to the point B2.
  • the shape of the input stamp is a circle
  • the second expansion point A2 is the center of the circle (ie, the center point of the input stamp).
  • the first expansion point C2 can be the connection between the second expansion point A2 and the point B2. The intersection between the extension line of the line and the first annular expansion line.
  • the shape of the middle seal is an ellipse
  • the initial second expansion point may be the midpoint (not shown) of the line connecting the two focal points of the ellipse.
  • the shape of the input stamp is also an ellipse, and after mapping, the second expansion point is also the midpoint of the line connecting the two focal points of the ellipse.
  • FIG. 2G is another schematic diagram of a seal image obtained by recognizing the input image shown in FIG. 2C .
  • the unfolding line includes a first annular unfolding line and a second annular unfolding line
  • the first annular unfolding line is an edge line of the input stamp
  • the second annular unfolding line is located in the first annular unfolding line.
  • the first object area is located in the annular area enclosed by the first annular expansion line and the second annular expansion line.
  • step S13 includes: processing the seal image and the feature point image based on the algorithm of OpenCV to obtain the initial first annular expansion line and the initial second annular expansion line; processing the seal image and the feature point image through the expansion point extraction model , to determine the characteristic object area corresponding to the first object in the seal image, determine the opening area in the seal image based on the characteristic object area, obtain any point in the opening area, and based on any point, the initial first annular expansion line and the initial first Two circular expansion lines, determine the initial first expansion point and the initial second expansion point; map the initial first expansion point, the initial second expansion point, the initial first circular expansion line and the initial second circular expansion line from the stamp image to An image is input to obtain a first unfolding point, a second unfolding point, a first circular unfolding line, and a second circular unfolding line.
  • the initial first annular development line, the initial second annular development line, the initial first development point, and the initial second development point are all located in the stamp image.
  • the connecting line segment between any two points among any point, the initial first unfolding point and the initial second unfolding point does not overlap with the feature object area.
  • the white circle 300 may be the initial second annular expansion line
  • the area 100 shown in FIG. 2G is the feature object area
  • the area 110 shown in FIG. 2G is the opening area.
  • step S13 determining the initial first expansion point and the initial second expansion point based on any point, the initial first annular expansion line and the initial second annular expansion line, including: obtaining the initial first annular expansion based on any point A point on the line corresponding to any point is used as an initial first expansion point; based on any point, a point corresponding to any point on the initial second annular expansion line is obtained as an initial second expansion point.
  • the any point B1 is a point in the opening area 110
  • the shape of the middle seal is a circle
  • the initial first expansion point C1 can be the radius of the circle including any point B1 and the initial first
  • the initial second development point A1 may be the intersection between the radius of the circle including any point B1 and the initial second annular development line 300 .
  • any point B1 may be the center point of the opening area.
  • the initial first unfolding point C1, the initial second unfolding point A1 and any point B1 are located on the same straight line, as shown in FIG. 2F, after the mapping, the first unfolding point C2, the second unfolding point A2 and the point B2 are also on the same straight line.
  • the initial first unfolding point C1 , the initial second unfolding point A1 and the arbitrary point B1 are located on a radius of the middle seal of the circle.
  • the distance between the initial first unfolding point C1 and the initial second unfolding point A1 is the radius of the middle seal of the circle.
  • the first unfolding point C2 the second unfolding point A2 and the point B2 are located on a radius of the circular input stamp, the first unfolding point C2 and the second unfolding point A2 The distance between is the radius of the input stamp for that circle.
  • first annular development line and the second annular development line are described as circles or elliptical circles as an example, the present disclosure is not limited to this, the first annular development line and/or the second annular development line are
  • the annular unfolding line may also be an unclosed arc or curve, and its specific shape is related to the shape of the first object area including the first object. For example, if the shape of the first object area is wavy, the first annular The unfolding line and/or the second annular unfolding line may also be a wavy line.
  • the first annular development line can also be a concentric ring line of the edge line of the input seal.
  • the edge line of the input seal is located in the area surrounded by the first annular development line.
  • the shape of the input seal is a circle
  • the shape of the first annular development line and the shape of the edge line of the input stamp may be the same, for example, both are circular, but the radius corresponding to the first annular development line is larger than the radius corresponding to the edge line of the input stamp.
  • FIG. 2H is a schematic diagram of an expanded seal image obtained by expanding the input seal in FIG. 2C
  • FIG. 3F is a schematic diagram of an expanded seal image obtained by expanding the input seal in FIG. 3C .
  • step S14 take the connecting line between the first development point and the second development point as the development reference line and the first development point as the development starting point, and place the input stamp along the development line Expand horizontally for expanded stamp image.
  • the unfolded seal image shown in FIG. 2H is based on the first unfolding point obtained by mapping the initial first unfolding point shown in FIG. 2G and the second unfolding point obtained by mapping the initial second unfolding point shown in FIG. 2G .
  • the connecting line is used as the unfolding reference line and the first unfolding point obtained by mapping the initial first unfolding point shown in FIG. 2G as the unfolding starting point, and is obtained by horizontally unfolding along the first annular unfolding line.
  • the first annular expansion line is expanded into a straight line in the expanded stamp image.
  • the straight line above the text is the expanded first circular expansion line; as shown in Figure 3F, the line above the text
  • the straight line is the first annular expansion line after expansion.
  • the second expansion point is also a point corresponding to any point on the second annular expansion line.
  • the shape of the expanded stamp image is a rectangle, the length of the rectangle is equal to the length of the first annular expansion line, and the width of the rectangle is the same as the first expansion point and the point obtained based on the mapping of any point (that is, the initial first The distance between the expansion point and any point) is equal.
  • the shape of the middle seal is a circle, the length of the rectangle is equal to the circumference of the circle, and the width of the rectangle is equal to the radius of the circle; for the example shown in Fig. 2G, the length of the middle seal is equal to the radius of the circle.
  • the shape is a circle, the length of the rectangle is equal to the circumference of the circle, and the width of the rectangle is less than the radius of the circle.
  • the connecting line between the initial first unfolding point and the initial second unfolding point may also be used as the unfolding reference line and the initial first unfolding point may be used as the unfolding starting point
  • the middle seal Expand horizontally along the expansion line to obtain the expanded stamp image, that is, at this time, the initial first expansion point is the first expansion point, the initial second expansion point is the second expansion point, and the initial first circular expansion line is the first expansion point.
  • the initial second annular expansion line is the second annular expansion line.
  • Fig. 2I is the schematic diagram of the first intermediate object region obtained by carrying out region recognition processing to the expanded seal image in Fig. 2H
  • Fig. 3G is the schematic diagram of the first intermediate object region obtained by performing region identification processing on the expanded seal image in Fig. 3F.
  • step S15 an area identification process is performed on the developed seal image to determine the first intermediate object area in the expanded seal image. For example, an area corresponding to the first intermediate object area in the input image is the first object area, and the first object is located in the first object area.
  • the shape of the first object area is an arc.
  • the shape of the first intermediate object area may be a rectangle. It should be noted that the specific unfolding methods of the circular seal and the oval seal may refer to the prior art, which will not be repeated here.
  • the first intermediate object region in the expanded stamp image can be identified by the region identification model.
  • the region recognition model can be implemented using machine learning technology, and the region recognition model is a pre-trained model.
  • the region recognition model can be implemented by neural networks such as deep convolutional neural network (CNN) or deep residual network (Resnet).
  • FIG. 2J is a schematic diagram of a first recognition result obtained by performing object recognition processing on the first intermediate object area in FIG. 2I ;
  • FIG. 3H is a schematic diagram of a first recognition result obtained by performing object recognition processing on the first intermediate object area in FIG. 3G .
  • step S16 an object recognition process is performed on the first intermediate object region to recognize and obtain a first recognition result.
  • the first recognition result is “Hangzhou Ruisheng Software Co., Ltd.”, that is, the first object.
  • the first object includes text
  • character recognition processing may be performed on the first intermediate object region through the first character recognition model to obtain the first recognition result, that is, the first object.
  • the accuracy of character recognition based on the first character recognition model is high.
  • the first character recognition model may be implemented based on technologies such as optical character recognition (Optical Character Recognition, OCR).
  • OCR Optical Character Recognition
  • the first character recognition model may also be a pre-trained model.
  • performing object recognition processing on the first intermediate object area to recognize and obtain the first recognition result may include: performing object recognition processing on the first intermediate object area to recognize and obtain the first intermediate recognition result; A check is performed to obtain the first identification result.
  • the first intermediate recognition result may have semantic errors, logical errors, etc. Therefore, it is necessary to verify the first intermediate recognition result, and correct the semantic errors and logical errors in the first intermediate recognition result, so as to obtain an accurate first intermediate recognition result. Identify the results.
  • the first intermediate recognition result may include "Hangzhou Ruisheng Software Co., Ltd.”, wherein the character “zhou” does not correspond to the text in the seal, and the word “Hangzhou” is in The semantics is wrong. After verification, "Hangzhou” can be corrected to "Hangzhou", so the first recognition result after verification is "Hangzhou Ruisheng Software Co., Ltd.”, thus obtaining an accurate recognition result .
  • the first identification result obtained by identification is "Hangzhou Ruisheng Software Co., Ltd.”, which is the first object in the input seal.
  • the image processing method can also determine the forward direction of the input image based on the regions corresponding to the irregularly arranged objects, so as to correct the input image and improve the recognition accuracy of the regularly arranged objects in the input image.
  • the image processing method further includes: determining the center point of the first intermediate object region; mapping the first intermediate object region back to the input image through the mapping relationship between the first intermediate object region and the input image to determine the first intermediate object region an object area, and mapping the center point of the first intermediate object area back to the input image to determine the center point of the first object area; determining the center point of the input seal; determining by the center point of the first object area and the center point of the input seal Correction angle for correcting the input image; correcting the input image based on the correction angle to obtain the corrected input image.
  • the first intermediate object region in the expanded seal image can be identified by the region recognition model, and the center point of the first intermediate object region can be determined, and the An intermediate object area is mapped into the input image (or the seal image), thereby determining the arc-shaped character area in the input image (or the seal image), that is, the first object area 200, and at the same time, the center point of the first intermediate object area is mapped to
  • the center point of the first object area is determined in the input image (or seal image), and the forward direction corresponding to the input image can be obtained through the center point of the first object area and the center point of the input seal (for example, the center of the circle of the input seal).
  • the angle between the forward direction and the reference direction (for example, the horizontal direction or the vertical direction) is the correction angle, Then, the input image can be corrected based on the correction direction, so that the forward direction and the reference direction overlap, so as to obtain the corrected input image, thus, it is convenient for the user to check and compare whether the first recognition result obtained by the recognition is correct, etc. , that is, it is determined whether the first recognition result obtained by the recognition is the same as the first object.
  • the first intermediate object region may also be mapped back to the intermediate input image, and a correction angle for correcting the intermediate input image is determined; and the intermediate input image is corrected based on the correction angle to obtain the correction After the intermediate input image.
  • the user can check whether the first recognition result obtained by the comparison and recognition based on the corrected intermediate input image is correct, etc.
  • the input stamp further includes a second object.
  • the image processing method further includes: performing region identification processing on the corrected input image to determine the second intermediate object region, wherein the region corresponding to the second intermediate object region in the input image is the second object region, and the second intermediate object region is the second object region.
  • the object is located in the second object area; the object recognition processing is performed on the second intermediate object area to obtain a second recognition result.
  • the shape of the second object area may be a rectangle.
  • the second object may include a plurality of characters arranged regularly, and a line connecting the center points of the plurality of characters is located on the same straight line.
  • the second object may include numbers and letters "91330108MA2CDKJ756", and the second object may also include the text "Invoice Special Seal”.
  • the center points of each character (numbers and letters) in "91330108MA2CDKJ756" are located on the same line (such as a horizontal line), and the center points of each character in the "Special Invoice Seal" are also located on the same line (such as a horizontal line).
  • character recognition processing can be performed on the second intermediate object area through the second character recognition model to obtain the second intermediate recognition result; the second intermediate recognition result is verified to obtain the second recognition result, and the second recognition result is for the second object.
  • the second character recognition model may be implemented based on technologies such as optical character recognition.
  • the second character recognition model may also be a pre-trained model.
  • first character recognition model and the second character recognition model may be the same model, or may be different models.
  • the image processing method may further include: outputting the first recognition result and the second recognition result.
  • the first recognition result and the second recognition result may be displayed on the display panel to achieve output.
  • the image processing method may further include: outputting the corrected input image and/or the corrected intermediate input image, so that the user can judge whether the outputted first recognition result and the second recognition result are correct.
  • the corrected input image and/or the corrected intermediate input image may also be displayed on the display panel for output.
  • the image processing method before acquiring the input image, the image processing method further includes: a training phase.
  • the training phase includes the process of training the models (image segmentation model, region recognition model, expansion point extraction model, seal region recognition model, character recognition model, etc.).
  • FIG. 4 is a schematic block diagram of an image processing apparatus according to some embodiments of the present disclosure.
  • At least one embodiment of the present disclosure further provides an image processing apparatus.
  • the image processing apparatus 400 includes a processor 402 and a memory 401 . It should be noted that the components of the image processing apparatus 400 shown in FIG. 4 are only exemplary and not restrictive, and the image processing apparatus 400 may also have other components according to actual application requirements.
  • the memory 401 is used for non-transitory storage of computer-readable instructions; the processor 402 is used for executing computer-readable instructions, and the computer-readable instructions are executed by the processor 402 when running the image processing method according to any of the above embodiments. one or more steps.
  • the network may include a wireless network, a wired network, and/or any combination of wireless and wired networks.
  • the network may include a local area network, the Internet, a telecommunication network, the Internet of Things (Internet of Things) based on the Internet and/or a telecommunication network, and/or any combination of the above networks, etc.
  • the wired network may use twisted pair, coaxial cable or optical fiber transmission for communication
  • the wireless network may use, for example, 3G/4G/5G mobile communication network, Bluetooth, Zigbee or WiFi and other communication methods.
  • the present disclosure does not limit the type and function of the network.
  • processor 402 may control other components in image processing apparatus 400 to perform desired functions.
  • the processor 402 may be a device with data processing capability and/or program execution capability, such as a central processing unit (CPU), a tensor processing unit (TPU), or a graphics processing unit (GPU).
  • the central processing unit (CPU) can be an X86 or an ARM architecture or the like.
  • the GPU can be individually integrated directly onto the motherboard, or built into the motherboard's Northbridge chip. GPUs can also be built into central processing units (CPUs).
  • memory 401 may include any combination of one or more computer program products, which may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory.
  • Volatile memory may include, for example, random access memory (RAM) and/or cache memory, among others.
  • Non-volatile memory may include, for example, read only memory (ROM), hard disk, erasable programmable read only memory (EPROM), portable compact disk read only memory (CD-ROM), USB memory, flash memory, and the like.
  • ROM read only memory
  • EPROM erasable programmable read only memory
  • CD-ROM portable compact disk read only memory
  • USB memory flash memory
  • One or more computer-readable instructions may be stored on the computer-readable storage medium, and the processor 402 may execute the computer-readable instructions to implement various functions of the image processing apparatus 400 .
  • Various application programs, various data and the like can also be stored in the storage medium.
  • FIG. 5 is a schematic block diagram of an intelligent invoice recognition device provided by some embodiments of the present disclosure.
  • the intelligent invoice recognition device 500 may include a memory 501 , a processor 502 and an image acquisition component 503 . It should be noted that the components of the smart invoice recognition device 500 shown in FIG. 5 are only exemplary, not limiting, and the smart invoice recognition device 500 may also have other components according to actual application requirements.
  • the image acquisition part 503 is used to acquire an invoice image of a paper invoice.
  • Memory 501 is used to store invoice images and computer readable instructions.
  • Processor 502 operates to read the invoice image and determine an input image based on the invoice image and execute computer readable instructions.
  • the computer readable instructions are executed by the processor 502 to perform one or more steps in the image processing method according to any of the above embodiments.
  • the invoice image may be the original image described in the embodiment of the image processing method.
  • the image acquisition component 503 is the image acquisition device described in the embodiments of the above image processing method.
  • the image acquisition component 503 may be a camera of a smartphone, a camera of a tablet computer, a camera of a personal computer, a lens of a digital camera, Or even a webcam.
  • the image of the invoice may be the image of the original invoice directly collected by the image acquisition component 503, or may be the image obtained after preprocessing the image of the original invoice.
  • Preprocessing can remove irrelevant information or noise information in the original invoice image to facilitate better processing of the invoice image.
  • the preprocessing may include, for example, performing image augmentation (Data Augment), image scaling, gamma (Gamma) correction, image enhancement or noise reduction filtering on the original invoice image.
  • processor 502 may control other components in intelligent invoice recognition device 500 to perform desired functions.
  • the processor 502 may be a device with data processing capability and/or program execution capability, such as a central processing unit (CPU), a tensor processing unit (TPU), or a graphics processing unit (GPU).
  • the central processing unit (CPU) can be an X86 or an ARM architecture or the like.
  • the GPU can be individually integrated directly onto the motherboard, or built into the motherboard's Northbridge chip. GPUs can also be built into central processing units (CPUs).
  • memory 501 may include any combination of one or more computer program products, which may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory.
  • Volatile memory may include, for example, random access memory (RAM) and/or cache memory, among others.
  • Non-volatile memory may include, for example, read only memory (ROM), hard disk, erasable programmable read only memory (EPROM), portable compact disk read only memory (CD-ROM), USB memory, flash memory, and the like.
  • ROM read only memory
  • EPROM erasable programmable read only memory
  • CD-ROM portable compact disk read only memory
  • USB memory flash memory, and the like.
  • One or more computer-readable instructions may be stored on the computer-readable storage medium, and the processor 502 may execute the computer-readable instructions to implement various functions of the intelligent invoice recognition device 500.
  • FIG. 6 is a schematic diagram of a storage medium provided by some embodiments of the present disclosure.
  • one or more computer-readable instructions 601 may be non-transitory stored on storage medium 600 .
  • the computer readable instructions 601 when executed by a computer, one or more steps in the image processing method according to the above description may be performed.
  • storage medium 600 is a non-transitory computer-readable storage medium.
  • the storage medium 600 can be applied to the above-mentioned image processing apparatus 400 and/or the smart invoice recognition apparatus 500 , for example, it can be the memory 401 in the image processing apparatus 400 and/or the memory 501 in the smart invoice recognition apparatus 500 .
  • the description of the storage medium 600 reference may be made to the description of the memory in the embodiments of the image processing apparatus 400 and/or the smart invoice recognition device 500, and the repetition will not be repeated.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

Au moins un mode de réalisation de la présente divulgation concerne un procédé de traitement d'image, un appareil de traitement d'image, un dispositif de reconnaissance de facture intelligent et un support de stockage. Le procédé de traitement d'image consiste à : obtenir une image d'entrée, l'image d'entrée comprenant un cachet d'entrée ; reconnaître le cachet d'entrée dans l'image d'entrée pour obtenir une image de cachet, l'image de cachet comprenant un cachet intermédiaire correspondant au cachet d'entrée ; effectuer un traitement d'extraction de caractéristiques sur l'image de cachet pour obtenir une image de point caractéristique ; traiter l'image de cachet et l'image de point caractéristique pour obtenir un premier point de dépliage, un second point de dépliage et une ligne de dépliage ; en utilisant la ligne de liaison entre le premier point de dépliage et le second point de dépliage en tant que ligne de référence de dépliage et le premier point de dépliage en tant que point de départ de dépliage, déplier le cachet d'entrée horizontalement le long de la ligne de dépliage pour obtenir une image de cachet déplié ; effectuer un traitement de reconnaissance de zone sur l'image de cachet déplié pour déterminer une première zone d'objet intermédiaire ; et effectuer un traitement de reconnaissance d'objet sur la première zone d'objet intermédiaire pour obtenir un premier résultat de reconnaissance.
PCT/CN2022/076400 2021-03-04 2022-02-16 Procédé et appareil de traitement d'image, dispositif de reconnaissance de facture intelligent et support de stockage WO2022183907A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110241290.9 2021-03-04
CN202110241290.9A CN113033325A (zh) 2021-03-04 2021-03-04 图像处理方法及装置、智能发票识别设备和存储介质

Publications (1)

Publication Number Publication Date
WO2022183907A1 true WO2022183907A1 (fr) 2022-09-09

Family

ID=76467781

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/076400 WO2022183907A1 (fr) 2021-03-04 2022-02-16 Procédé et appareil de traitement d'image, dispositif de reconnaissance de facture intelligent et support de stockage

Country Status (2)

Country Link
CN (1) CN113033325A (fr)
WO (1) WO2022183907A1 (fr)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113033325A (zh) * 2021-03-04 2021-06-25 杭州睿胜软件有限公司 图像处理方法及装置、智能发票识别设备和存储介质

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105631447A (zh) * 2015-12-18 2016-06-01 杭州仁盈科技股份有限公司 一种识别圆形公章中文字的方法
CN110659647A (zh) * 2019-09-11 2020-01-07 杭州睿琪软件有限公司 印章图像识别方法及装置、智能发票识别设备和存储介质
CN111178355A (zh) * 2019-12-27 2020-05-19 中化资本有限公司 印章识别方法、装置和存储介质
CN111767911A (zh) * 2020-06-22 2020-10-13 平安科技(深圳)有限公司 面向复杂环境的印章文字检测识别方法、装置及介质
CN113033325A (zh) * 2021-03-04 2021-06-25 杭州睿胜软件有限公司 图像处理方法及装置、智能发票识别设备和存储介质

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7209161B2 (ja) * 2019-05-20 2023-01-20 京都電子計算株式会社 文字読取装置及び文字読取装置用プログラム
CN110443250B (zh) * 2019-07-31 2022-06-10 天津车之家数据信息技术有限公司 一种合同印章的类别识别方法、装置和计算设备

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105631447A (zh) * 2015-12-18 2016-06-01 杭州仁盈科技股份有限公司 一种识别圆形公章中文字的方法
CN110659647A (zh) * 2019-09-11 2020-01-07 杭州睿琪软件有限公司 印章图像识别方法及装置、智能发票识别设备和存储介质
CN111178355A (zh) * 2019-12-27 2020-05-19 中化资本有限公司 印章识别方法、装置和存储介质
CN111767911A (zh) * 2020-06-22 2020-10-13 平安科技(深圳)有限公司 面向复杂环境的印章文字检测识别方法、装置及介质
CN113033325A (zh) * 2021-03-04 2021-06-25 杭州睿胜软件有限公司 图像处理方法及装置、智能发票识别设备和存储介质

Also Published As

Publication number Publication date
CN113033325A (zh) 2021-06-25

Similar Documents

Publication Publication Date Title
CN110659647B (zh) 印章图像识别方法及装置、智能发票识别设备和存储介质
US9811749B2 (en) Detecting a label from an image
CN110866495A (zh) 票据图像识别方法及装置和设备、训练方法和存储介质
US11823358B2 (en) Handwritten content removing method and device and storage medium
US20230222631A1 (en) Method and device for removing handwritten content from text image, and storage medium
CN113486828B (zh) 图像处理方法、装置、设备和存储介质
US9959475B2 (en) Table data recovering in case of image distortion
JP2014524085A (ja) オブジェクトにおけるテキスト情報を認識するための方法およびシステム
WO2020097909A1 (fr) Procédé et appareil de détection de texte et support de stockage
WO2022134771A1 (fr) Procédé et appareil de traitement de table, dispositif électronique et support de stockage
CN112819720B (zh) 图像处理方法、装置、电子设备及存储介质
CN111709414A (zh) Ar设备及其文字识别方法、装置和计算机可读存储介质
WO2022166833A1 (fr) Procédé et appareil de traitement d'image, ainsi que dispositif électronique et support de stockage
WO2022183907A1 (fr) Procédé et appareil de traitement d'image, dispositif de reconnaissance de facture intelligent et support de stockage
CN112149561A (zh) 图像处理方法和装置、电子设备和存储介质
CN113436222A (zh) 图像处理方法、图像处理装置、电子设备及存储介质
WO2022166707A1 (fr) Procédé et appareil de traitement d'image, dispositif électronique et support de stockage
CN116597466A (zh) 一种基于改进YOLOv5s的工程图纸文字检测识别方法及系统
CN111767924B (zh) 图像处理方法、图像处理装置、电子设备、存储介质
KR20160146355A (ko) 이미지 내의 텍스트를 검출하는 방법 및 장치
CN115527215A (zh) 包含文本的图像处理方法、系统及存储介质
CN111914840A (zh) 一种文本识别方法、模型训练方法、装置及设备
CN112861861A (zh) 识别数码管文本的方法、装置及电子设备
CN114627457A (zh) 一种票面信息识别方法及装置
US9020295B1 (en) Image glyph enhancement

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22762373

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22762373

Country of ref document: EP

Kind code of ref document: A1