CN113095327A - Method and system for positioning optical character recognition area and storage medium thereof - Google Patents

Method and system for positioning optical character recognition area and storage medium thereof Download PDF

Info

Publication number
CN113095327A
CN113095327A CN202110280359.9A CN202110280359A CN113095327A CN 113095327 A CN113095327 A CN 113095327A CN 202110280359 A CN202110280359 A CN 202110280359A CN 113095327 A CN113095327 A CN 113095327A
Authority
CN
China
Prior art keywords
image
binary image
axis
width
height
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110280359.9A
Other languages
Chinese (zh)
Other versions
CN113095327B (en
Inventor
丁凡
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Emperor Technology Co Ltd
Original Assignee
Shenzhen Emperor Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Emperor Technology Co Ltd filed Critical Shenzhen Emperor Technology Co Ltd
Priority to CN202110280359.9A priority Critical patent/CN113095327B/en
Publication of CN113095327A publication Critical patent/CN113095327A/en
Application granted granted Critical
Publication of CN113095327B publication Critical patent/CN113095327B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/148Segmentation of character regions
    • G06V30/153Segmentation of character regions using recognition of characters or words
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/187Segmentation; Edge detection involving region growing; involving region merging; involving connected component labelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/62Analysis of geometric attributes of area, perimeter, diameter or volume
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/22Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Geometry (AREA)
  • Character Input (AREA)

Abstract

The invention relates to the technical field of character recognition, in particular to a method, a system and a storage medium for positioning an optical character recognition area, wherein the method comprises the following steps: acquiring an original image, converting the original image into a gray image, and performing binarization processing to obtain an original binary image; marking connected domains in the image; acquiring height and width parameters of a connected domain; performing self-adaptive threshold value calculation on a current pixel point in the gray level image to obtain a first binary image; then removing a noise area in the image to obtain a second binary image; respectively carrying out integral and haar characteristic diagram calculation on the first binary image to obtain a gradient diagram; obtaining a y-axis positioning interval according to the gradient frequency of the gradient map; determining a target area according to the y-axis positioning interval, intercepting the second binary image, counting the distance between connected domains in the image, and determining an x-axis positioning interval according to the distance between the connected domains; the invention can realize the quick positioning of the positioning area and improve the stability and the accuracy of character recognition.

Description

Method and system for positioning optical character recognition area and storage medium thereof
Technical Field
The invention relates to the technical field of character recognition, in particular to a method for positioning an optical character recognition area, a system for positioning the optical character recognition area and a readable storage medium storing the method.
Background
With the development of information technology, people often use Optical Character Recognition (OCR) in social life, which is a technology that electronic equipment is used to optically convert characters in a paper document into an image file of a black-and-white dot matrix, and Recognition software is used to convert the characters in the image into a text format for further editing and processing by word processing software; therefore, the OCR technology is one of the key technologies for implementing paperless computer automatic processing, and the main indexes for measuring the performance of the OCR system include: rejection rate, false recognition rate, recognition speed, etc.
In the OCR technology, one of the key steps is to locate the character area; in the prior art, when complex background interference exists in a paper material, the conventional OCR recognition method cannot meet the requirement of quickly positioning and segmenting an OCR character area, so that the stability of OCR recognition is poor and the accuracy of character recognition is low.
Disclosure of Invention
In order to overcome the above-mentioned drawbacks, the present invention provides a positioning method, a positioning system and a readable storage medium storing the positioning method, which can quickly position and segment an OCR character area.
The purpose of the invention is realized by the following technical scheme:
the invention relates to a method for positioning an optical character recognition area, which comprises the following steps:
acquiring an original image to be identified and width parameters thereof, carrying out gray level conversion on the original image to obtain a gray level image, and then carrying out binarization processing on the gray level image to obtain an original binary image;
marking a plurality of connected domains in the original binary image by a connected domain marking algorithm; acquiring the height and width attribute parameters of each connected domain; the attribute parameters of the height and the width of all connected domains are counted, and the attribute parameters of the width and the height with the highest frequency of occurrence are defined as the actual attribute parameters of the width, the height and the area;
performing threshold calculation on the gray values of the current pixel point and the neighborhood pixel block in the gray image by using a self-adaptive threshold binarization algorithm to obtain a first binary image;
removing a noise area in the first binary image according to the height and width attribute parameters of the connected domain to obtain a second binary image;
integrating the first binary image to obtain an integral image, determining a haar characteristic value according to actual width and height attribute parameters, and calculating the haar characteristic value of the integral image according to the haar characteristic value to obtain a gradient image; after binarization processing is carried out on the gradient map, traversing the gradient frequency of the gradient map along the direction from small to large of the y-axis coordinate, and determining a y-axis starting point and a y-axis end point by judging whether the gradient frequency is greater than a character number threshold value or not to obtain a y-axis positioning interval;
determining a target area according to the y-axis positioning interval and the width parameter of the original image, and intercepting the second binary image according to the target area to obtain a third binary image; counting the distance between adjacent connected domains in the third binary image, and calculating the average distance range; and determining an x-axis starting point and an x-axis end point according to the distance between the connected domains and the actual width attribute parameters of the connected domains to obtain an x-axis positioning interval.
In the present invention, traversing the gradient frequency of the y-axis coordinate from small to large and determining the y-axis starting point and the y-axis end point by judging whether the gradient frequency is greater than the character number threshold, and obtaining the y-axis positioning interval includes:
traversing the gradient frequency of the y-axis coordinate from small to large and judging whether the gradient frequency is greater than a character number threshold, and if so, determining the current y-direction position as a y-axis starting point; and continuously judging whether the gradient frequency is greater than the character number threshold value or not along the y-axis direction, and if the gradient frequency is less than the threshold value, determining the current y-direction position as a y-axis terminal point to obtain a y-axis positioning interval.
In the present invention, the determining an x-axis starting point and an x-axis ending point according to the distance between the connected domains and the actual width attribute parameters of the connected domains to obtain an x-axis positioning interval includes:
traversing the distance between the connected domains from small to large along the x-axis coordinate, determining two adjacent connected domains which meet the average distance range, taking the x-axis coordinate of the connected domain with the small x-axis coordinate as the x-axis starting point, continuously judging whether the distance between the subsequent connected domains and the actual wide attribute parameter are in the preset range along the x-axis direction, and if not, taking the x-axis coordinate of the current connected domain as the x-axis end point to obtain an x-axis positioning interval.
In the present invention, the removing the noise region in the first binary image according to the height and width attribute parameters of the connected domain to obtain the second binary image includes:
in the first binary image, marking a plurality of second connected regions by a connected domain marking algorithm of eight adjacent points, defining the second connected regions as connected domains, and acquiring the height and width attribute parameters of each connected domain;
and removing a noise region in the first binary image according to the high and wide attribute parameters of the connected domain to obtain a second binary image.
In the present invention, the removing the noise region in the first binary image according to the height and width attribute parameters of the connected domain to obtain the second binary image includes:
acquiring maximum area parameters and minimum area parameters in all connected domains, determining a maximum area threshold value through the maximum area parameters, determining a minimum area threshold value through the minimum area parameters, and calculating an area threshold value range;
and respectively judging whether the area of each connected domain is within the area threshold range, if not, defining the connected domain as a noise region, and removing the noise region from the first binary image.
In the present invention, said marking a plurality of connected regions in the original binary image by a connected component marking algorithm includes:
in the original binary image, a plurality of first connected regions are marked through a connected domain marking algorithm of four adjacent points, and the first connected regions are defined as connected domains.
In the present invention, before performing threshold calculation on the gray values of the current pixel point and the neighboring pixel block in the gray image by using an adaptive threshold binarization algorithm, the method further includes:
calculating the size ratio value between the actual width and height attribute parameters and the preset standard width and height attribute parameters, and zooming the gray-scale image according to the size ratio value.
In the present invention, the performing threshold calculation on the gray values of the current pixel point and the neighboring pixel block in the gray image by using the adaptive threshold binarization algorithm to obtain a first binary image comprises:
acquiring the gray value of a pixel point adjacent to the current pixel point in the gray image, and setting a gray threshold value according to the gray value of the adjacent pixel point; and then carrying out self-adaptive threshold value binarization calculation according to the set gray threshold value to obtain a first binary image.
Based on the same concept, the present invention also provides a positioning system of an optical character recognition area, comprising:
the system comprises an original image acquisition module, a width parameter acquisition module and a recognition module, wherein the original image acquisition module is used for acquiring an original image to be recognized and a width parameter thereof;
the gray level image conversion module is connected with the original image acquisition module and is used for carrying out gray level conversion on the original image to obtain a gray level image;
the original binary image generation module is connected with the gray level image conversion module and is used for carrying out binarization processing on the gray level image to obtain an original binary image;
the connected domain generating module is connected with the original binary image generating module and used for marking a plurality of connected domains in the original binary image through a connected domain marking algorithm and acquiring the height and width attribute parameters of each connected domain;
the actual parameter acquisition module is connected with the connected domain generation module and is used for defining the attribute parameters with the highest frequency of occurrence, such as the width and the height, as the actual attribute parameters of the width, the height and the area by counting the attribute parameters of the height and the width of all the connected domains;
the first binary image generation module is connected with the gray level image conversion module and is used for performing threshold calculation on gray levels of a current pixel point and a neighborhood pixel block in the gray level image by using a self-adaptive threshold binary algorithm to obtain a first binary image;
the second binary image generation module is connected with the connected domain generation module and the first binary image generation module and is used for removing a noise area in the first binary image according to the height and width attribute parameters of the connected domain to obtain a second binary image;
the y-axis interval positioning module is connected with the first binary image generating module and the actual parameter acquiring module and used for integrating the first binary image to obtain an integral map, determining a haar characteristic value according to actual width and height attribute parameters, and calculating the integral map according to the haar characteristic value to obtain a gradient map; after binarization processing is carried out on the gradient map, traversing the gradient frequency of the gradient map along the direction from small to large of the y-axis coordinate, and determining a y-axis starting point and a y-axis end point by judging whether the gradient frequency is greater than a character number threshold value or not to obtain a y-axis positioning interval;
the x-axis interval positioning module is respectively connected with the y-axis interval positioning module, the second binary image generating module and the original image acquiring module and is used for determining a target area according to a y-axis positioning interval and width parameters of the original image, and intercepting the second binary image according to the target area to obtain a third binary image; counting the distance between adjacent connected domains in the third binary image, and calculating the average distance range; and determining an x-axis starting point and an x-axis end point according to the distance between the connected domains and the actual width attribute parameters of the connected domains to obtain an x-axis positioning interval.
Based on the same concept, the present invention also provides a computer-readable program storage medium storing computer program instructions which, when executed by a computer, cause the computer to perform the method as described above.
According to the method, the position of the y-axis direction in the positioning area is quickly obtained by carrying out normalization processing on the image and according to the gradient frequency distribution in the image; the position in the x-axis direction in the positioning area is quickly obtained through the character width and character detection characteristics in the x-axis direction; therefore, the method and the device can realize quick positioning of the positioning area, further greatly improve the stability of character recognition and improve the accuracy of character recognition.
Drawings
For the purpose of easy explanation, the present invention will be described in detail with reference to the following preferred embodiments and the accompanying drawings.
FIG. 1 is a flowchart illustrating an embodiment of a method for locating an OCR region according to the present invention;
FIG. 2 is a schematic diagram illustrating a haar feature calculation in the method for locating an optical character recognition area according to the present invention;
FIG. 3 is a schematic diagram illustrating the calculation of haar features in the method for locating an optical character recognition area according to the present invention;
FIG. 4 is a schematic diagram illustrating the principle of integral calculation for each point in an integral graph in the method for locating an optical character recognition area according to the present invention;
FIG. 5 is a flowchart illustrating a method for locating an OCR region according to another embodiment of the present invention;
FIG. 6 is a schematic logical structure diagram of an embodiment of a system for locating an OCR region according to the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
In the description of the present invention, it is to be understood that the terms "center", "longitudinal", "lateral", "length", "width", "thickness", "upper", "lower", "front", "rear", "left", "right", "vertical", "horizontal", "top", "bottom", "inner", "outer", "clockwise", "counterclockwise", and the like, indicate orientations and positional relationships based on those shown in the drawings, and are used only for convenience of description and simplicity of description, and do not indicate or imply that the device or element being referred to must have a particular orientation, be constructed and operated in a particular orientation, and thus, should not be considered as limiting the present invention. Furthermore, the terms "first", "second" and "first" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, features defined as "first", "second", may explicitly or implicitly include one or more of the described features. In the description of the present invention, "a plurality" means two or more unless specifically defined otherwise.
In the description of the present invention, it should be noted that the terms "mounted," "connected," and "connected" are to be construed broadly and may be, for example, fixedly connected, detachably connected, or integrally connected unless otherwise explicitly stated or limited. Either mechanically or electrically. Either directly or indirectly through intervening media, either internally or in any other relationship. The specific meanings of the above terms in the present invention can be understood by those skilled in the art according to specific situations.
An embodiment of a method for locating an optical character recognition area according to the present invention is described in detail below with reference to fig. 1, which includes:
s101, acquiring an original image to be identified
Obtaining an image containing characters by image scanning or importing the image from a database, positioning and roughly cutting an original image containing character areas and obtaining a width parameter X of the image by an empirical parameter configuration modewidth
S102, obtaining a gray level image through gray level conversion
The original image is subjected to gray scale conversion to obtain a gray scale image I, which is specifically calculated by a formula I (x, y) +1/3 × I _ R (x, y) +1/3 × I _ G (x, y) +1/3 × I _ B (x, y), wherein R, G, B is red, green and blue colors in the figure respectively.
S103, carrying out binarization processing to obtain an original binary image
Carrying out binarization processing on the gray level image I by utilizing an Otsu (otsu) algorithm to obtain an original binary image Bo; wherein, otsu algorithm is also called as: the maximum inter-class variance method.
S104, marking a plurality of connected domains through a connected domain marking algorithm
Marking a plurality of connected domains in the original binary image Bo through a connected domain marking algorithm; the connected domain refers to an image region which is formed by foreground pixel points with the same pixel value and adjacent positions in an image, and each connected domain is marked as a blob. Connected component analysis refers to finding and labeling each connected component in an image. In an image, the smallest unit is a pixel point, 8 adjacent pixels are arranged around each pixel point, and 2 common adjacent relations are provided: 4 contiguous with 8 contiguous. 4 are adjacent to a total of 4 points, namely, the upper, lower, left and right directions of the pixel points, 8 are adjacent to a total of 8 points, and the 4 points adjacent to the 4 points also comprise 4 points of upper left, lower left, upper right and lower right; in this embodiment, it marks the connected domain in an 8-way adjacency. And (3) performing connected domain analysis on the pixel points with the gray value Bo (x, y) of 255 in the original binary image Bo, and calculating to obtain the attribute parameters of the height and the width of each connected domain.
S105, acquiring actual width and height attribute parameters
The method comprises the steps that the attribute parameters of the height and the width of all connected domains are counted, and the attribute parameters of the width and the height with the highest occurrence frequency are defined as the actual attribute parameters of the width, the height and the area; since it is known that there are many characters in the original image to be recognized, and the aspect ratio of the characters is generally consistent, the actual width and height of the image of most of the character numbers can be obtained through statistics in this step.
S106, obtaining a first binary image by using a self-adaptive threshold value binarization algorithm
Performing threshold calculation on the gray values of the current pixel point and the neighborhood pixel block in the gray image I by using a self-adaptive threshold binarization algorithm to obtain a first binary image B1; the neighborhood pixels are square, and the size of the neighborhood pixels is set to be one fourth of the actual height of the connected domain; specifically, the gray value of the current pixel point is compared with the gray value of the neighborhood pixel block, if the gray value of the current pixel point is smaller than the gray value of the neighborhood pixel block, the current pixel point is set to 0, otherwise, the pixel point is set to 255, and thus the first binary image B1 is generated.
S107, removing the noise area to obtain a second binary image
Removing a noise region in the first binary image B1 according to the height and width attribute parameters of the connected domain to obtain a second binary image B2;
s108, calculating and obtaining a y-axis positioning interval by combining a haar characteristic diagram
Integrating the first binary image B1 to obtain an integral map Sat, wherein the conversion formula of the integral map Sat is as follows:
Figure BDA0002978018000000081
determining a haar-like characteristic value according to the actual width and height attribute parameters, wherein the haar characteristic is a digital image characteristic used for object identification; then carrying out haar characteristic diagram calculation on the integral diagram according to the haar characteristic value to obtain a gradient diagram T; wherein, haar edge characteristics are as shown in fig. 2 to 4, and the characteristics are described as follows: the width of the black part is one third of the actual width of the connected domain, the height is the actual width of the connected domain, the width and height of the white part are the standard width and height, and the specific formula is as follows:
HarrA-B=Sum(A)-Sum(B)
=[SAT6+SAT2-SAT2-SAT5]-[SAT5+SAT1-SAT2-SAT4]
wherein, the integral calculation of each point in the integral graph is specifically as follows:
SAT1=Sum(Ra)
SAT2=Sum(Ra)+Sum(Rb)
SAT3=Sum(Ra)+Sum(Rc)
Sum(Rd)=SAT1+SAT4-SAT2-SAT3
after the gradient map T is subjected to binarization processing, the gradient frequency F of the gradient map T is subjected to binarization processing along the direction from small to large of the y-axis coordinateocrTraversing and determining a y-axis starting point and a y-axis end point by judging whether the number of characters is greater than a character number threshold value Tresh to obtain a y-axis positioning interval;
s109, obtaining an x-axis positioning interval according to the distance of the connected domains and the actual wide attribute parameters
Positioning an interval according to the y axis and the width parameter X of the original imagewidthDetermining a target area, and intercepting the second binary image B2 according to the target area to obtain a third binary image B3; counting the distances between adjacent connected domains blob in the third binary image B3, and calculating the average distance range; and determining an x-axis starting point and an x-axis terminal point according to the distance between connected domain blobs and the actual width attribute parameters of the connected domain blobs to obtain an x-axis positioning interval.
In another embodiment, a method for positioning an optical character recognition area according to the present invention is described in detail below with reference to fig. 5, which includes:
s201, obtaining an original image to be identified
Obtaining an image containing characters by image scanning or importing the image from a database, positioning and roughly cutting an original image containing character areas and obtaining a width parameter X of the image by an empirical parameter configuration modewidth
S202, obtaining a gray level image through gray level conversion
The original image is subjected to gray scale conversion to obtain a gray scale image I, which is specifically calculated by a formula I (x, y) +1/3 × I _ R (x, y) +1/3 × I _ G (x, y) +1/3 × I _ B (x, y), wherein R, G, B is red, green and blue colors in the figure respectively.
S203, carrying out binarization processing to obtain an original binary image
Carrying out binarization processing on the gray level image I by utilizing an Otsu (otsu) algorithm to obtain an original binary image Bo; wherein, otsu algorithm is also called as: the maximum inter-class variance method.
S204, marking a plurality of first communication areas through a communication area marking algorithm
In the original binary image Bo, performing connected domain analysis and marking on pixel points Bo (x, y) ═ 255 by using a connected domain marking algorithm of four adjacent points to obtain a plurality of first connected regions, and defining the first connected regions as connected domains; the connected domain refers to an image region which is formed by foreground pixel points with the same pixel value and adjacent positions in an image, and each connected domain is marked as a blob. Connected component analysis refers to finding and labeling each connected component in an image. In an image, the smallest unit is a pixel point, 8 adjacent pixels are arranged around each pixel point, and 2 common adjacent relations are provided: 4 contiguous with 8 contiguous. 4 are adjacent to a total of 4 points, namely, the upper, lower, left and right directions of the pixel points, 8 are adjacent to a total of 8 points, and the 4 points adjacent to the 4 points also comprise 4 points of upper left, lower left, upper right and lower right; in this embodiment, it marks the connected domain in an 8-way adjacency. And (3) performing connected domain analysis on the pixel points with the gray value Bo (x, y) of 255 in the original binary image Bo, and calculating to obtain the attribute parameters of the height and the width of each connected domain.
S205, obtaining actual width and height attribute parameters
Calculating the aspect ratios of all connected domains according to the height and width attribute parameters of all connected domains, putting the calculated aspect ratios into corresponding aspect ratio intervals according to the respective aspect ratios, counting the aspect ratio intervals in which the most connected domains fall, and defining the width and height attribute parameters corresponding to the aspect ratio intervals as actual width and height attribute parameters; since it is known that there are many characters in the original image to be recognized, and the aspect ratio of the characters is generally consistent, the actual width and height of the image of most of the character numbers can be obtained through statistics in this step. In this embodiment, a first communication area is obtained by means of communication area marking of four adjacent points, and actual width and height are obtained according to statistics of the first communication area; it can quickly obtain the actual width and height of the collocation.
S206, zooming the gray level image
Calculating the size ratio value between the actual width and height attribute parameters and the preset standard width and height attribute parameters, and zooming the gray-scale image according to the size ratio value. Specifically, it uses the formula according to the actual width and height of the connected domain in the image and the width and height of the target character: the proportion is actual height, width/target height and width, and the gray level image is subjected to size scaling according to the calculation result of the ratio height and width to obtain a gray level image Id, so that the size of the processed character can be ensured to be kept unchanged; facilitating subsequent processing. Wherein the width and height of the target character are set empirically.
S207, obtaining a first binary image by using a self-adaptive threshold value binarization algorithm
Acquiring the gray value of a pixel point adjacent to the current pixel point in the gray image Id, and setting a gray threshold value according to the gray value of the adjacent pixel point; then, performing adaptive threshold binarization calculation according to the set gray threshold to obtain a first binary image B1; the neighborhood pixels are square, and the size of the neighborhood pixels is set to be one fourth of the actual height of the connected domain; specifically, the gray value of the current pixel point is compared with the gray value of the neighborhood pixel block, if the gray value of the current pixel point is smaller than the gray value of the neighborhood pixel block, the current pixel point is set to 0, otherwise, the pixel point is set to 255, and thus the first binary image B1 is generated.
S208, marking a plurality of second connected regions through a connected region marking algorithm
In the first binary image B1, performing connected domain analysis on the pixel points B1(x, y) ═ 255 by using an eight-adjacency-point connected domain labeling algorithm to label a plurality of second connected regions, defining the second connected regions as connected domains, and acquiring attribute parameters of height, width and area of each connected domain.
S209, removing the noise area to obtain a second binary image
Acquiring maximum area parameters and minimum area parameters in all connected domains, determining a maximum area threshold value through the maximum area parameters, determining a minimum area threshold value through the minimum area parameters, and calculating an area threshold value range;
respectively judging whether the area of each connected domain is within the area threshold range, if not, defining the connected domain as a noise region, and removing the noise region from the first binary image; the method comprises the following steps: counting the area distribution of a class of connected domains which meet the requirements of width and height of characters to obtain the maximum area Smax and the minimum area Smin; clearing the connected region smaller than Tmin in the first binary image B1 by taking the size of one fourth of Smin as a threshold Tmin; and (3) clearing a connected region larger than Tmax in the first binary image B1 by taking the size of 2 times of Smax as a threshold Tmax to obtain a second binary image B2.
S210, calculating and obtaining a y-axis positioning interval by combining a haar characteristic diagram
Integrating the first binary image B1 to obtain an integral map Sat, wherein the conversion formula of the integral map Sat is as follows:
Figure BDA0002978018000000111
determining a haar-like characteristic value according to the actual width and height attribute parameters, wherein the haar characteristic is a digital image characteristic used for object identification; then carrying out haar characteristic diagram calculation on the integral diagram according to the haar characteristic value to obtain a gradient diagram T; wherein, haar edge characteristics are as shown in fig. 2 to 4, and the characteristics are described as follows: the width of the black part is one third of the actual width of the connected domain, the height is the actual width of the connected domain, the width and height of the white part are the standard width and height, and the specific formula is as follows:
HarrA-B=Sum(A)-Sum(B)
=[SAT6+SAT2-SAT2-SAT5]-[SAT5+SAT1-SAT2-SAT4]
wherein, the integral calculation of each point in the integral graph is specifically as follows:
SAT1=Sum(Ra)
SAT2=Sum(Ra)+Sum(Rb)
SAT3=Sum(Ra)+Sum(Rc)
Sum(Rd)=SAT1+SAT4-SAT2-SAT3
carrying out binarization processing on the gradient map T, wherein the threshold value is standard width multiplied by standard width/4; gradient frequency F along y-axis coordinate from small to largeocrTraversing and judging whether the width is larger than a character number threshold value Tresh or not, wherein the width W of the connected domain is knownocrAnd width W of the imageTObtaining the threshold value Tresh ═ α × WT/WocrWherein, alpha is (1/2, 2/3); if the current y-direction position is larger than the threshold value Tresh, determining the current y-direction position as a y-axis starting point; and continuously judging the gradient frequency F along the y-axis directionocrWhether the number of the characters is larger than a character number threshold value or not, if the number of the characters is smaller than a threshold value Tresh, determining the current y-direction position as a y-axis terminal point to obtain a y-axis positioning interval; it may specifically be: if the gradient frequency F of the current lineocrIf the current Y-direction position is larger than the threshold value Tresh, the current Y-direction position is determined as the Y-axis starting point YstartAt this time, OCR flagocrSelf-adding 1; if the gradient frequency F of the current lineocrIf the value is less than the threshold value Tresh, the Y-axis starting point Y is determinedstartSet to zero at this time OCR flagocrSetting zero; if the gradient frequency F of the current lineocrLess than threshold Tresh and meets OCR flagocrAdding 1, determining the current Y-direction position as the Y-axis terminal point YendBy the coordinate Ystart、YendAnd determining a positioning interval of the Y-axis direction of the OCR.
S211, obtaining an x-axis positioning interval according to the distance of the connected domains and the actual wide attribute parameters
Positioning an interval according to the y axis and the width parameter X of the original imagewidthDetermining a target area, wherein the specific target area is (0, Y)start,Xwidth,Yend-Ystart) The second binary image B2 is processed according to the target areaIntercepting to obtain a third binary image B3; counting the distances between adjacent connected domains blob in the third binary image B3, and calculating the average distance range; traversing the distance between the connected domains from small to large along the X-axis coordinate, determining two adjacent connected domains which meet the average distance range, taking the X-axis coordinate of the connected domain with the small X-axis coordinate as the X-axis starting point, and continuously judging the distance between the subsequent connected domains and the actual wide attribute parameter X along the X-axis directionwidthAnd if not, taking the x-axis coordinate of the current connected domain as an x-axis terminal point to obtain an x-axis positioning interval. Specifically, calculating the distribution interval of each connected domain blob in the x-axis direction, and counting the distance of each adjacent blob block according to the sorting from small to large according to the x coordinate of the center of the blob block, wherein the distance is the character interval; and the normal OCR character interval should have small difference, based on the property, the mean gap of blob interval distribution is calculated firstavg(ii) a Then, evaluating the minimum starting point in the x-axis direction with uniform interval, and taking the blob with the minimum central coordinate x value as the starting block blob of the line of OCR0,blob0The starting position X of the X-axis direction of the OCR area is determinedstart(ii) a 5) From XstartStarting with the actual wide attribute parameter of the connected field, the character pitch gapavgTraversing from left to right until the size of the character does not meet the requirement, and obtaining the OCR end position X in the X-axis direction at the momentend(ii) a Location area (X)start,Ystart,Xend-Xstart,Yend-Ystart) I.e. a cut-out box of the OCR area that needs to be located.
The present invention includes a computer readable storage medium having stored thereon a program product capable of implementing the above-described method of the present specification. In some possible embodiments, aspects of the invention may also be implemented in the form of a program product comprising program code means for causing a terminal device to carry out the steps according to various exemplary embodiments of the invention described in the above section "exemplary methods" of the present description, when said program product is run on the terminal device.
The program product may employ any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. A readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium include: an electrical connection having one or more wires, a portable disk, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
A computer readable signal medium may include a propagated data signal with readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A readable signal medium may also be any readable medium that is not a readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on the above readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
An embodiment of a positioning system for an optical character recognition area according to the present invention is described in detail below, with reference to fig. 6, which includes:
the system comprises an original image acquisition module 101, wherein the original image acquisition module 101 is used for acquiring an original image to be identified and width parameters thereof; specifically, the method acquires an image containing characters through image scanning or importing the image from a database, and positions and cuts rough original images containing character areas and acquires width parameters of the images through an empirical parameter configuration mode.
The grayscale image conversion module 102, the grayscale image conversion module 102 is connected to the original image acquisition module 101, and is configured to perform grayscale conversion on the original image to obtain a grayscale image; specifically, formula I (x, y) ═ 1/3 × I _ R (x, y) +1/3 × I _ G (x, y) +1/3 × I _ B (x, y), where R, G, B represents three colors of red, green, and blue in the figure.
An original binary image generation module 103, where the original binary image generation module 103 is connected to the grayscale image conversion module 102, and is configured to perform binarization processing on the grayscale image to obtain an original binary image; in the present embodiment, it performs binarization processing on the grayscale image by using an otsu (otsu) algorithm.
A connected domain generating module 104, where the connected domain generating module 104 is connected to the original binary image generating module 103, and is configured to mark a plurality of connected domains in the original binary image through a connected domain marking algorithm, and obtain attribute parameters of the height and width of each connected domain; the connected domain refers to an image region which is formed by foreground pixel points with the same pixel value and adjacent positions in an image, and each connected domain is marked as a blob. Connected component analysis refers to finding and labeling each connected component in an image. In an image, the smallest unit is a pixel point, 8 adjacent pixels are arranged around each pixel point, and 2 common adjacent relations are provided: 4 contiguous with 8 contiguous. 4 are adjacent to a total of 4 points, namely, the upper, lower, left and right directions of the pixel points, 8 are adjacent to a total of 8 points, and the 4 points adjacent to the 4 points also comprise 4 points of upper left, lower left, upper right and lower right; in this embodiment, it marks the connected domain in an 8-way adjacency. And (3) performing connected domain analysis on the pixel points with the gray value Bo (x, y) of 255 in the original binary image Bo, and calculating to obtain the attribute parameters of the height and the width of each connected domain.
An actual parameter obtaining module 105, where the actual parameter obtaining module 105 is connected to the connected domain generating module 104, and is configured to define the attribute parameters with the highest occurrence frequency, such as the width and height, as actual width, height and area attribute parameters by performing statistics on the attribute parameters of the height and width of all connected domains; specifically, a plurality of aspect ratio sections are preset, the aspect ratios of all connected domains are calculated according to the height attribute parameters and the width attribute parameters of all connected domains, the connected domains are placed into the corresponding aspect ratio sections according to the respective aspect ratios, the aspect ratio section where the most connected domains fall is counted, and the width attribute parameters and the height attribute parameters corresponding to the aspect ratio sections are defined as the actual width attribute parameters and the actual height attribute parameters.
A first binary image generation module 106, where the first binary image generation module 106 is connected to the grayscale image conversion module 102, and is configured to perform threshold calculation on the grayscale values of the current pixel point and the neighborhood pixel block in the grayscale image by using an adaptive threshold binarization algorithm to obtain a first binary image; specifically, the gray values of the current pixel point and the neighborhood pixel block are compared, if the gray values of the current pixel point and the neighborhood pixel block are smaller than the gray value of the neighborhood pixel block, the current pixel point is set to be 0, otherwise, the pixel point is set to be 255, and therefore the first binary image is generated.
A second binary image generating module 107, where the second binary image generating module 107 is connected to the connected domain generating module 104 and the first binary image generating module 106, and is configured to remove a noise region in the first binary image according to the high and wide attribute parameters of the connected domain to obtain a second binary image; the method comprises the following steps: acquiring maximum area parameters and minimum area parameters in all connected domains, determining a maximum area threshold value through the maximum area parameters, determining a minimum area threshold value through the minimum area parameters, and calculating an area threshold value range; and respectively judging whether the area of each connected domain is within the area threshold range, if not, defining the connected domain as a noise region, and removing the noise region from the first binary image.
A y-axis interval positioning module 108, where the y-axis interval positioning module 108 is connected to the first binary image generating module 106 and the actual parameter obtaining module 105, and is configured to integrate the first binary image to obtain an integral map, determine a haar feature value according to actual width and height attribute parameters, and calculate the haar feature value according to the haar feature value to obtain a gradient map; after binarization processing is carried out on the gradient map, traversing the gradient frequency of the gradient map along the direction from small to large of the y-axis coordinate, and determining a y-axis starting point and a y-axis end point by judging whether the gradient frequency is greater than a character number threshold value or not to obtain a y-axis positioning interval; specifically, after the gradient map is subjected to binarization processing, traversing the gradient frequency of the gradient map along the y-axis coordinate from small to large and judging whether the gradient frequency is greater than a character number threshold value, and if so, determining the current y-direction position as a y-axis starting point; and continuously judging whether the gradient frequency is greater than the character number threshold value or not along the y-axis direction, and if the gradient frequency is less than the threshold value, determining the current y-direction position as a y-axis terminal point to obtain a y-axis positioning interval.
The x-axis interval positioning module 109 is connected with the y-axis interval positioning module 108, the second binary image generating module 107 and the original image acquiring module 101 respectively, and is used for determining a target area according to the y-axis positioning interval and the width parameter of the original image, and intercepting the second binary image according to the target area to obtain a third binary image; counting the distance between adjacent connected domains in the third binary image, and calculating the average distance range; determining an x-axis starting point and an x-axis terminal point according to the distance between the connected domains and the actual width attribute parameters of the connected domains to obtain an x-axis positioning interval; specifically, after the average distance range is calculated, traversing the distance between the connected domains along the x-axis coordinate from small to large, determining two adjacent connected domains which meet the average distance range, taking the x-axis coordinate of the connected domain with the small x-axis coordinate as the x-axis starting point, continuously judging whether the distance between the subsequent connected domains and the actual wide attribute parameter are within a preset range along the x-axis direction, and if not, taking the x-axis coordinate of the current connected domain as the x-axis end point to obtain an x-axis positioning interval.
In the description of the present specification, reference to the description of the terms "one embodiment", "some embodiments", "an illustrative embodiment", "an example", "a specific example", or "some examples", etc., means that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the present invention. In this specification, schematic representations of the above terms do not necessarily refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents and improvements made within the spirit and principle of the present invention are intended to be included within the scope of the present invention.

Claims (10)

1. A method for locating an optical character recognition area, comprising:
acquiring an original image to be identified and width parameters thereof, carrying out gray level conversion on the original image to obtain a gray level image, and then carrying out binarization processing on the gray level image to obtain an original binary image;
marking a plurality of connected domains in the original binary image by a connected domain marking algorithm; acquiring the height and width attribute parameters of each connected domain; the attribute parameters of the height and the width of all connected domains are counted, and the attribute parameters of the width and the height with the highest frequency of occurrence are defined as the actual attribute parameters of the width, the height and the area;
performing threshold calculation on the gray values of the current pixel point and the neighborhood pixel block in the gray image by using a self-adaptive threshold binarization algorithm to obtain a first binary image;
removing a noise area in the first binary image according to the height and width attribute parameters of the connected domain to obtain a second binary image;
integrating the first binary image to obtain an integral image, determining a haar characteristic value according to actual width and height attribute parameters, and calculating the haar characteristic value of the integral image according to the haar characteristic value to obtain a gradient image; after binarization processing is carried out on the gradient map, traversing the gradient frequency of the gradient map along the direction from small to large of the y-axis coordinate, and determining a y-axis starting point and a y-axis end point by judging whether the gradient frequency is greater than a character number threshold value or not to obtain a y-axis positioning interval;
determining a target area according to the y-axis positioning interval and the width parameter of the original image, and intercepting the second binary image according to the target area to obtain a third binary image; counting the distance between adjacent connected domains in the third binary image, and calculating the average distance range; and determining an x-axis starting point and an x-axis end point according to the distance between the connected domains and the actual width attribute parameters of the connected domains to obtain an x-axis positioning interval.
2. The method for locating an optical character recognition area according to claim 1, wherein traversing the gradient frequency of the y-axis coordinate from small to large and determining a y-axis starting point and a y-axis ending point by determining whether the gradient frequency is greater than a character number threshold value or not to obtain a y-axis locating interval comprises:
traversing the gradient frequency of the y-axis coordinate from small to large and judging whether the gradient frequency is greater than a character number threshold, and if so, determining the current y-direction position as a y-axis starting point; and continuously judging whether the gradient frequency is greater than the character number threshold value or not along the y-axis direction, and if the gradient frequency is less than the threshold value, determining the current y-direction position as a y-axis terminal point to obtain a y-axis positioning interval.
3. The method of claim 2, wherein determining an x-axis start point and an x-axis end point according to the distance between connected components and the actual wide property parameters of the connected components to obtain an x-axis positioning interval comprises:
traversing the distance between the connected domains from small to large along the x-axis coordinate, determining two adjacent connected domains which meet the average distance range, taking the x-axis coordinate of the connected domain with the small x-axis coordinate as the x-axis starting point, continuously judging whether the distance between the subsequent connected domains and the actual wide attribute parameter are in the preset range along the x-axis direction, and if not, taking the x-axis coordinate of the current connected domain as the x-axis end point to obtain an x-axis positioning interval.
4. The method of claim 3, wherein the removing the noise region in the first binary image according to the property parameters of the height and width of the connected component to obtain the second binary image comprises:
in the first binary image, marking a plurality of second connected regions by a connected domain marking algorithm of eight adjacent points, defining the second connected regions as connected domains, and acquiring the height and width attribute parameters of each connected domain;
and removing a noise region in the first binary image according to the high and wide attribute parameters of the connected domain to obtain a second binary image.
5. The method of claim 4, wherein the removing the noise region in the first binary image according to the property parameters of the height and width of the connected component to obtain the second binary image comprises:
acquiring maximum area parameters and minimum area parameters in all connected domains, determining a maximum area threshold value through the maximum area parameters, determining a minimum area threshold value through the minimum area parameters, and calculating an area threshold value range;
and respectively judging whether the area of each connected domain is within the area threshold range, if not, defining the connected domain as a noise region, and removing the noise region from the first binary image.
6. The method for locating an optical character recognition area according to claim 5, wherein the marking a plurality of connected components in the original binary image by a connected component marking algorithm comprises:
in the original binary image, a plurality of first connected regions are marked through a connected domain marking algorithm of four adjacent points, and the first connected regions are defined as connected domains.
7. The method as claimed in claim 6, wherein the step of performing threshold calculation on the gray values of the current pixel point and the neighborhood pixel block in the gray image by using an adaptive threshold binarization algorithm further comprises:
calculating the size ratio value between the actual width and height attribute parameters and the preset standard width and height attribute parameters, and zooming the gray-scale image according to the size ratio value.
8. The method for locating an optical character recognition area according to claim 7, wherein the performing a threshold calculation on the gray values of the current pixel point and the neighboring pixel block in the gray image by using an adaptive threshold binarization algorithm to obtain a first binary image comprises:
acquiring the gray value of a pixel point adjacent to the current pixel point in the gray image, and setting a gray threshold value according to the gray value of the adjacent pixel point; and then carrying out self-adaptive threshold value binarization calculation according to the set gray threshold value to obtain a first binary image.
9. A system for locating an optical character recognition area, comprising:
the system comprises an original image acquisition module, a width parameter acquisition module and a recognition module, wherein the original image acquisition module is used for acquiring an original image to be recognized and a width parameter thereof;
the gray level image conversion module is connected with the original image acquisition module and is used for carrying out gray level conversion on the original image to obtain a gray level image;
the original binary image generation module is connected with the gray level image conversion module and is used for carrying out binarization processing on the gray level image to obtain an original binary image;
the connected domain generating module is connected with the original binary image generating module and used for marking a plurality of connected domains in the original binary image through a connected domain marking algorithm and acquiring the height and width attribute parameters of each connected domain;
the actual parameter acquisition module is connected with the connected domain generation module and is used for defining the attribute parameters with the highest frequency of occurrence, such as the width and the height, as the actual attribute parameters of the width, the height and the area by counting the attribute parameters of the height and the width of all the connected domains;
the first binary image generation module is connected with the gray level image conversion module and is used for performing threshold calculation on gray levels of a current pixel point and a neighborhood pixel block in the gray level image by using a self-adaptive threshold binary algorithm to obtain a first binary image;
the second binary image generation module is connected with the connected domain generation module and the first binary image generation module and is used for removing a noise area in the first binary image according to the height and width attribute parameters of the connected domain to obtain a second binary image;
the y-axis interval positioning module is connected with the first binary image generating module and the actual parameter acquiring module and used for integrating the first binary image to obtain an integral map, determining a haar characteristic value according to actual width and height attribute parameters, and calculating the integral map according to the haar characteristic value to obtain a gradient map; after binarization processing is carried out on the gradient map, traversing the gradient frequency of the gradient map along the direction from small to large of the y-axis coordinate, and determining a y-axis starting point and a y-axis end point by judging whether the gradient frequency is greater than a character number threshold value or not to obtain a y-axis positioning interval;
the x-axis interval positioning module is respectively connected with the y-axis interval positioning module, the second binary image generating module and the original image acquiring module and is used for determining a target area according to a y-axis positioning interval and width parameters of the original image, and intercepting the second binary image according to the target area to obtain a third binary image; counting the distance between adjacent connected domains in the third binary image, and calculating the average distance range; and determining an x-axis starting point and an x-axis end point according to the distance between the connected domains and the actual width attribute parameters of the connected domains to obtain an x-axis positioning interval.
10. A computer-readable program storage medium storing computer program instructions which, when executed by a computer, cause the computer to perform the method of any one of claims 1 to 8.
CN202110280359.9A 2021-03-16 2021-03-16 Method and system for positioning optical character recognition area and storage medium thereof Active CN113095327B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110280359.9A CN113095327B (en) 2021-03-16 2021-03-16 Method and system for positioning optical character recognition area and storage medium thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110280359.9A CN113095327B (en) 2021-03-16 2021-03-16 Method and system for positioning optical character recognition area and storage medium thereof

Publications (2)

Publication Number Publication Date
CN113095327A true CN113095327A (en) 2021-07-09
CN113095327B CN113095327B (en) 2022-10-14

Family

ID=76668148

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110280359.9A Active CN113095327B (en) 2021-03-16 2021-03-16 Method and system for positioning optical character recognition area and storage medium thereof

Country Status (1)

Country Link
CN (1) CN113095327B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114387292A (en) * 2022-03-25 2022-04-22 北京市农林科学院信息技术研究中心 Image edge pixel point optimization method, device and equipment
CN115497109A (en) * 2022-11-17 2022-12-20 山东思玛特教育科技有限公司 Character and image preprocessing method based on intelligent translation
CN115546232A (en) * 2022-10-12 2022-12-30 什维新智医疗科技(上海)有限公司 Liver ultrasonic image working area extraction method and system and electronic equipment
CN117253233A (en) * 2023-09-05 2023-12-19 广东奥普特科技股份有限公司 Character erasing method, device and equipment
CN117831037A (en) * 2024-01-04 2024-04-05 北京和气聚力教育科技有限公司 Method and device for determining answer condition of objective questions in answer sheet
CN117877037A (en) * 2024-03-11 2024-04-12 杭州汇萃智能科技有限公司 OCR single character positioning and identifying method, system and computer storage medium

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001283153A (en) * 2000-01-27 2001-10-12 Fujitsu Ltd Device and method for extracting character string pattern from color picture
JP2007094806A (en) * 2005-09-29 2007-04-12 Victor Co Of Japan Ltd Image display method, image display apparatus, and image display program
US20070292029A1 (en) * 2006-06-16 2007-12-20 Shen-Zheng Wang Cascade plate recognition system
US20110013847A1 (en) * 2009-07-18 2011-01-20 Abbyy Software Ltd Identifying picture areas based on gradient image analysis
CN102902957A (en) * 2012-09-05 2013-01-30 佳都新太科技股份有限公司 Video-stream-based automatic license plate recognition method
CN103020618A (en) * 2011-12-19 2013-04-03 北京捷成世纪科技股份有限公司 Detection method and detection system for video image text
CN103870803A (en) * 2013-10-21 2014-06-18 北京邮电大学 Vehicle license plate recognition method and system based on coarse positioning and fine positioning fusion
CN104462380A (en) * 2014-12-11 2015-03-25 北京中细软移动互联科技有限公司 Trademark retrieval method
CN105373794A (en) * 2015-12-14 2016-03-02 河北工业大学 Vehicle license plate recognition method
WO2018028306A1 (en) * 2016-08-11 2018-02-15 杭州海康威视数字技术股份有限公司 Method and device for recognizing license plate number
CN109740603A (en) * 2019-01-21 2019-05-10 闽江学院 Based on the vehicle character identifying method under CNN convolutional neural networks
CN110443184A (en) * 2019-07-31 2019-11-12 上海海事大学 ID card information extracting method, device and computer storage medium
CN111598104A (en) * 2020-06-30 2020-08-28 成都鹏业软件股份有限公司 License plate character recognition method and system

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001283153A (en) * 2000-01-27 2001-10-12 Fujitsu Ltd Device and method for extracting character string pattern from color picture
JP2007094806A (en) * 2005-09-29 2007-04-12 Victor Co Of Japan Ltd Image display method, image display apparatus, and image display program
US20070292029A1 (en) * 2006-06-16 2007-12-20 Shen-Zheng Wang Cascade plate recognition system
US20110013847A1 (en) * 2009-07-18 2011-01-20 Abbyy Software Ltd Identifying picture areas based on gradient image analysis
CN103020618A (en) * 2011-12-19 2013-04-03 北京捷成世纪科技股份有限公司 Detection method and detection system for video image text
CN102902957A (en) * 2012-09-05 2013-01-30 佳都新太科技股份有限公司 Video-stream-based automatic license plate recognition method
CN103870803A (en) * 2013-10-21 2014-06-18 北京邮电大学 Vehicle license plate recognition method and system based on coarse positioning and fine positioning fusion
CN104462380A (en) * 2014-12-11 2015-03-25 北京中细软移动互联科技有限公司 Trademark retrieval method
CN105373794A (en) * 2015-12-14 2016-03-02 河北工业大学 Vehicle license plate recognition method
WO2018028306A1 (en) * 2016-08-11 2018-02-15 杭州海康威视数字技术股份有限公司 Method and device for recognizing license plate number
CN109740603A (en) * 2019-01-21 2019-05-10 闽江学院 Based on the vehicle character identifying method under CNN convolutional neural networks
CN110443184A (en) * 2019-07-31 2019-11-12 上海海事大学 ID card information extracting method, device and computer storage medium
CN111598104A (en) * 2020-06-30 2020-08-28 成都鹏业软件股份有限公司 License plate character recognition method and system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
周明辉: "运动车辆车牌识别算法的研究与实现", 《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》 *
童威: "车牌定位及畸变矫正技术的研究", 《中国优秀博硕士学位论文全文数据库(硕士)》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114387292A (en) * 2022-03-25 2022-04-22 北京市农林科学院信息技术研究中心 Image edge pixel point optimization method, device and equipment
CN115546232A (en) * 2022-10-12 2022-12-30 什维新智医疗科技(上海)有限公司 Liver ultrasonic image working area extraction method and system and electronic equipment
CN115497109A (en) * 2022-11-17 2022-12-20 山东思玛特教育科技有限公司 Character and image preprocessing method based on intelligent translation
CN115497109B (en) * 2022-11-17 2023-03-24 山东思玛特教育科技有限公司 Character and image preprocessing method based on intelligent translation
CN117253233A (en) * 2023-09-05 2023-12-19 广东奥普特科技股份有限公司 Character erasing method, device and equipment
CN117253233B (en) * 2023-09-05 2024-05-17 广东奥普特科技股份有限公司 Character erasing method, device and equipment
CN117831037A (en) * 2024-01-04 2024-04-05 北京和气聚力教育科技有限公司 Method and device for determining answer condition of objective questions in answer sheet
CN117877037A (en) * 2024-03-11 2024-04-12 杭州汇萃智能科技有限公司 OCR single character positioning and identifying method, system and computer storage medium

Also Published As

Publication number Publication date
CN113095327B (en) 2022-10-14

Similar Documents

Publication Publication Date Title
CN113095327B (en) Method and system for positioning optical character recognition area and storage medium thereof
CN107093172B (en) Character detection method and system
US6771813B1 (en) Image processing apparatus and pattern extraction apparatus
JP3748164B2 (en) Pattern extraction device
CN115147414B (en) Surface breakdown defect detection method for bipolar power transistor
US20050031208A1 (en) Apparatus for extracting ruled line from multiple-valued image
CN108805128B (en) Character segmentation method and device
US20130194448A1 (en) Rules for merging blocks of connected components in natural images
CN111353961B (en) Document curved surface correction method and device
US7146047B2 (en) Image processing apparatus and method generating binary image from a multilevel image
CN115619656A (en) Digital file deviation rectifying method and system
CN111461131A (en) Identification method, device, equipment and storage medium for ID card number information
CN110135407B (en) Sample labeling method and computer storage medium
CN103093245A (en) Signal lamp identifying method in video image
CN115082942A (en) Document image flow chart identification method, device and medium based on YOLO v5
US6532303B2 (en) Line direction deciding device, image inclination detecting device and image inclination correcting device
CN110210467A (en) A kind of formula localization method, image processing apparatus, the storage medium of text image
CN101799922A (en) Method and device for detecting strokes of characters, method and device for locating lines of characters, and method and device for judging repeat of subtitles
JPH09311905A (en) Line detecting method and character recognition device
CN112560856A (en) License plate detection and identification method, device, equipment and storage medium
CN111325199B (en) Text inclination angle detection method and device
JP2021149452A (en) Image processing device, control method and control program
JP2008028716A (en) Image processing method and apparatus
CN116503871A (en) Character segmentation preprocessing method, terminal device and computer readable storage medium
CN112257764A (en) License plate classification method, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant