WO2014129018A1 - Character recognition device, character recognition method, and recording medium - Google Patents

Character recognition device, character recognition method, and recording medium Download PDF

Info

Publication number
WO2014129018A1
WO2014129018A1 PCT/JP2013/079265 JP2013079265W WO2014129018A1 WO 2014129018 A1 WO2014129018 A1 WO 2014129018A1 JP 2013079265 W JP2013079265 W JP 2013079265W WO 2014129018 A1 WO2014129018 A1 WO 2014129018A1
Authority
WO
WIPO (PCT)
Prior art keywords
character
image
width direction
candidate
range
Prior art date
Application number
PCT/JP2013/079265
Other languages
French (fr)
Japanese (ja)
Inventor
洋平 小島
杉本 喜一
拓馬 岡▲崎▼
健太 中尾
Original Assignee
三菱重工業株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 三菱重工業株式会社 filed Critical 三菱重工業株式会社
Priority to SG11201506568XA priority Critical patent/SG11201506568XA/en
Priority to MYPI2015702775A priority patent/MY177406A/en
Publication of WO2014129018A1 publication Critical patent/WO2014129018A1/en

Links

Images

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/017Detecting movement of traffic to be counted or controlled identifying vehicles
    • G08G1/0175Detecting movement of traffic to be counted or controlled identifying vehicles by photographing vehicles, e.g. when violating traffic rules
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • G06V20/63Scene text, e.g. street names
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/18Extraction of features or characteristics of the image
    • G06V30/18086Extraction of features or characteristics of the image by performing operations within image blocks or by using histograms
    • G06V30/18095Summing image-intensity values; Projection and histogram analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • G06V20/625License plates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Definitions

  • the present invention relates to a character recognition device, a character recognition method, and a recording medium, and more particularly, to a character recognition device, a character recognition method, and a recording medium for performing character recognition on a grayscale image.
  • the technology for performing character recognition on a captured image captured by an imaging apparatus is applied to various uses including license plate recognition.
  • character recognition for a captured image generally, processing for generating a binarized image from the captured image is performed, and character recognition processing is further performed for the binarized image.
  • Such a technique is disclosed in, for example, Japanese Patent No. 3411785 (Patent Document 1).
  • Patent Document 2 discloses a technique for separating characters in contact in a binarized image, but this technique has a problem that a good binarized image cannot be obtained. It is not a fundamental solution.
  • an object of the present invention is to provide a technique for appropriately setting a character candidate region when performing character recognition processing on a grayscale image.
  • a character recognition device includes a character height direction range detection unit, a character width direction range detection unit, a character center candidate region determination unit, a character candidate region setting unit, a character recognition unit, Character identification result determination means.
  • the character height direction range detecting means is a grayscale image and the character width direction in which the character string exists for a target image that is an image showing a character string including a plurality of characters arranged in the character width direction.
  • the character height direction range which is the range in the character height direction perpendicular to the character height is detected.
  • the character width direction range detecting means detects a character width direction range that is a range in the character width direction in which each character of the character string exists for the target image.
  • the character center candidate area determining means includes, for the target image, a character center for each character of the character string as an area including the center of the character height direction range in the character height direction and the center of the character width direction range in the character width direction.
  • a candidate area is determined.
  • the character candidate area setting means sets, for each point in the character center candidate area, a plurality of rectangular character candidate areas in which the intersections of diagonal lines match each point.
  • the character recognition means performs character recognition processing on each of the plurality of character candidate area portions of the target image, and obtains the respective character recognition results of the plurality of character candidate areas for each character of the character string.
  • the character identification result determination means determines the character identification result for each character of the character string from the character recognition results of each of the plurality of character candidate areas.
  • the character height direction range detecting means calculates a first projection histogram, which is a distribution of the sum of luminance values of pixels arranged in the character width direction, calculated for each position in the character height direction of the target image. Generate and smooth the first projection histogram to calculate first smoothed data, and set the upper end of the range in the character height direction as the position where the sign of the difference between the first projection histogram and the smoothed data is inverted. And the lower end are detected.
  • a first projection histogram is a distribution of the sum of luminance values of pixels arranged in the character width direction, calculated for each position in the character height direction of the target image.
  • the character width direction range detection means generates a second projection histogram that is a distribution of the sum of luminance values of pixels arranged in the character height direction, calculated for each position in the character width direction of the target image. Then, smoothing processing is performed on the second projection histogram to calculate second smoothed data, and the start point candidate and the end point candidate are determined from the position where the sign of the difference between the second projection histogram and the second smoothed data is inverted. Is detected, the start point of the character width direction range of each character of the character string is selected from the start point candidates, and the end point of the character width direction range of each character of the character string is selected from the end point candidates.
  • the present invention is applied when the target image is a plate area image on which a license plate is projected or an image obtained by performing image processing on a plate area image.
  • the target image may be an image obtained by performing preprocessing on a plate area image on which a license plate is displayed.
  • pre-processing the luminance of the high-brightness portion is suppressed with respect to the plate region image, and linear conversion is performed so that the average value and the standard deviation of the pixel luminances have predetermined values, respectively.
  • a brightness correction process is performed, an image obtained by the brightness correction process is processed by a differential filter to generate an edge image, and a hole filling process is performed to fill the inside of the character outline included in the edge image. , Processing to remove low-frequency components in the character width direction is performed on the image obtained by the filling process
  • the character recognition method is a grayscale image, and there is a character string for a target image that is an image showing a character string including a plurality of characters arranged in the character width direction.
  • a step of detecting a character height direction range that is a range in a character height direction perpendicular to the character width direction, and a character width direction range that is a range in the character width direction in which each character of the character string exists for the target image And a character center candidate area for each character of the character string as an area including the center of the character height direction range in the character height direction and the center of the character width direction range in the character width direction.
  • a program recorded on a recording medium causes an arithmetic unit to execute the following steps: a character string that is a grayscale image and includes a plurality of characters arranged in the character width direction A step of detecting a character height direction range that is a range in a character height direction perpendicular to the character width direction, in which the character string exists, and for each character of the character string for the target image A step of detecting a character width direction range that is a character width direction range in which the character string exists, including the center of the character height direction range in the character height direction and the center of the character width direction range in the character width direction for the target image Determining a character center candidate region for each character of the character string as a region, setting a plurality of character candidate regions for each position of the character center candidate region, A step of performing character recognition processing for each of a plurality of character candidate areas of an image to obtain a character recognition result of each of the plurality of character candidate areas, and a character string from the character recognition
  • character candidate areas can be appropriately set when character recognition processing is performed on a grayscale image.
  • FIG. 1 is a block diagram showing a configuration of a character recognition device 1 according to an embodiment of the present invention.
  • the character recognition device 1 is used for license plate recognition.
  • the character recognition device 1 performs character recognition processing on the plate area image data 21 supplied from the outside to generate plate recognition data 22.
  • the plate area image data 21 is image data of a plate area image, that is, an image in which a license plate (automobile registration number mark or vehicle number mark) is projected. For example, from a captured image obtained by photographing a vehicle, It is obtained by image processing that cuts out a portion corresponding to the license plate.
  • the plate area image data 21 is grayscale image data.
  • the plate recognition data 22 is data indicating a character recognition result (that is, a recognized character) obtained by the character recognition process.
  • the character recognition device 1 includes an image processing IC (integrated circuit) 2, an external interface 3, an external storage device 4, a memory 5, and a ROM (read only memory) 6.
  • the external interface 3 supplies the plate area image data 21 received from the outside to the image processing IC 2.
  • the external storage device 4 stores data generated in the character recognition process in the character recognition device 1.
  • Data stored in the external storage device 4 includes plate recognition data 22.
  • the memory 5 is used as a working area for arithmetic processing by the image processing IC 2.
  • the ROM 6 stores a program executed by the image processing IC 2.
  • the program stored in the ROM 6 includes character recognition software 6a that is a program for executing character recognition processing.
  • the recording medium storing the character recognition software 6a may be used for installing the character recognition software 6a in the ROM 6.
  • the character recognition software 6a may be stored in the external storage device 4.
  • a recording medium for storing the character recognition software 6a may be used for installing the character recognition software 6a in the external storage device 4.
  • the image processing IC 2 includes an arithmetic module 11, an image input interface 12, a data input / output interface 13, a memory controller 14, and a ROM controller 15.
  • the arithmetic module 11, the image input interface 12, the data input / output interface 13, the memory controller 14, and the ROM controller 15 are connected by an internal bus 16.
  • the arithmetic module 11 executes the character recognition software 6 a while using the memory 5 as a working area, and performs character recognition processing on the plate area image data 21.
  • the image input interface 12 is an interface used for inputting the plate area image data 21 to the image processing IC 2.
  • the data input / output interface 13 is an interface for accessing the external storage device 4.
  • the memory controller 14 is an interface for accessing the memory 5.
  • the ROM controller 15 is an interface for accessing the ROM 6.
  • FIG. 2 is a flowchart showing a character recognition method in this embodiment.
  • the character recognition software 6a described above is a program code group for executing this character recognition method.
  • the character recognition method of the present embodiment generally includes pre-processing (step S01), character height direction range detection processing (step S02), character width direction range detection processing (step S03), and character center candidate position region setting. Processing (step S04) and character region / character recognition result determination processing (step S05) are included. Hereinafter, each processing will be described in detail.
  • preprocessing is performed on the plate area image data 21.
  • the preprocessing is image processing for eliminating disturbance components around the character as much as possible and emphasizing a line (character line) constituting the character. It should be noted that the preprocessing described below is suitable for performing highly accurate character recognition, but is not essential in principle.
  • a luminance correction process is performed (step S11).
  • a process of suppressing the brightness of a portion that is high due to halation or the like is performed, and further, the average value and standard deviation of the brightness of the pixels are set to predetermined values, respectively. Linear transformation is performed.
  • step S12 the image data obtained by the brightness correction in step S01 is processed by a Sobel filter or other differential filter, and an edge image is generated (step S12). A character outline appears in the edge image.
  • a process of filling the inside of the character outline included in the edge image is performed on the obtained edge image, and a filled image is generated (step S13).
  • Such hole filling processing can be realized, for example, by performing processing using a maximum value filter and a minimum value filter on the edge image.
  • the filter size in the character width direction (that is, horizontal direction) is selected to be, for example, about the thickness of the character line
  • the filter size in the character height direction that is, vertical direction
  • the plate area image and the edge image as a whole have characters arranged in the horizontal direction.
  • a process of removing low frequency components in the character width direction is performed on the hole-filled image obtained by the process in step S13 (step S14).
  • This processing is performed in order to remove an image other than the letters of the license plate, such as a bumper or trunk of a vehicle, when the image is included in the hole-filled image. More specifically, the processing is performed by performing a process using a minimum value filter, and further performing a process using a maximum value filter, and then calculating a difference from the hole-filled image obtained in step S13.
  • the difference image data is the image data obtained in step S14.
  • Image data of the image obtained by the preprocessing is stored in the external storage device 4.
  • the image obtained by the preprocessing is a grayscale image (not a binarized image).
  • character height direction range detection processing is performed on the image data obtained by the preprocessing (step S01) (step S02).
  • the upper and lower ends of the character height direction range (character height direction range) in which the character string exists in the image to be processed (that is, the image obtained by the preprocessing) It is a process to detect.
  • FIG. 4A is a flowchart illustrating processing performed in the character height direction range detection processing
  • FIG. 4B is a diagram conceptually illustrating the character height direction range detection processing.
  • a y-axis projection histogram is generated (step S21).
  • the y-axis projection histogram is a distribution of the sum of luminance values of pixels arranged in the character width direction, calculated for each position in the character height direction of the image. As shown in FIG.
  • the y-axis projection histogram S ⁇ 1 (y) is processed by a smoothing filter (smoothing process) to calculate smoothed data (step S22).
  • the filter size of the smoothing filter is selected as a size corresponding to the height of one character.
  • a process for calculating a moving average is used as the smoothing process. That is, as the smoothed data, moving average S ⁇ .SIGMA.1 the y-axis projection histogram S ⁇ 1 (y) (y) is calculated.
  • a graph 32 represents the moving average S ⁇ ⁇ 1 (y).
  • the upper end and the lower end of the range in the character height direction are detected using the y-axis projection histogram S ⁇ 1 (y) and the moving average S ⁇ ⁇ 1 (y) (step S24).
  • the peak point 33 which is the position in the character height direction where the smoothed data takes the maximum value.
  • the y-coordinate at which the moving average S ⁇ ⁇ 1 (y) has the maximum value is detected as the peak point 33.
  • the moving average S ⁇ ⁇ 1 (y) is used as the smoothed data
  • the position where the moving average S ⁇ ⁇ 1 (y) becomes larger than the value of the y-axis projection histogram S ⁇ 1 (y) for the first time may be detected as the positions of the upper end 34 and the lower end 46.
  • any one of the pixels may be detected as the positions of the upper end 34 and the lower end 46.
  • step S24 it is determined whether the plate area image is an image of the first plate or the second plate.
  • the plate region image is a two-stage plate image
  • processing for detecting the upper end 34 and the lower end 35 is performed for one of the two character strings included in the license plate.
  • the plate region image is a two-stage plate image
  • the same processing as the processing in steps S21 to S23 is performed on the other character string on which the processing for detecting the upper end 34 and the lower end 35 is not performed.
  • the upper end and the lower end of the character height direction range of the other character string are detected (step S25).
  • step S24 may not be performed. If the license plate imaged in the plate area image is a one-stage plate, the character height direction range detection processing is completed in the processing in steps S21 to S23 described above. If the license plate transferred to the plate area image is a two-stage plate, the above-mentioned character string for which the processing for detecting the upper end 34 and the lower end 35 is not performed in step S25 is performed again. The same processing as that in steps S21 to S23 is performed.
  • the characters written on the license plate in Japan are the upper row in which the characters of the place name 53 and the characters of the classification number 54, which are composed of relatively small characters, are arranged in the horizontal direction.
  • the upper end 34 and the lower end 35 of the character height direction range are detected for the lower character string 52 in the processing of steps S21 to S23. It will be.
  • the same processing as the processing of steps S21 to S23 is performed on the upper character string 51, and the upper character string is processed.
  • the upper end and the lower end of 51 in the character height direction range may be detected (step S26).
  • the upper and lower end data indicating the positions of the upper end and the lower end detected by the above-described character height direction range detection process are stored in the external storage device 4.
  • character width direction range detection processing is performed on the image data obtained by the preprocessing (step S01) (step S03).
  • the character width direction range detection process is a horizontal start point candidate (start point candidate) and end point candidate (end point) of a horizontal range in which each character is considered to exist (hereinafter referred to as “character width direction range”). This is a process for detecting a candidate.
  • the starting point candidates detected by the character width direction range detection process are only “candidates” and are not finally determined as the starting point in the horizontal direction of the area where each character exists.
  • the end point candidate is merely a “candidate”, and is not finally determined as the end point in the horizontal direction of the region where each character exists.
  • FIG. 6 is a flowchart showing processing performed in the character width direction range detection processing
  • FIG. 7 conceptually illustrates the character width direction range detection processing.
  • an x-axis projection histogram is generated (step S31).
  • the x-axis projection histogram is a distribution of the sum of luminance values of pixels arranged in the vertical direction, calculated for each position in the horizontal direction of the image.
  • the x-axis projection histogram is expressed as S ⁇ 2 (x) defined by the following equation (2).
  • S ⁇ 2 (x) ⁇ B (x, j) (2)
  • is the sum for all pixels whose x-axis coordinates are x.
  • a graph 41 represents an x-axis projection histogram S ⁇ 2 (x).
  • the smoothing data is calculated by performing processing (smoothing processing) by the smoothing filter on the x-axis projection histogram S ⁇ 2 (x) (step S32).
  • the filter size of the smoothing filter is selected as a size corresponding to the width of one character.
  • a process for calculating a moving average is used as the smoothing process. That is, as the smoothed data, moving average S ⁇ .SIGMA.2 the x-axis projection histogram S ⁇ 2 (x) (x) is calculated.
  • a graph 42 represents the moving average S ⁇ ⁇ 2 (x).
  • the start point candidate and the end point candidate of the character width direction range of each character are detected (step S24).
  • the start point candidate and the end point candidate are detected by searching from one end of the image, more specifically from the left end. That is, the start point candidate is detected as the start point candidate of the character width direction range of each character, and the end point candidate is detected as the end point candidate of the character width direction range.
  • the difference between the moving average and the x-axis projection histogram is calculated for each x coordinate, and the sign of the difference between the moving average and the x-axis projection histogram is reversed for the start point candidate and the end point candidate. It is detected as a position.
  • the difference S ⁇ .SIGMA.2 minus x-axis projection histogram from the moving average (x) -S ⁇ 2 (x) is calculated for each x coordinate
  • the starting point candidates, the difference S ⁇ ⁇ 2 (x) -S ⁇ 2 ( x) is calculated as an x coordinate in which the sign of “+” is inverted from plus to minus
  • the end point candidate is calculated as an x coordinate in which the sign of the difference S ⁇ 2 (x) ⁇ S ⁇ 2 (x) is inverted from minus to plus.
  • the detected start point candidate is indicated by a symbol “ ⁇ ”
  • the detected end point candidate is indicated by a symbol “ ⁇ ”.
  • the processing for detecting the start point candidate and the end point candidate is performed for one of the character strings arranged in the horizontal direction.
  • the plate region image is a two-stage plate image
  • the above-described processing of steps S31 to S33 is performed for the other character string for which the start point candidate and the end point candidate are not detected (step S34).
  • the positions of the detected start point candidates and end point candidates are stored in the external storage device 4 as start / end point candidate data.
  • Character center candidate position area setting process In the character center candidate position area setting process (step S04), the positions of the upper and lower ends of the character height direction range detected by the above-described character height direction range detection process, and the character width direction Based on the data of the start point candidate and the end point candidate detected by the range detection process, the character center candidate region of each character is determined.
  • the character center candidate region is a region that is a candidate for a position where the center of each character exists.
  • any position within the character center candidate area is determined as the position where the center of each character exists.
  • the range in the character height direction (that is, the y-axis direction) of the character center candidate region is determined as a region including the center position of the character height direction range detected by the character height direction range detection process (step S41). .
  • the range in the character height direction of the character center candidate region is common to all characters included in the character string arranged in the horizontal direction.
  • the position candidate in the vertical direction of the center of each character has a degree of freedom of (2m) pixels.
  • the reason why a certain degree of freedom is given to the candidate of the position where the center of each character exists is that it is assumed that the license plate is photographed somewhat obliquely in the plate region image.
  • the range in the horizontal direction (that is, the x-axis direction) of the character center candidate region is determined (step S42).
  • the character width direction range corresponding to each character is determined from the start point candidate and the end point candidate detected by the character width direction range detection process described above. A start and end pair is selected.
  • the start point of the character width direction range is selected from the start point candidates detected by the character width direction range detection process, and the end point of the character width direction range is selected from the end point candidates.
  • the start point candidate and the end point candidate are detected by the search from the left side
  • the i th start point candidate from the left is the start point
  • the i th from the left A certain end point candidate may be determined as the end point.
  • there is a range that is considered to be appropriate for the width of the character and when a start point and an end point that are outside the range are determined, a pair of start point and end point is used by using adjacent start point candidates or end point candidates instead. Is determined. By such a procedure, the pair of the start point and end point of the character width direction range of each character is determined.
  • the range in the horizontal direction (that is, the x-axis direction) of the character center candidate region is determined as a region including the center of the character width direction range determined for each character.
  • the position candidate in the horizontal direction of the center of each character has a degree of freedom of (2k) pixels. This is due to the fact that there is some variation in the width of the characters shown in the plate area image.
  • the processing for determining the character center candidate region for each character is performed for one of the character strings arranged in the horizontal direction.
  • the plate region image is a two-stage plate image
  • the above-described processing of steps S41 to S42 is performed for the other character string that has not been subjected to the processing for determining the character center candidate region (step S43).
  • Data indicating the range of the determined character center candidate region (character center candidate region data) is stored in the external storage device 4.
  • step S05 a process for determining the area where each character exists and the character recognition result is performed.
  • FIG. 10 is a flowchart showing processing performed in the character area / character recognition result determination processing
  • FIGS. 11 and 12 are diagrams conceptually explaining the character width direction range detection processing.
  • a character center candidate region corresponding to the character to be processed first (in this embodiment, the character located at the leftmost) is selected (step S50).
  • a plurality of character candidate areas having different heights and / or widths are set for each point (each pixel) inside the selected character center candidate area (step S51). For example, when the character center candidate region is 9 pixels of 3 ⁇ 3, nine character candidate regions corresponding to each pixel of the 9 pixels are set.
  • the character candidate area is a rectangle (that is, a rectangle (including a square)).
  • FIG. 11 is a diagram illustrating a relationship between a specific point 61 inside the character center candidate region and the character candidate region 62 set for the specific point. The character candidate region 62 set for a specific point 61 inside the character center candidate region is determined so that the intersection of the diagonal lines 63 of the character candidate region 62 coincides with the specific point 61.
  • the character candidate area has a reference size (reference height and width), and each character candidate area has the same height and width as the reference size. And the width or the height and width increased or decreased from the reference size.
  • the maximum increase / decrease amount of the height and width is set in advance as a parameter.
  • a character recognition process is performed on the image in each character candidate area (step S52).
  • character recognition processing is performed by pattern matching using a template (template matching method), and in the character recognition processing, a character recognition result and its evaluation value are determined.
  • the evaluation value for example, the similarity calculated in the template matching method may be used.
  • the character candidate area having the highest evaluation value and the character recognition result of the image of the character candidate area are extracted (step S53). For example, when the character center candidate region is 9 pixels, the character candidate region having the highest evaluation value among the nine character candidate regions and the character recognition result of the image of the character candidate region are extracted.
  • step S54 When the evaluation value is greater than or equal to a predetermined reference value (step S54: Yes), the character candidate area with the highest evaluation value and the character recognition result are the character area and character recognition result in which the character to be processed exists. Confirmed (step S55). Otherwise (step S54: No), the character recognition result is rejected assuming that a region not including characters is given as a character candidate region.
  • steps S51 to S55 are repeated until they are executed for all character center candidate regions (step S56). If there is a character center candidate area that has not been subjected to the processes of steps S51 to S55, the character center candidate area corresponding to the next character is selected (step S57), and the processes of steps S51 to S55 are performed again.
  • the processing of steps S51 to S55 is performed for all the character center candidate regions, the character region in which each character exists and the character recognition result are determined for all characters (step S58). Data indicating the confirmed character region and character recognition result is stored in the external storage device 4 as plate recognition data 22.
  • a character center candidate region is determined for each character based on the y-axis projection histogram and the x-axis projection histogram, and the character center candidate region inside the character center candidate region is determined.
  • a character candidate area is determined for each point.
  • a character candidate area is set in a state where a certain degree of freedom is given to the center of each character, and erroneous recognition can be suppressed.
  • the character center candidate area is limited to some extent by the detection result of the character height direction range by the character height direction range detection process and the detection result of the character width direction range by the character width direction range detection process. The number of character candidate areas to be set is suppressed to a certain extent. For this reason, the data processing amount of character recognition processing can be suppressed and processing time can be shortened.
  • the present invention has been specifically described above, but the present invention is not limited to the above embodiment. It will be apparent to those skilled in the art that the present invention can be implemented with various modifications. For example, in the above description, an embodiment in which the present invention is applied to license plate recognition (recognition of characters written on a license plate) is described. However, the present invention generally recognizes characters for a captured image that is a grayscale image. Applicable.

Abstract

In a target image, said target image being a grayscale image, this character recognition device detects both a character-height-direction extent, i.e. the extent in a character-height direction perpendicular to a character-width direction within which a string exists, and a character-width-direction extent for each character in said string, i.e. the extent in the character-width direction within which each character exists. For each character in the string, the character recognition device then determines a candidate character-center region that contains the center of the character-height-direction extent in the character-height direction and the center of the character-width-direction extent of that character in the character-width direction, and for each point in that candidate character-center region, the character recognition device sets a plurality of rectangular candidate character regions such that the diagonals of each of said candidate character regions intersect at that point. The character recognition device performs a character recognition process on the part of the target image corresponding to each of the plurality of candidate character regions, thereby obtaining a character recognition result for each of the plurality of candidate character regions for each character in the string, and for each character in the string, the character recognition device finalizes a character recognition result from the character recognition results for the plurality of candidate character regions for that character.

Description

文字認識装置、文字認識方法及び記録媒体Character recognition device, character recognition method, and recording medium
 本発明は、文字認識装置、文字認識方法及び記録媒体に関し、特に、グレースケール画像に対して文字認識を行うための文字認識装置、文字認識方法及び記録媒体に関する。 The present invention relates to a character recognition device, a character recognition method, and a recording medium, and more particularly, to a character recognition device, a character recognition method, and a recording medium for performing character recognition on a grayscale image.
 撮像装置によって撮像された撮像画像について文字認識を行う技術は、ナンバープレート認識を始めとして、様々な用途に応用されている。撮像画像に対する文字認識では、一般的に、撮像画像から2値化画像を生成する処理が行われ、更に、該2値化画像について文字認識処理が行われる。このような技術は、例えば、特許第3411795号公報(特許文献1)に開示されている。 The technology for performing character recognition on a captured image captured by an imaging apparatus is applied to various uses including license plate recognition. In character recognition for a captured image, generally, processing for generating a binarized image from the captured image is performed, and character recognition processing is further performed for the binarized image. Such a technique is disclosed in, for example, Japanese Patent No. 3411785 (Patent Document 1).
 撮像画像から2値化画像を生成し、該2値化画像に対して文字認識処理を行う技術の一つの問題は、撮像画像の画質が悪いと、良好な2値化画像が得られないことである。例えば、撮像画像においてコントラストが不十分であると、2値化処理において隣接した文字が連結した2値化画像が生成され、これは、文字認識の精度を低下させる。特許第3798582号公報(特許文献2)は、2値化画像において接触している文字を分離する技術を開示しているが、この技術は、良好な2値化画像が得られないという問題を根本的に解決するものではない。 One problem with the technique of generating a binarized image from a captured image and performing character recognition processing on the binarized image is that a good binarized image cannot be obtained if the image quality of the captured image is poor. It is. For example, if the contrast is insufficient in the captured image, a binarized image in which adjacent characters are connected in the binarization process is generated, which reduces the accuracy of character recognition. Japanese Patent No. 3798582 (Patent Document 2) discloses a technique for separating characters in contact in a binarized image, but this technique has a problem that a good binarized image cannot be obtained. It is not a fundamental solution.
 このような問題に対処するために、発明者らは、2値化処理を行わずに文字認識処理を行うこと、即ち、グレースケール画像(濃淡画像)に対して文字認識処理を行うことを検討している。そして、発明者らの検討によれば、2値化を行わずに文字認識処理を行う上での一つの問題は、各文字が存在する領域の候補(以下では、「文字候補領域」ということがある)を適切に設定することである。グレースケール画像に対して文字認識処理を行う場合に文字候補領域を不適切に設定すると、誤った文字認識結果が得られる、即ち、文字の誤認識が発生することになる。2値化画像では、ノイズを除去した後、(バックグラウンドが白である場合には)黒の画素が連続して存在する部分を文字候補領域であるとして設定する処理を行えばよいが、グレースケール画像については、このような処理を行うことはできない。多数の文字候補領域を適宜に設定することで、文字の誤認識を抑制することはできるが、その一方で、過剰に多数の文字候補領域を設定することは、文字認識処理におけるデータ処理量の増大、即ち、処理時間の長大化を招く。 In order to deal with such a problem, the inventors consider performing character recognition processing without performing binarization processing, that is, performing character recognition processing on a grayscale image (grayscale image). is doing. According to the study by the inventors, one problem in performing character recognition processing without performing binarization is that each character exists in an area candidate (hereinafter referred to as a “character candidate area”). There is an appropriate setting. If character candidate regions are set inappropriately when performing character recognition processing on a grayscale image, an incorrect character recognition result is obtained, that is, erroneous character recognition occurs. In a binarized image, after removing noise, a process of setting a portion where black pixels are continuously present as a character candidate region (when the background is white) may be performed. Such processing cannot be performed on the scale image. Although it is possible to suppress erroneous recognition of characters by appropriately setting a large number of character candidate regions, on the other hand, setting an excessively large number of character candidate regions is an important factor in the amount of data processing in character recognition processing. Increase, that is, increase the processing time.
 このような背景から、グレースケール画像に対して文字認識処理を行う際に、適切に文字候補領域を設定するための技術を提供することが求められている。 From such a background, it is required to provide a technique for appropriately setting a character candidate area when performing character recognition processing on a grayscale image.
特許第3411795号公報Japanese Patent No. 3411795 特許第3798582号公報Japanese Patent No. 3798582
 したがって、本発明の目的は、グレースケール画像に対して文字認識処理を行う際に、適切に文字候補領域を設定するための技術を提供することにある。 Therefore, an object of the present invention is to provide a technique for appropriately setting a character candidate region when performing character recognition processing on a grayscale image.
 本発明の一の観点では、文字認識装置が、文字高さ方向範囲検出手段と、文字幅方向範囲検出手段と、文字中心候補領域決定手段と、文字候補領域設定手段と、文字認識手段と、文字識別結果確定手段とを具備する。文字高さ方向範囲検出手段は、グレースケール画像であり、且つ、文字幅方向に並んだ複数の文字を含む文字列が映された画像である対象画像について、文字列が存在する、文字幅方向と垂直な文字高さ方向の範囲である文字高さ方向範囲を検出する。文字幅方向範囲検出手段は、対象画像について、文字列の各文字が存在する文字幅方向の範囲である文字幅方向範囲を検出する。文字中心候補領域決定手段は、対象画像について、文字高さ方向における文字高さ方向範囲の中心を含み、文字幅方向における文字幅方向範囲の中心を含む領域として、文字列の各文字について文字中心候補領域を決定する。文字候補領域設定手段は、文字中心候補領域の各点について、対角線の交点が各点に一致する複数の矩形の文字候補領域を設定する。文字認識手段は、対象画像の複数の文字候補領域の部分のそれぞれに対して文字認識処理を行って、文字列の各文字について、複数の文字候補領域のそれぞれの文字認識結果を得る。文字識別結果確定手段は、複数の文字候補領域それぞれの文字認識結果から、文字列の各文字についての文字識別結果を確定する。 In one aspect of the present invention, a character recognition device includes a character height direction range detection unit, a character width direction range detection unit, a character center candidate region determination unit, a character candidate region setting unit, a character recognition unit, Character identification result determination means. The character height direction range detecting means is a grayscale image and the character width direction in which the character string exists for a target image that is an image showing a character string including a plurality of characters arranged in the character width direction. The character height direction range which is the range in the character height direction perpendicular to the character height is detected. The character width direction range detecting means detects a character width direction range that is a range in the character width direction in which each character of the character string exists for the target image. The character center candidate area determining means includes, for the target image, a character center for each character of the character string as an area including the center of the character height direction range in the character height direction and the center of the character width direction range in the character width direction. A candidate area is determined. The character candidate area setting means sets, for each point in the character center candidate area, a plurality of rectangular character candidate areas in which the intersections of diagonal lines match each point. The character recognition means performs character recognition processing on each of the plurality of character candidate area portions of the target image, and obtains the respective character recognition results of the plurality of character candidate areas for each character of the character string. The character identification result determination means determines the character identification result for each character of the character string from the character recognition results of each of the plurality of character candidate areas.
 一実施形態では、文字高さ方向範囲検出手段は、対象画像の文字高さ方向の各位置について算出された、文字幅方向に並んだ画素の輝度値の和の分布である第1射影ヒストグラムを生成し、第1射影ヒストグラムに対して平滑化処理を行って第1平滑化データを算出し、第1射影ヒストグラムと平滑化データとの差の符号が反転する位置として文字高さ方向範囲の上端と下端とを検出する。 In one embodiment, the character height direction range detecting means calculates a first projection histogram, which is a distribution of the sum of luminance values of pixels arranged in the character width direction, calculated for each position in the character height direction of the target image. Generate and smooth the first projection histogram to calculate first smoothed data, and set the upper end of the range in the character height direction as the position where the sign of the difference between the first projection histogram and the smoothed data is inverted. And the lower end are detected.
 一実施形態では、文字幅方向範囲検出手段は、対象画像の文字幅方向の各位置について算出された、文字高さ方向に並んだ画素の輝度値の和の分布である第2射影ヒストグラムを生成し、第2射影ヒストグラムに対して平滑化処理を行って第2平滑化データを算出し、第2射影ヒストグラムと第2平滑化データとの差の符号が反転する位置から始点候補と終点候補とを検出し、始点候補のうちから文字列の各文字の文字幅方向範囲の始点を選択し、終点候補のうちから文字列の各文字の文字幅方向範囲の終点を選択する。 In one embodiment, the character width direction range detection means generates a second projection histogram that is a distribution of the sum of luminance values of pixels arranged in the character height direction, calculated for each position in the character width direction of the target image. Then, smoothing processing is performed on the second projection histogram to calculate second smoothed data, and the start point candidate and the end point candidate are determined from the position where the sign of the difference between the second projection histogram and the second smoothed data is inverted. Is detected, the start point of the character width direction range of each character of the character string is selected from the start point candidates, and the end point of the character width direction range of each character of the character string is selected from the end point candidates.
 好適な応用例としては、本発明は、対象画像が、ナンバープレートが映されたプレート領域画像、又は、プレート領域画像に対して画像処理を行うことで得られる画像である場合に適用される。一実施形態では、対象画像が、ナンバープレートが映されたプレート領域画像に対して前処理を行うことで得られる画像であってもよい。前処理では、プレート領域画像に対して、高輝度になっている部分の輝度を抑制すると共に、画素の輝度の平均値と標準偏差とが、それぞれ、所定の値になるように線形変換を行う輝度補正処理が行われ、輝度補正処理によって得られた画像に対して微分フィルタによる処理が行われてエッジ画像が生成され、エッジ画像に含まれる文字輪郭線の内側を穴埋めする穴埋め処理が行われ、穴埋め処理で得られた画像に対して文字幅方向の低周波成分を除去する処理が行われる As a preferred application example, the present invention is applied when the target image is a plate area image on which a license plate is projected or an image obtained by performing image processing on a plate area image. In one embodiment, the target image may be an image obtained by performing preprocessing on a plate area image on which a license plate is displayed. In the pre-processing, the luminance of the high-brightness portion is suppressed with respect to the plate region image, and linear conversion is performed so that the average value and the standard deviation of the pixel luminances have predetermined values, respectively. A brightness correction process is performed, an image obtained by the brightness correction process is processed by a differential filter to generate an edge image, and a hole filling process is performed to fill the inside of the character outline included in the edge image. , Processing to remove low-frequency components in the character width direction is performed on the image obtained by the filling process
 本発明の他の観点においては、文字認識方法が、グレースケール画像であり、且つ、文字幅方向に並んだ複数の文字を含む文字列が映された画像である対象画像について、文字列が存在する、文字幅方向と垂直な文字高さ方向の範囲である文字高さ方向範囲を検出するステップと、対象画像について、文字列の各文字が存在する文字幅方向の範囲である文字幅方向範囲を検出するステップと、対象画像について、文字高さ方向における文字高さ方向範囲の中心を含み、文字幅方向における文字幅方向範囲の中心を含む領域として、文字列の各文字について文字中心候補領域を決定するステップと、文字中心候補領域の各点について、対角線の交点が各点に一致する複数の矩形の文字候補領域を設定するステップと、対象画像の複数の文字候補領域の部分のそれぞれに対して文字認識処理を行って、文字列の各文字について、複数の文字候補領域のそれぞれの文字認識結果を得るステップと、複数の文字候補領域それぞれの文字認識結果から、文字列の各文字についての文字識別結果を確定するステップとを具備する。 In another aspect of the present invention, the character recognition method is a grayscale image, and there is a character string for a target image that is an image showing a character string including a plurality of characters arranged in the character width direction. A step of detecting a character height direction range that is a range in a character height direction perpendicular to the character width direction, and a character width direction range that is a range in the character width direction in which each character of the character string exists for the target image And a character center candidate area for each character of the character string as an area including the center of the character height direction range in the character height direction and the center of the character width direction range in the character width direction. Determining, for each point of the character center candidate region, a step of setting a plurality of rectangular character candidate regions whose diagonal intersection points coincide with each point, and a plurality of character candidates of the target image Character recognition processing is performed on each part of the region, and for each character of the character string, a step of obtaining each character recognition result of the plurality of character candidate regions, and a character recognition result of each of the plurality of character candidate regions, Determining a character identification result for each character of the character string.
 本発明の更に他の観点においては、記録媒体に記録されたプログラムが、演算装置に、下記ステップを実行させる:グレースケール画像であり、且つ、文字幅方向に並んだ複数の文字を含む文字列が映された画像である対象画像について、文字列が存在する、文字幅方向と垂直な文字高さ方向の範囲である文字高さ方向範囲を検出するステップ、対象画像について、文字列の各文字が存在する文字幅方向の範囲である文字幅方向範囲を検出するステップ、対象画像について、文字高さ方向における文字高さ方向範囲の中心を含み、文字幅方向における文字幅方向範囲の中心を含む領域として、文字列の各文字について文字中心候補領域を決定するステップ、文字中心候補領域の各位置について、複数の文字候補領域を設定するステップ、対象画像の複数の文字候補領域のそれぞれに対して文字認識処理を行って、複数の文字候補領域のそれぞれの文字認識結果を得るステップ、及び、複数の文字候補領域それぞれの文字認識結果から、文字列の各文字についての文字識別結果を確定するステップ。 In still another aspect of the present invention, a program recorded on a recording medium causes an arithmetic unit to execute the following steps: a character string that is a grayscale image and includes a plurality of characters arranged in the character width direction A step of detecting a character height direction range that is a range in a character height direction perpendicular to the character width direction, in which the character string exists, and for each character of the character string for the target image A step of detecting a character width direction range that is a character width direction range in which the character string exists, including the center of the character height direction range in the character height direction and the center of the character width direction range in the character width direction for the target image Determining a character center candidate region for each character of the character string as a region, setting a plurality of character candidate regions for each position of the character center candidate region, A step of performing character recognition processing for each of a plurality of character candidate areas of an image to obtain a character recognition result of each of the plurality of character candidate areas, and a character string from the character recognition results of each of the plurality of character candidate areas Determining a character identification result for each of the characters.
 本発明によれば、グレースケール画像に対して文字認識処理を行う際に、適切に文字候補領域を設定することができる。 According to the present invention, character candidate areas can be appropriately set when character recognition processing is performed on a grayscale image.
本発明の一実施形態における文字認識装置の構成を示すブロック図である。It is a block diagram which shows the structure of the character recognition apparatus in one Embodiment of this invention. 本発明の一実施形態における文字認識方法の概要を示すフローチャートである。It is a flowchart which shows the outline | summary of the character recognition method in one Embodiment of this invention. 本実施形態の文字認識方法における前処理の内容を示すフローチャートである。It is a flowchart which shows the content of the pre-process in the character recognition method of this embodiment. 本実施形態の文字認識方法における文字高さ方向範囲検出処理の内容を示すフローチャートである。It is a flowchart which shows the content of the character height direction range detection process in the character recognition method of this embodiment. 本実施形態の文字認識方法における文字高さ方向範囲検出処理の内容を示す概念図である。It is a conceptual diagram which shows the content of the character height direction range detection process in the character recognition method of this embodiment. ナンバープレートの構成の例を示す概念図である。It is a conceptual diagram which shows the example of a structure of a license plate. 図5Aのナンバープレートについて行われる文字高さ方向範囲検出処理の内容を示すフローチャートである。It is a flowchart which shows the content of the character height direction range detection process performed about the number plate of FIG. 5A. 本実施形態の文字認識方法における文字幅方向範囲検出処理の内容を示すフローチャートである。It is a flowchart which shows the content of the character width direction range detection process in the character recognition method of this embodiment. 本実施形態の文字認識方法における文字幅方向範囲検出処理の内容を示す概念図である。It is a conceptual diagram which shows the content of the character width direction range detection process in the character recognition method of this embodiment. 本実施形態の文字認識方法における文字中心候補領域設定処理の内容を示すフローチャートである。It is a flowchart which shows the content of the character center candidate area | region setting process in the character recognition method of this embodiment. 本実施形態の文字認識方法における文字中心候補領域設定処理の内容を示す概念図である。It is a conceptual diagram which shows the content of the character center candidate area | region setting process in the character recognition method of this embodiment. 本実施形態の文字認識方法における文字領域・文字認識結果確定処理の内容を示すフローチャートである。It is a flowchart which shows the content of the character area and character recognition result decision process in the character recognition method of this embodiment. 文字中心候補領域の特定の点と、該特定の点について設定される文字候補領域との関係を示す概念図である。It is a conceptual diagram which shows the relationship between the specific point of a character center candidate area | region, and the character candidate area | region set about this specific point. 本実施形態における文字領域・文字認識結果確定処理において設定される文字候補領域を示す概念図である。It is a conceptual diagram which shows the character candidate area | region set in the character area and character recognition result confirmation process in this embodiment.
 図1は、本発明の一実施形態における文字認識装置1の構成を示すブロック図である。本実施形態では、文字認識装置1が、ナンバープレート認識に用いられる。即ち、文字認識装置1は、外部から供給されたプレート領域画像データ21に対して文字認識処理を行ってプレート認識データ22を生成する。ここで、プレート領域画像データ21とは、プレート領域画像、即ち、ナンバープレート(自動車登録番号標又は車両番号標)が映された画像の画像データであり、例えば、車両を撮影した撮像画像から、ナンバープレートに対応する部分を切り出す画像処理によって得られる。プレート領域画像データ21は、グレースケール画像の画像データであることに留意されたい。また、プレート認識データ22とは、文字認識処理によって得られた文字認識結果(即ち、認識された文字)を示すデータである。 FIG. 1 is a block diagram showing a configuration of a character recognition device 1 according to an embodiment of the present invention. In the present embodiment, the character recognition device 1 is used for license plate recognition. In other words, the character recognition device 1 performs character recognition processing on the plate area image data 21 supplied from the outside to generate plate recognition data 22. Here, the plate area image data 21 is image data of a plate area image, that is, an image in which a license plate (automobile registration number mark or vehicle number mark) is projected. For example, from a captured image obtained by photographing a vehicle, It is obtained by image processing that cuts out a portion corresponding to the license plate. It should be noted that the plate area image data 21 is grayscale image data. The plate recognition data 22 is data indicating a character recognition result (that is, a recognized character) obtained by the character recognition process.
 文字認識装置1は、画像処理IC(integrated circuit)2と、外部インターフェース3と、外部記憶装置4と、メモリ5と、ROM(read only memory)6とを備えている。外部インターフェース3は、外部から受け取ったプレート領域画像データ21を、画像処理IC2に供給する。外部記憶装置4は、文字認識装置1における文字認識処理において生成されるデータを保存する。外部記憶装置4に保存されるデータは、プレート認識データ22を含んでいる。メモリ5は、画像処理IC2による演算処理のワーキングエリアとして使用される。ROM6は、画像処理IC2によって実行されるプログラムを保存している。ROM6に記憶されているプログラムは、文字認識処理を実行するためのプログラムである文字認識ソフトウェア6aを含んでいる。 The character recognition device 1 includes an image processing IC (integrated circuit) 2, an external interface 3, an external storage device 4, a memory 5, and a ROM (read only memory) 6. The external interface 3 supplies the plate area image data 21 received from the outside to the image processing IC 2. The external storage device 4 stores data generated in the character recognition process in the character recognition device 1. Data stored in the external storage device 4 includes plate recognition data 22. The memory 5 is used as a working area for arithmetic processing by the image processing IC 2. The ROM 6 stores a program executed by the image processing IC 2. The program stored in the ROM 6 includes character recognition software 6a that is a program for executing character recognition processing.
 ROM6として書き換え可能な不揮発性メモリ(例えば、フラッシュメモリ)が使用される場合、ROM6への文字認識ソフトウェア6aのインストールは、文字認識ソフトウェア6aを記憶する記録媒体が用いられても良い。また、文字認識ソフトウェア6aは、外部記憶装置4に記憶されてもよい。この場合、外部記憶装置4への文字認識ソフトウェア6aのインストールには、文字認識ソフトウェア6aを記憶する記録媒体が用いられても良い。 When a rewritable nonvolatile memory (for example, a flash memory) is used as the ROM 6, the recording medium storing the character recognition software 6a may be used for installing the character recognition software 6a in the ROM 6. The character recognition software 6a may be stored in the external storage device 4. In this case, a recording medium for storing the character recognition software 6a may be used for installing the character recognition software 6a in the external storage device 4.
 画像処理IC2は、演算モジュール11と、画像入力インターフェース12と、データ入出力インターフェース13と、メモリコントローラ14と、ROMコントローラ15とを備えている。演算モジュール11と、画像入力インターフェース12と、データ入出力インターフェース13と、メモリコントローラ14と、ROMコントローラ15とは、内部バス16によって接続されている。演算モジュール11は、メモリ5をワーキングエリアとして使用しながら文字認識ソフトウェア6aを実行して、プレート領域画像データ21に対する文字認識処理を行う。画像入力インターフェース12は、画像処理IC2にプレート領域画像データ21を入力するために用いられるインターフェースである。データ入出力インターフェース13は、外部記憶装置4へのアクセスを行うためのインターフェースである。メモリコントローラ14は、メモリ5へのアクセスを行うためのインターフェースである。また、ROMコントローラ15は、ROM6へのアクセスを行うためのインターフェースである。 The image processing IC 2 includes an arithmetic module 11, an image input interface 12, a data input / output interface 13, a memory controller 14, and a ROM controller 15. The arithmetic module 11, the image input interface 12, the data input / output interface 13, the memory controller 14, and the ROM controller 15 are connected by an internal bus 16. The arithmetic module 11 executes the character recognition software 6 a while using the memory 5 as a working area, and performs character recognition processing on the plate area image data 21. The image input interface 12 is an interface used for inputting the plate area image data 21 to the image processing IC 2. The data input / output interface 13 is an interface for accessing the external storage device 4. The memory controller 14 is an interface for accessing the memory 5. The ROM controller 15 is an interface for accessing the ROM 6.
 以下に述べられる文字認識方法の各処理は、図1に示されているハードウェアを用いて実行される。以下では、本実施形態の文字認識方法について、詳細に説明する。 Each process of the character recognition method described below is executed using the hardware shown in FIG. Below, the character recognition method of this embodiment is demonstrated in detail.
 図2は、本実施形態における文字認識方法を示すフローチャートである。上述の文字認識ソフトウェア6aは、この文字認識方法を実行するためのプログラムコード群である。 FIG. 2 is a flowchart showing a character recognition method in this embodiment. The character recognition software 6a described above is a program code group for executing this character recognition method.
 本実施形態の文字認識方法は、概略的には、前処理(ステップS01)、文字高さ方向範囲検出処理(ステップS02)、文字幅方向範囲検出処理(ステップS03)、文字中心候補位置領域設定処理(ステップS04)、及び、文字領域、文字認識結果確定処理(ステップS05)を含んでいる。以下、それぞれの処理について、詳細に説明する。 The character recognition method of the present embodiment generally includes pre-processing (step S01), character height direction range detection processing (step S02), character width direction range detection processing (step S03), and character center candidate position region setting. Processing (step S04) and character region / character recognition result determination processing (step S05) are included. Hereinafter, each processing will be described in detail.
1.前処理
 まず、プレート領域画像データ21に対して、前処理(ステップS01)が行われる。前処理とは、文字の周辺の外乱成分を可能な限り排除し、また、文字を構成する線(文字線)を強調するための画像処理である。なお、以下に説明される前処理は、精度の高い文字認識を行うために好適であるが、原理的には必須ではないことに留意されたい。
1. Preprocessing First, preprocessing (step S01) is performed on the plate area image data 21. The preprocessing is image processing for eliminating disturbance components around the character as much as possible and emphasizing a line (character line) constituting the character. It should be noted that the preprocessing described below is suitable for performing highly accurate character recognition, but is not essential in principle.
 前処理(ステップS01)では、まず、輝度補正処理が行われる(ステップS11)。輝度補正処理では、まず、ハレーション等により高輝度になっている部分の輝度を抑制する処理が行われ、更に、画素の輝度の平均値と標準偏差とが、それぞれ、所定の値になるように線形変換が行われる。 In the pre-processing (step S01), first, a luminance correction process is performed (step S11). In the brightness correction process, first, a process of suppressing the brightness of a portion that is high due to halation or the like is performed, and further, the average value and standard deviation of the brightness of the pixels are set to predetermined values, respectively. Linear transformation is performed.
 更に、ステップS01の輝度補正によって得られた画像データに対し、Sobelフィルタその他の微分フィルタによる処理が行われ、エッジ画像が生成される(ステップS12)。エッジ画像には、文字輪郭線が現れる。 Further, the image data obtained by the brightness correction in step S01 is processed by a Sobel filter or other differential filter, and an edge image is generated (step S12). A character outline appears in the edge image.
 得られたエッジ画像に対し、エッジ画像に含まれる文字輪郭線の内側を穴埋めする処理が行われ、穴埋め画像が生成される(ステップS13)。このような穴埋め処理は、例えば、エッジ画像に対し、最大値フィルタと最小値フィルタによる処理を行うことで実現できる。この場合、文字幅方向(即ち、水平方向)におけるフィルタサイズは、例えば、文字線の太さ程度に選択され、文字高さ方向(即ち、垂直方向)におけるフィルタサイズは、1画素に選択される。ここで、プレート領域画像及びエッジ画像には、全体としては、水平方向に文字が並んでいることに留意されたい。 A process of filling the inside of the character outline included in the edge image is performed on the obtained edge image, and a filled image is generated (step S13). Such hole filling processing can be realized, for example, by performing processing using a maximum value filter and a minimum value filter on the edge image. In this case, the filter size in the character width direction (that is, horizontal direction) is selected to be, for example, about the thickness of the character line, and the filter size in the character height direction (that is, vertical direction) is selected to be one pixel. . Here, it should be noted that the plate area image and the edge image as a whole have characters arranged in the horizontal direction.
 更に、ステップS13における処理で得られた穴埋め画像に対し、文字幅方向の低周波成分を除去する処理が行われる(ステップS14)。この処理は、車両のバンパーやトランク等、ナンバープレートの文字以外の像が穴埋め画像に含まれている場合に、該像を除去するために行われる。より具体的には、最小値フィルタによる処理を行い、更に、最大値フィルタによる処理を行った後、ステップS13で得られる穴埋め画像との差分を算出することで行われる。該差分の画像データが、ステップS14で得られる画像データである。 Further, a process of removing low frequency components in the character width direction is performed on the hole-filled image obtained by the process in step S13 (step S14). This processing is performed in order to remove an image other than the letters of the license plate, such as a bumper or trunk of a vehicle, when the image is included in the hole-filled image. More specifically, the processing is performed by performing a process using a minimum value filter, and further performing a process using a maximum value filter, and then calculating a difference from the hole-filled image obtained in step S13. The difference image data is the image data obtained in step S14.
 以上で、前処理が完了する。前処理によって得られた画像の画像データは、外部記憶装置4に保存される。ここで、前処理によって得られる画像は、(2値化画像ではなく)グレースケール画像であることに留意されたい。 This completes the pre-processing. Image data of the image obtained by the preprocessing is stored in the external storage device 4. Here, it should be noted that the image obtained by the preprocessing is a grayscale image (not a binarized image).
2.文字高さ方向範囲検出処理
 続いて、前処理(ステップS01)によって得られた画像データに対して、文字高さ方向範囲検出処理が行われる(ステップS02)。文字高さ方向範囲検出処理とは、処理対象の画像(即ち、前処理によって得られた画像)において、文字列が存在する文字高さ方向の範囲(文字高さ方向範囲)の上端と下端を検出する処理である。図4Aは、文字高さ方向範囲検出処理において行われる処理を示すフローチャートであり、また、図4Bは、文字高さ方向範囲検出処理を概念的に説明する図である。
2. Character Height Direction Range Detection Processing Subsequently, character height direction range detection processing is performed on the image data obtained by the preprocessing (step S01) (step S02). In the character height direction range detection process, the upper and lower ends of the character height direction range (character height direction range) in which the character string exists in the image to be processed (that is, the image obtained by the preprocessing) It is a process to detect. FIG. 4A is a flowchart illustrating processing performed in the character height direction range detection processing, and FIG. 4B is a diagram conceptually illustrating the character height direction range detection processing.
 まず、y軸射影ヒストグラムが生成される(ステップS21)。ここで、y軸射影ヒストグラムとは、画像の文字高さ方向の各位置について算出された、文字幅方向に並んだ画素の輝度値の和の分布である。図4Bに図示されているように、y軸が画像の垂直方向に、x軸が水平方向に定められたxy座標系を規定した場合、座標(x,y)にある画素の輝度をB(x,y)として、y軸射影ヒストグラムは、次式(1)で定義されるSΣ1(y)として得られる:
 SΣ1(y)=ΣB(i、y)    ・・・(1)
ここで、Σは、y軸座標がyである全画素についての和である。図4Bにおいて、グラフ31は、y軸射影ヒストグラムSΣ1(y)を表わしている。
First, a y-axis projection histogram is generated (step S21). Here, the y-axis projection histogram is a distribution of the sum of luminance values of pixels arranged in the character width direction, calculated for each position in the character height direction of the image. As shown in FIG. 4B, when an xy coordinate system in which the y axis is defined in the vertical direction of the image and the x axis is defined in the horizontal direction is defined, the luminance of the pixel at the coordinates (x, y) is represented by B ( x, y), the y-axis projection histogram is obtained as S Σ1 (y) defined by the following equation (1):
S Σ1 (y) = ΣB (i, y) (1)
Here, Σ is the sum for all pixels whose y-axis coordinates are y. In FIG. 4B, a graph 31 represents a y-axis projection histogram S Σ1 (y).
 更に、y軸射影ヒストグラムSΣ1(y)に対して平滑化フィルタによる処理(平滑化処理)が行われて平滑化データが算出される(ステップS22)。平滑化フィルタのフィルタサイズは、1文字の高さに相当するサイズに選ばれる。本実施形態では、平滑化処理として移動平均を算出する処理が用いられる。即ち、平滑化データとして、y軸射影ヒストグラムSΣ1(y)の移動平均S^Σ1(y)が算出される。図4Bにおいて、グラフ32は、移動平均S^Σ1(y)を表わしている。 Further, the y-axis projection histogram S Σ1 (y) is processed by a smoothing filter (smoothing process) to calculate smoothed data (step S22). The filter size of the smoothing filter is selected as a size corresponding to the height of one character. In the present embodiment, a process for calculating a moving average is used as the smoothing process. That is, as the smoothed data, moving average S ^ .SIGMA.1 the y-axis projection histogram S Σ1 (y) (y) is calculated. In FIG. 4B, a graph 32 represents the moving average S ^ Σ1 (y).
 更に、y軸射影ヒストグラムSΣ1(y)と、その移動平均S^Σ1(y)とを用いて、文字高さ方向範囲の上端と下端が検出される(ステップS24)。ステップS24における、文字高さ方向範囲の上端と下端の検出では、まず、平滑化データが最大値をとる文字高さ方向の位置であるピーク点33が検出される。本実施形態では、移動平均S^Σ1(y)が最大値をとるy座標が、ピーク点33として検出される。更に、ピーク点33を始点として平滑化データの値がy軸射影ヒストグラムSΣ1(y)の値よりも大きくなる位置が検索され、平滑化データの値がy軸射影ヒストグラムSΣ1(y)の値よりも大きくなる位置が、文字高さ方向範囲の上端及び下端の位置として検出される。図4Bにおいて、符号34は、ステップS24で検出された、文字高さ方向範囲の上端を示しており、符号35は、文字高さ方向範囲の下端を示している。 Furthermore, the upper end and the lower end of the range in the character height direction are detected using the y-axis projection histogram S Σ1 (y) and the moving average S ^ Σ1 (y) (step S24). In the detection of the upper and lower ends of the character height direction range in step S24, first, the peak point 33, which is the position in the character height direction where the smoothed data takes the maximum value, is detected. In the present embodiment, the y-coordinate at which the moving average S ^ Σ1 (y) has the maximum value is detected as the peak point 33. Further, a position where the value of the smoothed data is larger than the value of the y-axis projection histogram S Σ1 (y) starting from the peak point 33 is searched, and the value of the smoothed data is the y-axis projection histogram S Σ1 (y). Positions larger than the value are detected as the positions of the upper end and the lower end of the character height direction range. In FIG. 4B, the code | symbol 34 has shown the upper end of the character height direction range detected by step S24, and the code | symbol 35 has shown the lower end of the character height direction range.
 平滑化データとして移動平均S^Σ1(y)が用いられる本実施形態では、移動平均S^Σ1(y)がy軸射影ヒストグラムSΣ1(y)の値よりも大きくなる位置が、文字が存在する領域の上端34及び下端35の位置として検出される。ここで、移動平均S^Σ1(y)がy軸射影ヒストグラムSΣ1(y)の値よりも初めて大きくなる位置が上端34及び下端46の位置として検出されてもよい。また、文字高さ方向の所定数の一連の画素にわたって移動平均S^Σ1(y)がy軸射影ヒストグラムSΣ1(y)の値よりも大きくなる場合に、当該一連の画素のいずれか(典型的には中央の画素)のy座標が、上端34及び下端46の位置として検出されてもよい。 In this embodiment in which the moving average S ^ Σ1 (y) is used as the smoothed data, there is a character at a position where the moving average S ^ Σ1 (y) is larger than the value of the y-axis projection histogram SΣ1 (y). It is detected as the position of the upper end 34 and the lower end 35 of the area to be performed. Here, the position where the moving average S ^ Σ1 (y) becomes larger than the value of the y-axis projection histogram SΣ1 (y) for the first time may be detected as the positions of the upper end 34 and the lower end 46. Further, when the moving average S ^ Σ1 (y) is larger than the value of the y-axis projection histogram SΣ1 (y) over a predetermined number of pixels in the character height direction, any one of the pixels (typically Specifically, the y coordinate of the center pixel) may be detected as the positions of the upper end 34 and the lower end 46.
 更に、プレート領域画像が、1段プレート、2段プレートのいずれの画像であるかが判定される(ステップS24)。プレート領域画像が2段プレートの画像である場合、上記のステップS21~S23の処理では、ナンバープレートに含まれる2つの文字列の一方について、上端34及び下端35を検出する処理が行われることになる。そこで、プレート領域画像が2段プレートの画像である場合、上端34、下端35を検出する処理が行われていない他方の文字列について上記のステップS21~S23の処理と同じ処理が行われ、該他方の文字列の、文字高さ方向範囲の上端と下端とが検出される(ステップS25)。 Further, it is determined whether the plate area image is an image of the first plate or the second plate (step S24). When the plate region image is a two-stage plate image, in the processing of steps S21 to S23, processing for detecting the upper end 34 and the lower end 35 is performed for one of the two character strings included in the license plate. Become. Therefore, when the plate region image is a two-stage plate image, the same processing as the processing in steps S21 to S23 is performed on the other character string on which the processing for detecting the upper end 34 and the lower end 35 is not performed. The upper end and the lower end of the character height direction range of the other character string are detected (step S25).
 ここで、プレート領域画像に移されているナンバープレートが1段プレート、2段プレートのいずれであるかが予め分かっている場合には、1段プレート、2段プレートのいずれの画像であるかの判定(ステップS24)は行わなくてもよい。プレート領域画像に撮像されているナンバープレートが1段プレートである場合には、上記のステップS21~S23の処理で、文字高さ方向範囲検出処理が完了する。また、プレート領域画像に移されているナンバープレートが2段プレートである場合には、ステップS25において、上端34、下端35を検出する処理が行われていない他方の文字列について、再度、上記のステップS21~S23の処理と同じ処理が行われる。 Here, if it is known in advance whether the license plate transferred to the plate area image is the first plate or the second plate, it is either the first plate or the second plate. The determination (step S24) may not be performed. If the license plate imaged in the plate area image is a one-stage plate, the character height direction range detection processing is completed in the processing in steps S21 to S23 described above. If the license plate transferred to the plate area image is a two-stage plate, the above-mentioned character string for which the processing for detecting the upper end 34 and the lower end 35 is not performed in step S25 is performed again. The same processing as that in steps S21 to S23 is performed.
 例えば、日本国のナンバープレートに記載される文字は、図5Aに図示されているように、相対的に小さな文字で構成される地名53の文字と分類番号54の文字が水平方向に並んだ上段の文字列51と、相対的に大きな文字で構成される一連指定番号55と平仮名56とが並んだ下段の文字列52とを含んでいる。そこで、日本国のナンバープレートについての文字認識処理においては、ステップS24における1段プレート、2段プレートのいずれの画像であるかの判定は行われなくてもよい。 For example, as shown in FIG. 5A, the characters written on the license plate in Japan are the upper row in which the characters of the place name 53 and the characters of the classification number 54, which are composed of relatively small characters, are arranged in the horizontal direction. , And a lower character string 52 in which a series designation number 55 composed of relatively large characters and a hiragana 56 are arranged. Therefore, in the character recognition process for the license plate in Japan, it is not necessary to determine which image is the first plate or the second plate in step S24.
 そして、日本国のナンバープレートのプレート領域画像について文字認識を行う場合、ステップS21~S23の処理では、下側の文字列52について、文字高さ方向範囲の上端34と下端35とが検出されることになる。この場合、図5Bに図示されているように、ステップS21~S23の処理が行われた後、上段の文字列51について上記のステップS21~S23の処理と同じ処理が行われ、上段の文字列51の、文字高さ方向範囲の上端と下端とが検出されてもよい(ステップS26)。 When character recognition is performed on a plate area image of a license plate in Japan, the upper end 34 and the lower end 35 of the character height direction range are detected for the lower character string 52 in the processing of steps S21 to S23. It will be. In this case, as shown in FIG. 5B, after the processing of steps S21 to S23 is performed, the same processing as the processing of steps S21 to S23 is performed on the upper character string 51, and the upper character string is processed. The upper end and the lower end of 51 in the character height direction range may be detected (step S26).
 上記の文字高さ方向範囲検出処理によって検出された上端、下端の位置を示す上下端データは、外部記憶装置4に保存される。 The upper and lower end data indicating the positions of the upper end and the lower end detected by the above-described character height direction range detection process are stored in the external storage device 4.
3.文字幅方向範囲検出処理
 更に、前処理(ステップS01)によって得られた画像データに対して、文字幅方向範囲検出処理が行われる(ステップS03)。文字幅方向範囲検出処理とは、各文字が存在すると考えられる水平方向の範囲(以下、「文字幅方向範囲」という。)の、水平方向における始点の候補(始点候補)及び終点の候補(終点の候補)を検出する処理である。なお、文字幅方向範囲検出処理で検出される始点候補は、あくまで「候補」であり、各文字が存在する領域の、水平方向における始点として最終的に決定されるわけではないことに留意されたい。同様に、終点候補は、あくまで「候補」であり、各文字が存在する領域の、水平方向における終点として最終的に決定されるわけではない。
3. Character Width Direction Range Detection Processing Further, character width direction range detection processing is performed on the image data obtained by the preprocessing (step S01) (step S03). The character width direction range detection process is a horizontal start point candidate (start point candidate) and end point candidate (end point) of a horizontal range in which each character is considered to exist (hereinafter referred to as “character width direction range”). This is a process for detecting a candidate. It should be noted that the starting point candidates detected by the character width direction range detection process are only “candidates” and are not finally determined as the starting point in the horizontal direction of the area where each character exists. . Similarly, the end point candidate is merely a “candidate”, and is not finally determined as the end point in the horizontal direction of the region where each character exists.
 図6は、文字幅方向範囲検出処理において行われる処理を示すフローチャートであり、また、図7は、文字幅方向範囲検出処理を概念的に説明する図である。まず、x軸射影ヒストグラムが生成される(ステップS31)。ここで、x軸射影ヒストグラムとは、画像の水平方向の各位置について算出された、垂直方向に並んだ画素の輝度値の和の分布である。図7を参照して、座標(x,y)にある画素の輝度をB(x,y)とした場合、x軸射影ヒストグラムは、次式(2)で定義されるSΣ2(x)として得られる:
 SΣ2(x)=ΣB(x、j)    ・・・(2)
ここで、Σは、x軸座標がxである全画素についての和である。図7において、グラフ41は、x軸射影ヒストグラムSΣ2(x)を表わしている。
FIG. 6 is a flowchart showing processing performed in the character width direction range detection processing, and FIG. 7 conceptually illustrates the character width direction range detection processing. First, an x-axis projection histogram is generated (step S31). Here, the x-axis projection histogram is a distribution of the sum of luminance values of pixels arranged in the vertical direction, calculated for each position in the horizontal direction of the image. Referring to FIG. 7, when the luminance of a pixel at coordinates (x, y) is B (x, y), the x-axis projection histogram is expressed as S Σ2 (x) defined by the following equation (2). can get:
S Σ2 (x) = ΣB (x, j) (2)
Here, Σ is the sum for all pixels whose x-axis coordinates are x. In FIG. 7, a graph 41 represents an x-axis projection histogram S Σ2 (x).
 更に、x軸射影ヒストグラムSΣ2(x)に対して平滑化フィルタによる処理(平滑化処理)が行われて平滑化データが算出される(ステップS32)。平滑化フィルタのフィルタサイズは、1文字の幅に相当するサイズに選ばれる。本実施形態では、平滑化処理として移動平均を算出する処理が用いられる。即ち、平滑化データとして、x軸射影ヒストグラムSΣ2(x)の移動平均S^Σ2(x)が算出される。図7において、グラフ42は、移動平均S^Σ2(x)を表わしている。 Further, the smoothing data is calculated by performing processing (smoothing processing) by the smoothing filter on the x-axis projection histogram S Σ2 (x) (step S32). The filter size of the smoothing filter is selected as a size corresponding to the width of one character. In the present embodiment, a process for calculating a moving average is used as the smoothing process. That is, as the smoothed data, moving average S ^ .SIGMA.2 the x-axis projection histogram S Σ2 (x) (x) is calculated. In FIG. 7, a graph 42 represents the moving average S ^ Σ2 (x).
 更に、x軸射影ヒストグラムSΣ2(x)と、その移動平均S^Σ2(x)とを用いて、各文字の文字幅方向範囲の始点候補及び終点候補が検出される(ステップS24)。本実施形態では、始点候補及び終点候補とは、画像の一方の端、より具体的には、左端から探索して検出される。即ち、始点候補は、各文字の文字幅方向範囲の始点の候補、終点候補は、文字幅方向範囲の終点の候補として検出される。始点候補及び終点候補の検出においては、各x座標について移動平均とx軸射影ヒストグラムとの差が算出され、始点候補及び終点候補は、移動平均とx軸射影ヒストグラムとの差の符号が逆転する位置として検出される。例えば、移動平均からx軸射影ヒストグラムを減じた差S^Σ2(x)-SΣ2(x)が各x座標について算出される場合、始点候補は、差S^Σ2(x)-SΣ2(x)の符号がプラスからマイナスに反転するx座標として算出され、終点候補は、差S^Σ2(x)-SΣ2(x)の符号がマイナスからプラスに反転するx座標として算出される。図7において、検出された始点候補は、記号“○”によって示されており、検出された終点候補は、記号“□”によって示されている。 Further, using the x-axis projection histogram S Σ2 (x) and the moving average S ^ Σ2 (x), the start point candidate and the end point candidate of the character width direction range of each character are detected (step S24). In the present embodiment, the start point candidate and the end point candidate are detected by searching from one end of the image, more specifically from the left end. That is, the start point candidate is detected as the start point candidate of the character width direction range of each character, and the end point candidate is detected as the end point candidate of the character width direction range. In the detection of the start point candidate and the end point candidate, the difference between the moving average and the x-axis projection histogram is calculated for each x coordinate, and the sign of the difference between the moving average and the x-axis projection histogram is reversed for the start point candidate and the end point candidate. It is detected as a position. For example, if the difference S ^ .SIGMA.2 minus x-axis projection histogram from the moving average (x) -S Σ2 (x) is calculated for each x coordinate, the starting point candidates, the difference S ^ Σ2 (x) -S Σ2 ( x) is calculated as an x coordinate in which the sign of “+” is inverted from plus to minus, and the end point candidate is calculated as an x coordinate in which the sign of the difference S Σ2 (x) −S Σ2 (x) is inverted from minus to plus. In FIG. 7, the detected start point candidate is indicated by a symbol “◯”, and the detected end point candidate is indicated by a symbol “□”.
 以上の処理により、水平方向に並ぶ文字列の一つについて、始点候補及び終点候補を検出する処理が行われたことになる。プレート領域画像が2段プレートの画像である場合、始点候補及び終点候補の検出が行われていない他方の文字列について、上記のステップS31~S33の処理が行われる(ステップS34)。検出された始点候補及び終点候補の位置は、始終点候補データとして、外部記憶装置4に保存される。 Through the above processing, the processing for detecting the start point candidate and the end point candidate is performed for one of the character strings arranged in the horizontal direction. When the plate region image is a two-stage plate image, the above-described processing of steps S31 to S33 is performed for the other character string for which the start point candidate and the end point candidate are not detected (step S34). The positions of the detected start point candidates and end point candidates are stored in the external storage device 4 as start / end point candidate data.
4.文字中心候補位置領域設定処理
 文字中心候補位置領域設定処理(ステップS04)では、上述の文字高さ方向範囲検出処理によって検出された文字高さ方向範囲の上端及び下端の位置、及び、文字幅方向範囲検出処理によって検出された始点候補及び終点候補のデータに基づいて、各文字の文字中心候補領域を決定する。文字中心候補領域とは、各文字の中心が存在する位置の候補となる領域である。後で行われる文字領域、文字認識結果確定処理(ステップS05)においては、文字中心候補領域の内部のいずれかの位置が、各文字の中心が存在する位置として決定される。
4). Character center candidate position area setting process In the character center candidate position area setting process (step S04), the positions of the upper and lower ends of the character height direction range detected by the above-described character height direction range detection process, and the character width direction Based on the data of the start point candidate and the end point candidate detected by the range detection process, the character center candidate region of each character is determined. The character center candidate region is a region that is a candidate for a position where the center of each character exists. In the character area / character recognition result determination process (step S05) performed later, any position within the character center candidate area is determined as the position where the center of each character exists.
 文字中心候補領域の文字高さ方向(即ち、y軸方向)の範囲は、文字高さ方向範囲検出処理で検出された文字高さ方向範囲の中心位置を含む領域として決定される(ステップS41)。文字中心候補領域の文字高さ方向の範囲は、水平方向に並ぶ文字列に含まれる全ての文字について共通である。一例としては、文字高さ方向範囲検出処理で検出された上端のy座標yMAX、下端のy座標yMIN、及び、所定値mを用いて、(yMIN+yMAX)/2-mから(yMIN+yMAX)/2+mまでの範囲と決定される。この場合、各文字の中心の垂直方向における位置の候補に、(2m)画素の自由度があることになる。各文字の中心が存在する位置の候補に一定程度の自由度が与えられるのは、プレート領域画像にはナンバープレートが多少斜めに撮影されることが想定されるためである。 The range in the character height direction (that is, the y-axis direction) of the character center candidate region is determined as a region including the center position of the character height direction range detected by the character height direction range detection process (step S41). . The range in the character height direction of the character center candidate region is common to all characters included in the character string arranged in the horizontal direction. As an example, using the y coordinate y MAX at the upper end, the y coordinate y MIN at the lower end, and the predetermined value m detected by the character height direction range detection process, (y MIN + y MAX ) / 2−m y MIN + y MAX ) / 2 + m. In this case, the position candidate in the vertical direction of the center of each character has a degree of freedom of (2m) pixels. The reason why a certain degree of freedom is given to the candidate of the position where the center of each character exists is that it is assumed that the license plate is photographed somewhat obliquely in the plate region image.
 更に、各文字のそれぞれについて、文字中心候補領域の水平方向(即ち、x軸方向)の範囲が決定される(ステップS42)。各文字のそれぞれについての文字中心候補領域の水平方向の範囲の決定では、まず、上述の文字幅方向範囲検出処理で検出された始点候補及び終点候補から、各文字に対応する文字幅方向範囲の始点及び終点のペアが選択される。ここで、文字幅方向範囲の始点は、文字幅方向範囲検出処理で検出された始点候補から選択され、文字幅方向範囲の終点は、終点候補から選択される。左側からの探索によって始点候補及び終点候補が検出される本実施形態では、基本的には、左からi番目にある文字については、左からi番目にある始点候補を始点、左からi番目にある終点候補を終点と決定すればよい。ただし、文字の幅には、妥当と考えられる範囲があり、その範囲から外れるような始点及び終点が決定される場合には、隣接する始点候補又は終点候補を代わりに用いて始点及び終点のペアが決定される。このような手順により、各文字の文字幅方向範囲の始点及び終点のペアが決定される。 Further, for each character, the range in the horizontal direction (that is, the x-axis direction) of the character center candidate region is determined (step S42). In the determination of the horizontal range of the character center candidate region for each character, first, the character width direction range corresponding to each character is determined from the start point candidate and the end point candidate detected by the character width direction range detection process described above. A start and end pair is selected. Here, the start point of the character width direction range is selected from the start point candidates detected by the character width direction range detection process, and the end point of the character width direction range is selected from the end point candidates. In the present embodiment in which the start point candidate and the end point candidate are detected by the search from the left side, basically, for the i th character from the left, the i th start point candidate from the left is the start point, and the i th from the left A certain end point candidate may be determined as the end point. However, there is a range that is considered to be appropriate for the width of the character, and when a start point and an end point that are outside the range are determined, a pair of start point and end point is used by using adjacent start point candidates or end point candidates instead. Is determined. By such a procedure, the pair of the start point and end point of the character width direction range of each character is determined.
 文字中心候補領域の水平方向(即ち、x軸方向)の範囲は、各文字について決定された文字幅方向範囲の中心を含む領域として決定される。本実施形態では、各文字に対応する始点のx座標xMIN及び終点のx座標xMAX、及び、所定値kを用いて、(xMIN+xMAX)/2-kから(xMIN+xMAX)/2+kまでの範囲と決定される。この場合、各文字の中心の水平方向における位置の候補に、(2k)画素の自由度があることになる。これは、プレート領域画像に映る文字の幅に多少の変動があることを考慮したものである。 The range in the horizontal direction (that is, the x-axis direction) of the character center candidate region is determined as a region including the center of the character width direction range determined for each character. In the present embodiment, from (x MIN + x MAX ) / 2−k to (x MIN + x MAX ) using the x coordinate x MIN of the start point and the x coordinate x MAX of the end point corresponding to each character and the predetermined value k. It is determined as a range up to / 2 + k. In this case, the position candidate in the horizontal direction of the center of each character has a degree of freedom of (2k) pixels. This is due to the fact that there is some variation in the width of the characters shown in the plate area image.
 以上の処理により、水平方向に並ぶ文字列の一つについて、各文字の文字中心候補領域を決定する処理が行われたことになる。プレート領域画像が、2段プレートの画像である場合、文字中心候補領域を決定する処理が行われていない他方の文字列について、上記のステップS41~S42の処理が行われる(ステップS43)。決定された文字中心候補領域の範囲を示すデータ(文字中心候補領域データ)は、外部記憶装置4に保存される。 With the above processing, the processing for determining the character center candidate region for each character is performed for one of the character strings arranged in the horizontal direction. When the plate region image is a two-stage plate image, the above-described processing of steps S41 to S42 is performed for the other character string that has not been subjected to the processing for determining the character center candidate region (step S43). Data indicating the range of the determined character center candidate region (character center candidate region data) is stored in the external storage device 4.
5.文字領域・文字認識結果確定処理
 続いて、各文字が存在する領域、及び、文字認識結果を確定する処理が行われる(ステップS05)。
5. Character Area / Character Recognition Result Determination Process Subsequently, a process for determining the area where each character exists and the character recognition result is performed (step S05).
 図10は、文字領域・文字認識結果確定処理において行われる処理を示すフローチャートであり、また、図11、図12は、文字幅方向範囲検出処理を概念的に説明する図である。 FIG. 10 is a flowchart showing processing performed in the character area / character recognition result determination processing, and FIGS. 11 and 12 are diagrams conceptually explaining the character width direction range detection processing.
 詳細には、図10に図示されているように、まず、最初に処理すべき文字(本実施形態では、最も左に位置する文字)に対応する文字中心候補領域が選択される(ステップS50)。 Specifically, as shown in FIG. 10, first, a character center candidate region corresponding to the character to be processed first (in this embodiment, the character located at the leftmost) is selected (step S50). .
 更に、選択された文字中心候補領域の内部の各点(各画素)に対し、高さ及び/又は幅が異なる複数の文字候補領域が設定される(ステップS51)。例えば、文字中心候補領域が3×3の9画素である場合、その9画素の各画素に対応する9つの文字候補領域が設定される。本実施形態では、文字候補領域は矩形である(即ち、(正方形を含む)長方形である)。図11は、文字中心候補領域の内部の特定の点61と、該特定の点について設定される文字候補領域62の関係を示す図である。文字中心候補領域の内部の特定の点61について設定される文字候補領域62は、当該文字候補領域62の対角線63の交点が、該特定の点61に一致するように決定される。図12に図示されているように、文字候補領域には、基準サイズ(基準の高さ及び幅)が定められており、各文字候補領域の高さ及び幅は、基準サイズと同一の高さ及び幅、又は、基準サイズから増減された高さ及び幅に決定される。高さ及び幅の最大の増減量は、予め、パラメータとして設定される。 Further, a plurality of character candidate areas having different heights and / or widths are set for each point (each pixel) inside the selected character center candidate area (step S51). For example, when the character center candidate region is 9 pixels of 3 × 3, nine character candidate regions corresponding to each pixel of the 9 pixels are set. In this embodiment, the character candidate area is a rectangle (that is, a rectangle (including a square)). FIG. 11 is a diagram illustrating a relationship between a specific point 61 inside the character center candidate region and the character candidate region 62 set for the specific point. The character candidate region 62 set for a specific point 61 inside the character center candidate region is determined so that the intersection of the diagonal lines 63 of the character candidate region 62 coincides with the specific point 61. As shown in FIG. 12, the character candidate area has a reference size (reference height and width), and each character candidate area has the same height and width as the reference size. And the width or the height and width increased or decreased from the reference size. The maximum increase / decrease amount of the height and width is set in advance as a parameter.
 続いて、各文字候補領域内の画像に対して文字認識処理が行われる(ステップS52)。本実施形態では、文字認識処理が、テンプレートを用いたパターンマッチング(テンプレートマッチング法)によって行われ、文字認識処理においては、文字認識結果と、その評価値とが決定される。評価値としては、例えば、テンプレートマッチング法において算出される類似度を用いてもよい。 Subsequently, a character recognition process is performed on the image in each character candidate area (step S52). In this embodiment, character recognition processing is performed by pattern matching using a template (template matching method), and in the character recognition processing, a character recognition result and its evaluation value are determined. As the evaluation value, for example, the similarity calculated in the template matching method may be used.
 更に、評価値が最も高い文字候補領域と、その文字候補領域の画像の文字認識結果とが抽出される(ステップS53)。例えば、文字中心候補領域が9画素である場合、9つの文字候補領域のうち、評価値が最も高い文字候補領域と、その文字候補領域の画像の文字認識結果とが抽出される。 Further, the character candidate area having the highest evaluation value and the character recognition result of the image of the character candidate area are extracted (step S53). For example, when the character center candidate region is 9 pixels, the character candidate region having the highest evaluation value among the nine character candidate regions and the character recognition result of the image of the character candidate region are extracted.
 評価値が、所定の基準値以上である場合(ステップS54:Yes)、評価値が最も高い文字候補領域、及び、その文字認識結果が、処理対象の文字が存在する文字領域及び文字認識結果として確定される(ステップS55)。そうでない場合(ステップS54:No)、文字を含まない領域が文字候補領域として与えられたとして、文字認識結果は棄却される。 When the evaluation value is greater than or equal to a predetermined reference value (step S54: Yes), the character candidate area with the highest evaluation value and the character recognition result are the character area and character recognition result in which the character to be processed exists. Confirmed (step S55). Otherwise (step S54: No), the character recognition result is rejected assuming that a region not including characters is given as a character candidate region.
 ステップS51~S55の処理は、それらが全ての文字中心候補領域について実行されるまで繰り返して行われる(ステップS56)。ステップS51~S55の処理が行われていない文字中心候補領域が存在する場合、次文字に対応する文字中心候補領域が選択され(ステップS57)、再度、ステップS51~S55の処理が行われる。全ての文字中心候補領域についてステップS51~S55の処理が行われると、全ての文字について、各文字が存在する文字領域及び文字認識結果が確定したことになる(ステップS58)。確定した文字領域及び文字認識結果を示すデータが、プレート認識データ22として、外部記憶装置4に保存される。 The processing in steps S51 to S55 is repeated until they are executed for all character center candidate regions (step S56). If there is a character center candidate area that has not been subjected to the processes of steps S51 to S55, the character center candidate area corresponding to the next character is selected (step S57), and the processes of steps S51 to S55 are performed again. When the processing of steps S51 to S55 is performed for all the character center candidate regions, the character region in which each character exists and the character recognition result are determined for all characters (step S58). Data indicating the confirmed character region and character recognition result is stored in the external storage device 4 as plate recognition data 22.
 以上に述べられているように、本実施形態の文字認識処理においては、y軸射影ヒストグラム及びx軸射影ヒストグラムに基づいて各文字について文字中心候補領域が決定され、その文字中心候補領域の内部の各点について文字候補領域が決定される。本実施形態では、各文字の中心にある程度の自由度が与えられた状態で文字候補領域が設定されることになり、誤認識を抑制することができる。その一方で、文字中心候補領域は、文字高さ方向範囲検出処理による文字高さ方向範囲の検出結果、及び、文字幅方向範囲検出処理による文字幅方向範囲の検出結果によってある程度制限されるので、設定する文字候補領域の数が、一定程度抑制される。このため、文字認識処理のデータ処理量を抑制し、処理時間を短縮することができる。 As described above, in the character recognition processing of the present embodiment, a character center candidate region is determined for each character based on the y-axis projection histogram and the x-axis projection histogram, and the character center candidate region inside the character center candidate region is determined. A character candidate area is determined for each point. In this embodiment, a character candidate area is set in a state where a certain degree of freedom is given to the center of each character, and erroneous recognition can be suppressed. On the other hand, the character center candidate area is limited to some extent by the detection result of the character height direction range by the character height direction range detection process and the detection result of the character width direction range by the character width direction range detection process. The number of character candidate areas to be set is suppressed to a certain extent. For this reason, the data processing amount of character recognition processing can be suppressed and processing time can be shortened.
 以上には、本発明の実施形態が具体的に述べられているが、本発明は上記の実施形態には限定されない。本発明が、様々な変更と共に実施可能であることは、当業者には自明的であろう。例えば、上述では、本発明がナンバープレート認識(ナンバープレートに記載された文字の認識)に適用された実施形態が記載されているが、本発明は、グレースケール画像である撮像画像に対する文字認識一般に適用可能である。 The embodiment of the present invention has been specifically described above, but the present invention is not limited to the above embodiment. It will be apparent to those skilled in the art that the present invention can be implemented with various modifications. For example, in the above description, an embodiment in which the present invention is applied to license plate recognition (recognition of characters written on a license plate) is described. However, the present invention generally recognizes characters for a captured image that is a grayscale image. Applicable.

Claims (7)

  1.  グレースケール画像であり、且つ、文字幅方向に並んだ複数の文字を含む文字列が映された画像である対象画像について、前記文字列が存在する、前記文字幅方向と垂直な文字高さ方向の範囲である文字高さ方向範囲を検出する文字高さ方向範囲検出手段と、
     前記対象画像について、前記文字列の各文字が存在する前記文字幅方向の範囲である文字幅方向範囲を検出する文字幅方向範囲検出手段と、
     前記対象画像について、前記文字高さ方向における前記文字高さ方向範囲の中心を含み、前記文字幅方向における前記文字幅方向範囲の中心を含む領域として、前記文字列の各文字について文字中心候補領域を決定する文字中心候補領域決定手段と、
     前記文字中心候補領域の各点について、対角線の交点が前記各点に一致する複数の矩形の文字候補領域を設定する文字候補領域設定手段と、
     前記対象画像の前記複数の文字候補領域の部分のそれぞれに対して文字認識処理を行って、前記文字列の各文字について、前記複数の文字候補領域のそれぞれの文字認識結果を得る文字認識手段と、
     前記複数の文字候補領域それぞれの文字認識結果から、前記文字列の各文字についての文字識別結果を確定する文字識別結果確定手段
    とを具備する
     文字認識装置。
    A character height direction perpendicular to the character width direction in which the character string exists for a target image that is a gray scale image and an image in which a character string including a plurality of characters arranged in the character width direction is displayed A character height direction range detecting means for detecting a character height direction range that is a range of
    Character width direction range detection means for detecting a character width direction range that is a range in the character width direction in which each character of the character string exists for the target image;
    About the target image, a character center candidate region for each character of the character string as a region including the center of the character height direction range in the character height direction and including the center of the character width direction range in the character width direction A character center candidate region determining means for determining
    For each point of the character center candidate region, a character candidate region setting means for setting a plurality of rectangular character candidate regions in which the intersections of diagonal lines match the points,
    Character recognition means for performing character recognition processing on each of the plurality of character candidate area portions of the target image and obtaining each character recognition result of the plurality of character candidate areas for each character of the character string; ,
    A character recognition apparatus comprising: a character identification result determining unit that determines a character identification result for each character of the character string from a character recognition result of each of the plurality of character candidate regions.
  2.  請求項1に記載の文字認識装置であって、
     前記文字高さ方向範囲検出手段は、前記対象画像の前記文字高さ方向の各位置について算出された、前記文字幅方向に並んだ画素の輝度値の和の分布である第1射影ヒストグラムを生成し、前記第1射影ヒストグラムに対して平滑化処理を行って第1平滑化データを算出し、前記第1射影ヒストグラムと前記平滑化データとの差の符号が反転する位置として前記文字高さ方向範囲の上端と下端とを検出する
     文字認識装置。
    The character recognition device according to claim 1,
    The character height direction range detection means generates a first projection histogram that is a distribution of the sum of luminance values of pixels arranged in the character width direction, calculated for each position in the character height direction of the target image. Then, smoothing processing is performed on the first projection histogram to calculate first smoothed data, and the character height direction is set as a position where the sign of the difference between the first projection histogram and the smoothed data is inverted. A character recognition device that detects the top and bottom edges of a range.
  3.  請求項1又は2に記載の文字認識装置であって、
     前記文字幅方向範囲検出手段は、前記対象画像の前記文字幅方向の各位置について算出された、前記文字高さ方向に並んだ画素の輝度値の和の分布である第2射影ヒストグラムを生成し、前記第2射影ヒストグラムに対して平滑化処理を行って第2平滑化データを算出し、前記第2射影ヒストグラムと前記第2平滑化データとの差の符号が反転する位置から始点候補と終点候補とを検出し、前記始点候補のうちから前記文字列の各文字の前記文字幅方向範囲の始点を選択し、前記終点候補のうちから前記文字列の各文字の前記文字幅方向範囲の終点を選択する
     文字認識装置。
    The character recognition device according to claim 1 or 2,
    The character width direction range detecting means generates a second projection histogram that is a distribution of the sum of luminance values of pixels arranged in the character height direction, calculated for each position in the character width direction of the target image. Then, smoothing processing is performed on the second projection histogram to calculate second smoothed data, and a start point candidate and an end point are determined from the position where the sign of the difference between the second projection histogram and the second smoothed data is inverted. A candidate is detected, the start point of the character width direction range of each character of the character string is selected from the start point candidates, and the end point of the character width direction range of each character of the character string is selected from the end point candidates Select a character recognition device.
  4.  請求項1乃至3のいずれかに記載の文字認識装置であって、
     前記対象画像が、ナンバープレートが映されたプレート領域画像、又は、前記プレート領域画像に対して画像処理を行うことで得られる画像である
     文字認識装置。
    A character recognition device according to any one of claims 1 to 3,
    The character recognition device, wherein the target image is a plate area image showing a license plate or an image obtained by performing image processing on the plate area image.
  5.  請求項1乃至3のいずれかに記載の文字認識装置であって、
     前記対象画像が、ナンバープレートが映されたプレート領域画像に対して前処理を行うことで得られる画像であり、
     前記前処理では、前記プレート領域画像に対して、高輝度になっている部分の輝度を抑制すると共に、画素の輝度の平均値と標準偏差とが、それぞれ、所定の値になるように線形変換を行う輝度補正処理が行われ、前記輝度補正処理によって得られた画像に対して微分フィルタによる処理が行われてエッジ画像が生成され、前記エッジ画像に含まれる文字輪郭線の内側を穴埋めする穴埋め処理が行われ、前記穴埋め処理で得られた画像に対して前記文字幅方向の低周波成分を除去する処理が行われる
     文字認識装置。
    A character recognition device according to any one of claims 1 to 3,
    The target image is an image obtained by performing preprocessing on a plate region image on which a license plate is projected,
    In the pre-processing, the plate region image is subjected to linear conversion so as to suppress the luminance of the portion where the luminance is high, and the average value and the standard deviation of the pixel luminance are respectively predetermined values. The edge correction is performed on the image obtained by the brightness correction process, an edge image is generated, and the inside of the character outline included in the edge image is filled. A character recognition device in which processing is performed and processing for removing low-frequency components in the character width direction is performed on an image obtained by the hole filling processing.
  6.  グレースケール画像であり、且つ、文字幅方向に並んだ複数の文字を含む文字列が映された画像である対象画像について、前記文字列が存在する、前記文字幅方向と垂直な文字高さ方向の範囲である文字高さ方向範囲を検出するステップと、
     前記対象画像について、前記文字列の各文字が存在する前記文字幅方向の範囲である文字幅方向範囲を検出するステップと、
     前記対象画像について、前記文字高さ方向における前記文字高さ方向範囲の中心を含み、前記文字幅方向における前記文字幅方向範囲の中心を含む領域として、前記文字列の各文字について文字中心候補領域を決定するステップと、
     前記文字中心候補領域の各点について、対角線の交点が前記各点に一致する複数の矩形の文字候補領域を設定するステップと、
     前記対象画像の前記複数の文字候補領域の部分のそれぞれに対して文字認識処理を行って、前記文字列の各文字について、前記複数の文字候補領域のそれぞれの文字認識結果を得るステップと、
     前記複数の文字候補領域それぞれの文字認識結果から、前記文字列の各文字についての文字識別結果を確定するステップ
    とを具備する
     文字認識方法。
    A character height direction perpendicular to the character width direction in which the character string exists for a target image that is a gray scale image and an image in which a character string including a plurality of characters arranged in the character width direction is displayed Detecting a range in the character height direction that is a range of
    For the target image, detecting a character width direction range that is a range in the character width direction in which each character of the character string exists;
    About the target image, a character center candidate region for each character of the character string as a region including the center of the character height direction range in the character height direction and including the center of the character width direction range in the character width direction A step of determining
    For each point of the character center candidate region, setting a plurality of rectangular character candidate regions whose intersections of diagonal lines match the points;
    Performing character recognition processing on each of the plurality of character candidate region portions of the target image to obtain each character recognition result of the plurality of character candidate regions for each character of the character string;
    Determining a character identification result for each character of the character string from a character recognition result for each of the plurality of character candidate regions.
  7.  プログラムを記憶する記録媒体であって、前記プログラムは、実行されたときに下記ステップ群:
     グレースケール画像であり、且つ、文字幅方向に並んだ複数の文字を含む文字列が映された画像である対象画像について、前記文字列が存在する、前記文字幅方向と垂直な文字高さ方向の範囲である文字高さ方向範囲を検出するステップ、
     前記対象画像について、前記文字列の各文字が存在する前記文字幅方向の範囲である文字幅方向範囲を検出するステップ、
     前記対象画像について、前記文字高さ方向における前記文字高さ方向範囲の中心を含み、前記文字幅方向における前記文字幅方向範囲の中心を含む領域として、前記文字列の各文字について文字中心候補領域を決定するステップ、
     前記文字中心候補領域の各位置について、複数の文字候補領域を設定するステップ、
     前記対象画像の前記複数の文字候補領域のそれぞれに対して文字認識処理を行って、前記複数の文字候補領域のそれぞれの文字認識結果を得るステップ、及び、
     前記複数の文字候補領域それぞれの文字認識結果から、前記文字列の各文字についての文字識別結果を確定するステップ
    を演算装置に実行させる
     記録媒体。
    A recording medium for storing a program, wherein when the program is executed, the following steps are performed:
    A character height direction perpendicular to the character width direction in which the character string exists for a target image that is a gray scale image and an image in which a character string including a plurality of characters arranged in the character width direction is displayed Detecting a range in the height direction of a character that is a range of
    Detecting a character width direction range that is a range in the character width direction in which each character of the character string exists for the target image;
    About the target image, a character center candidate region for each character of the character string as a region including the center of the character height direction range in the character height direction and including the center of the character width direction range in the character width direction Step to determine,
    Setting a plurality of character candidate regions for each position of the character center candidate region;
    Performing character recognition processing on each of the plurality of character candidate regions of the target image to obtain each character recognition result of the plurality of character candidate regions; and
    A recording medium that causes a computing device to execute a step of determining a character identification result for each character of the character string from character recognition results of each of the plurality of character candidate regions.
PCT/JP2013/079265 2013-02-25 2013-10-29 Character recognition device, character recognition method, and recording medium WO2014129018A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
SG11201506568XA SG11201506568XA (en) 2013-02-25 2013-10-29 Character recognition apparatus, character recognition method and recording medium
MYPI2015702775A MY177406A (en) 2013-02-25 2013-10-29 Character recognition apparatus, character recognition method and program

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2013-035143 2013-02-25
JP2013035143A JP6177541B2 (en) 2013-02-25 2013-02-25 Character recognition device, character recognition method and program

Publications (1)

Publication Number Publication Date
WO2014129018A1 true WO2014129018A1 (en) 2014-08-28

Family

ID=51390835

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2013/079265 WO2014129018A1 (en) 2013-02-25 2013-10-29 Character recognition device, character recognition method, and recording medium

Country Status (4)

Country Link
JP (1) JP6177541B2 (en)
MY (1) MY177406A (en)
SG (1) SG11201506568XA (en)
WO (1) WO2014129018A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107480648A (en) * 2017-08-23 2017-12-15 南京大学 A kind of method of natural scene text detection
CN108399405A (en) * 2017-02-07 2018-08-14 腾讯科技(上海)有限公司 Business license recognition methods and device
CN113807351A (en) * 2021-09-18 2021-12-17 京东鲲鹏(江苏)科技有限公司 Scene character detection method and device

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6550723B2 (en) * 2014-10-31 2019-07-31 オムロン株式会社 Image processing apparatus, character recognition apparatus, image processing method, and program
CN105185121B (en) * 2015-09-11 2018-03-02 安徽四创电子股份有限公司 A kind of method of virtual bayonet socket parallelism recognition car plate
JP7075770B2 (en) * 2018-02-07 2022-05-26 グローリー株式会社 Character recognition system, character sharpening system, character sharpening program, character sharpening method, and character sharpening image display device
JP2019139383A (en) * 2018-02-07 2019-08-22 グローリー株式会社 Character recognition system, character recognition program, character recognition method, character sharpening system, character sharpening program, and character sharpening method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH087041A (en) * 1994-06-16 1996-01-12 Hitachi Ltd Character recognition method and device
JPH08241374A (en) * 1995-03-02 1996-09-17 Canon Inc Device and method for reading character
JPH08305795A (en) * 1995-04-28 1996-11-22 Nippon Steel Corp Character recognizing method
JP2010004152A (en) * 2008-06-18 2010-01-07 Konica Minolta Business Technologies Inc Image processor and character region extraction method, and computer program

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0528321A (en) * 1991-07-19 1993-02-05 Hitachi Ltd Method and system for image recognition
JPH11306283A (en) * 1998-04-24 1999-11-05 Chuo Spring Co Ltd Number plate reader
JP4224328B2 (en) * 2003-03-18 2009-02-12 株式会社日立製作所 Car number recognition device
JP2009250739A (en) * 2008-04-04 2009-10-29 Nikon Corp Method for processing image and program
JP5709410B2 (en) * 2009-06-16 2015-04-30 キヤノン株式会社 Pattern processing apparatus and method, and program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH087041A (en) * 1994-06-16 1996-01-12 Hitachi Ltd Character recognition method and device
JPH08241374A (en) * 1995-03-02 1996-09-17 Canon Inc Device and method for reading character
JPH08305795A (en) * 1995-04-28 1996-11-22 Nippon Steel Corp Character recognizing method
JP2010004152A (en) * 2008-06-18 2010-01-07 Konica Minolta Business Technologies Inc Image processor and character region extraction method, and computer program

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SATOSHI SHIMADA ET AL.: "Fast telop character recognition without extracting character area", IEICE TECHNICAL REPORT, vol. 111, no. 317, 17 November 2011 (2011-11-17), pages 57 - 62 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108399405A (en) * 2017-02-07 2018-08-14 腾讯科技(上海)有限公司 Business license recognition methods and device
CN108399405B (en) * 2017-02-07 2023-06-27 腾讯科技(上海)有限公司 Business license identification method and device
CN107480648A (en) * 2017-08-23 2017-12-15 南京大学 A kind of method of natural scene text detection
CN107480648B (en) * 2017-08-23 2020-03-27 南京大学 Method for detecting characters in natural scene
CN113807351A (en) * 2021-09-18 2021-12-17 京东鲲鹏(江苏)科技有限公司 Scene character detection method and device
CN113807351B (en) * 2021-09-18 2024-01-16 京东鲲鹏(江苏)科技有限公司 Scene text detection method and device

Also Published As

Publication number Publication date
JP2014164528A (en) 2014-09-08
SG11201506568XA (en) 2015-09-29
JP6177541B2 (en) 2017-08-09
MY177406A (en) 2020-09-14

Similar Documents

Publication Publication Date Title
JP6177541B2 (en) Character recognition device, character recognition method and program
JP5699788B2 (en) Screen area detection method and system
JP5707423B2 (en) Pattern matching device and computer program
JP4966893B2 (en) Matching degree calculation device and method, program
JP2002133426A (en) Ruled line extracting device for extracting ruled line from multiple image
JP6115214B2 (en) Pattern processing apparatus, pattern processing method, and pattern processing program
KR101461108B1 (en) Recognition device, vehicle model recognition apparatus and method
JP6110174B2 (en) Image detection apparatus, control program, and image detection method
CN112419207A (en) Image correction method, device and system
JP5772675B2 (en) Gray image edge extraction method, edge extraction device, and gray image edge extraction program
JP5100688B2 (en) Object detection apparatus and program
WO2020209046A1 (en) Object detection device
JP2006155126A (en) Vehicle number recognition device
JP2010286995A (en) Image processing system for vehicle
KR101444259B1 (en) Method for generating a compensation matrix during a circuit board inspection
JP6772059B2 (en) Electronic control devices, electronic control systems and electronic control methods
JP2010191767A (en) Device and method for recognizing character
JP2006309402A (en) Character string recognition method, number plate recognition method and device
JP2019100937A (en) Defect inspection device and defect inspection method
JPH11219435A (en) White line detector for automobile
JP2011018175A (en) Character recognition apparatus and character recognition method
JP2014021510A (en) Information processor and information processing method, and, program
JP6173715B2 (en) Character recognition device, character recognition method and program
JP2008204307A (en) Image-collating device
JP2007102702A (en) Method and device for correcting inclination of character string

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13875826

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: IDP00201505168

Country of ref document: ID

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 13875826

Country of ref document: EP

Kind code of ref document: A1