JPH0433080A - Method for recognizing character in table - Google Patents
Method for recognizing character in tableInfo
- Publication number
- JPH0433080A JPH0433080A JP2134876A JP13487690A JPH0433080A JP H0433080 A JPH0433080 A JP H0433080A JP 2134876 A JP2134876 A JP 2134876A JP 13487690 A JP13487690 A JP 13487690A JP H0433080 A JPH0433080 A JP H0433080A
- Authority
- JP
- Japan
- Prior art keywords
- frame
- character
- line
- scanning direction
- main scanning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 13
- 238000000605 extraction Methods 0.000 claims abstract description 20
- 239000000284 extract Substances 0.000 description 8
- 238000010586 diagram Methods 0.000 description 3
- 239000000203 mixture Substances 0.000 description 3
- 238000005520 cutting process Methods 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
Landscapes
- Character Input (AREA)
Abstract
Description
【発明の詳細な説明】
〔産業上の利用分野〕
本発明は、文字認識装置における文書の表内文字認識方
法に関する。DETAILED DESCRIPTION OF THE INVENTION [Field of Industrial Application] The present invention relates to a method for recognizing characters within a document in a character recognition device.
文字認識装置においては5文書画像を文字領域、写真や
図形などのイメージ領域、表領域などに分割し、それぞ
れの領域に別の処理を行うことが多い。In character recognition devices, five document images are often divided into a character area, an image area such as a photograph or figure, a table area, etc., and different processing is performed on each area.
表領域に関しては、表を構成する罫線の位置を認識し、
罫線で囲まれた枠内の画像に対して連結した黒画素の追
跡を行い、黒画素連結の外接矩形を求め、それを統合し
て文字行を抽出し、文字認識している。Regarding the table area, it recognizes the position of the borders that make up the table,
Connected black pixels are traced in an image within a frame surrounded by ruled lines, a circumscribed rectangle of connected black pixels is found, and the circumscribed rectangles are integrated to extract character lines and character recognition is performed.
(発明が解決しようとする課題〕
しかし従来は1表の各枠内の文字が横書きまたは縦書き
のいずれか一方で印字されていることを前提に文字行抽
出処理をしているため、例えば第3図に示すような横書
きの文字行と縦書きの文字行が混在した表の場合、横書
き(縦書き)を前提としているときには縦書き文字行(
横書き文字行)の抽出が正確に行われず、結果として文
字認識が正確に行われないことがある。(Problem to be solved by the invention) However, conventionally, character line extraction processing is performed on the assumption that the characters in each frame of a table are printed either horizontally or vertically. In the case of a table with a mixture of horizontal and vertical text lines as shown in Figure 3, if horizontal text (vertical writing) is assumed, the vertical text lines (
(Horizontal character lines) may not be extracted accurately, and as a result, character recognition may not be performed accurately.
本発明の目的は、表内の横書き文字行も縦書き文字行も
正確に抽出して文字認識することができる表内文字認識
方法を提供することにある。An object of the present invention is to provide a table character recognition method that can accurately extract and recognize both horizontally written character lines and vertically written character lines in a table.
本発明は、文書画像の表領域より、主走査方向及び副走
査方向の線分で囲まれた枠を抽出し、各枠内の文字行を
抽出して文字認識する表内文字認識方法において、各粋
の主走査方向の長さ及び副走査方向の長さによって各枠
内の文字行が横書きであるか縦書きであるかを判別し、
この判別の結果に応じて各枠内の文字行の抽出方法を切
り替えることを特徴とする。The present invention provides a table character recognition method that extracts frames surrounded by line segments in the main scanning direction and sub-scanning direction from a table area of a document image, extracts character lines within each frame, and recognizes the characters. Determine whether the character line in each frame is written horizontally or vertically based on the length in the main scanning direction and the length in the sub-scanning direction of each frame,
The method is characterized in that the method for extracting character lines within each frame is switched depending on the result of this determination.
本発明によれば、表中の各枠内に印字された文字が横書
きであるか縦書きであるかが自動的に判別され5判別さ
れた方向に適した文字行抽出方法が適用されることによ
り、横書きの枠と縦書きの枠が混在した表においても、
各枠内の文字行が正確に抽出され、したがって各枠内の
文字の切り出し及び文字認識の精度が上がる。According to the present invention, it is automatically determined whether the characters printed in each frame in the table are written horizontally or vertically, and a character line extraction method suitable for the determined direction is applied. Therefore, even in tables with a mixture of horizontal and vertical frames,
The character lines within each frame are accurately extracted, thus increasing the accuracy of character extraction and character recognition within each frame.
第1図は本発明の一実施例を示すブロック図。 FIG. 1 is a block diagram showing one embodiment of the present invention.
第2図は処理のフローチャートである。FIG. 2 is a flowchart of the process.
スキャナーなどの2値画像入力部101によって文書を
読取り、その2値画像を2値イメージメモリ102に格
納する(処理ステップ201)。A document is read by a binary image input unit 101 such as a scanner, and the binary image is stored in the binary image memory 102 (processing step 201).
この文書画像に対して1表領域認識部103はランレン
グス分布などを利用して表領域を自動的に認識するか、
あるいはマウスなどを用いて操作者から指定された領域
を表領域として認識し、表領域のイメージを表領域イメ
ージメモリ104に格納する(処理ステップ202)。For this document image, the table area recognition unit 103 automatically recognizes the table area using run length distribution or the like.
Alternatively, an area specified by the operator using a mouse or the like is recognized as a table area, and an image of the table area is stored in the table area image memory 104 (processing step 202).
この表領域のイメージに対し、主走査方向線分抽出部1
05において、主走査方向に連結した黒画素を追跡して
主走査方向の線分を抽出し、その始点及び終点の座標を
主走査方向線分座標メモリ106に格納する(処理ステ
ップ203)。同様に副走査方向線分抽出部107にお
いて、表領域イメージ内の副走査方向に連結した黒画素
を追跡して副走査方向の線分を抽出し、その始点及び終
点の座標を副走査方向線分座標メモリ108に格納する
(処理ステップ204)。For this image of the table area, the line segment extraction unit 1 in the main scanning direction
05, black pixels connected in the main scanning direction are traced to extract a line segment in the main scanning direction, and the coordinates of the starting point and end point are stored in the main scanning direction line segment coordinate memory 106 (processing step 203). Similarly, the sub-scanning direction line segment extraction unit 107 traces the black pixels connected in the sub-scanning direction in the table area image, extracts a line segment in the sub-scanning direction, and calculates the coordinates of the starting point and end point along the sub-scanning direction line. The coordinates are stored in the minute coordinate memory 108 (processing step 204).
次に枠認識部109において、各メモリ106゜108
に格納された線分座標を参照し、主走査方向線分と副走
査方向線分の組合せにより表の枠を認識し、枠の座標例
えば対角頂点の座標を枠座標メモリ110に格納する(
処理ステップ205)。Next, in the frame recognition unit 109, each memory 106° 108
The frame of the table is recognized by the combination of the line segment in the main scanning direction and the line segment in the sub-scanning direction, and the coordinates of the frame, for example, the coordinates of the diagonal vertices, are stored in the frame coordinate memory 110 (
Process step 205).
また枠領域抽出部111において、枠座標メモリ115
内の枠座標を参照することにより、表領域イメージメモ
リ104より枠の領域の画像を抽出して枠領域画像メモ
リ112に格納する(処理ステップ206)。In addition, in the frame area extraction unit 111, the frame coordinate memory 115
By referring to the frame coordinates in , the image of the frame area is extracted from the table area image memory 104 and stored in the frame area image memory 112 (processing step 206).
次に行方向判定部113において、枠座標メモリ110
を参照して全ての枠に対して主走査方向及び副走査方向
の長さのヒストグラムを作成する(処理ステップ207
,208)、そして、&大度数の副走査方向の長さを持
つ枠は全て行方向が横書きの枠であると判別しく処理ス
テップ209゜210)、その長さと同じ主走査方向の
長さを持つ枠は行方向が縦書きの枠であると判別しく処
理ステップ211,212)、残った枠はそれまでに判
別された枠数が多いほうの行方向の枠であると判別する
(処理ステップ213)、なお、処理ステップ207,
208でヒストグラムを求める際には各走査方向の長さ
にある程度の幅を持たせ。Next, in the row direction determination unit 113, the frame coordinate memory 110
A histogram of lengths in the main scanning direction and the sub-scanning direction is created for all frames with reference to (processing step 207
, 208), and all frames with a length in the sub-scanning direction of & large degrees are determined to be frames with horizontal writing in the line direction. The remaining frames are determined to be frames with vertical writing in the line direction (processing steps 211 and 212), and the remaining frames are determined to be vertical frames with the largest number of frames determined up to that point (processing steps 211 and 212). 213), and processing steps 207,
When calculating the histogram in step 208, a certain amount of width is given to the length in each scanning direction.
同様に処理ステップ211で長さを判別する際にも、比
較する長さの差がある幅の範囲内のときは一致すると判
定する。求められた行方向の情報は外接矩形抽出部11
4を経由して行画像抽出部116へ伝えられる。Similarly, when determining the length in processing step 211, if the difference in length to be compared is within a certain width range, it is determined that they match. The obtained row direction information is sent to the circumscribed rectangle extraction unit 11
4 to the row image extraction unit 116.
例えば第3図に示した表の場合、横書きの枠の副走査方
向の長さは全て同一(あるいは、はぼ同一)であるので
、その頻度は最大である。したがって、この表の横書き
の枠はすべて処理ステップ210で横書きと判別される
。また、この表の縦書き文字列″データ″が印刷された
枠の主走査方向の長さは、最大頻度の副走査方向の長さ
とほぼ同一である(差が一定の幅の範囲である)ので。For example, in the case of the table shown in FIG. 3, the lengths of the horizontally written frames in the sub-scanning direction are all the same (or almost the same), so the frequency is maximum. Therefore, all horizontal writing frames in this table are determined to be horizontal writing in processing step 210. Also, the length of the frame in which the vertically written character string "data" in this table is printed in the main scanning direction is almost the same as the length of the maximum frequency in the sub-scanning direction (the difference is within a certain width range) So.
処理ステップ212で縦書きの枠と判別される。In processing step 212, it is determined that the frame is for vertical writing.
次に外接矩形抽出部114において、枠領域画像メモリ
112を参照し、各枠内の画像に対して連結した黒画素
を追跡し、黒画素連結の外接矩形を抽出して、その対角
頂点の座標を外接矩形座標メモリ115に格納する(処
理ステップ214)。Next, the circumscribed rectangle extraction unit 114 refers to the frame area image memory 112, traces the connected black pixels for the image within each frame, extracts the circumscribed rectangle of connected black pixels, and extracts the circumscribed rectangle of the connected black pixels, and The coordinates are stored in the circumscribed rectangle coordinate memory 115 (processing step 214).
次に行画像抽出部116において、外接矩形座標メモリ
115を参照し、枠領域画像メモリ112内の各枠領域
画像に対して黒画素連結の外接矩形を1行方向判別部1
13により判別された行方向へ統合することにより、枠
内の文字行画像(文字列画像)を抽出し行画像メモリ1
17に格納する(処理ステップ215,216)−
このように各枠毎に行方向すなわち横書き・縦書きのい
ずれであるかの判別を行い、判別した行方向に適した方
法により文字行画像抽出を行うため、横書きの枠と縦書
きの枠が混在した表領域において、いずれの行方向の文
字行画像も正確に抽出することが可能となる。Next, the row image extraction unit 116 refers to the circumscribed rectangle coordinate memory 115 and extracts a circumscribed rectangle connected to black pixels for each frame area image in the frame area image memory 112 by the line direction determination unit 11.
By integrating in the row direction determined by 13, character row images (character string images) within the frame are extracted and stored in row image memory 1.
17 (processing steps 215, 216) - In this way, the line direction for each frame is determined, that is, whether it is horizontal writing or vertical writing, and character line image extraction is performed using a method suitable for the determined line direction. Therefore, in a table area where horizontal writing frames and vertical writing frames coexist, it becomes possible to accurately extract character line images in any row direction.
次に文字認識部118において、行画像メモリ117内
の各粋の文字行画像より文字画像を切り出すが、前段の
文字行画像抽出が正確であるため。Next, in the character recognition unit 118, a character image is cut out from each type of character line image in the line image memory 117, since the character line image extraction in the previous stage is accurate.
この文字画像切出しも正確に行うことができる。This character image cutting out can also be performed accurately.
そして、切り出した文字画像の特徴を抽出し、認識辞書
とのマツチングを行って認識し、認識結果を外部へ出力
する(処理ステップ217)。Then, the features of the cut out character image are extracted, recognized by matching with the recognition dictionary, and the recognition result is output to the outside (processing step 217).
以上説明した如く、本発明によれば、表中の各枠内に印
字された文字行が横書きであるか縦書きであるかを自動
的に判別し、判別した方向に応じた方法により文字行抽
出を行うので、横書き枠と縦書き枠が混在した表におい
ても、各枠内の文字行の切出しを精度良く行うことがで
き、したがって枠内文字のi8g!精度と上げることが
できる。As explained above, according to the present invention, it is automatically determined whether a character line printed in each frame in a table is written horizontally or vertically, and the character line is Since extraction is performed, even in a table with a mixture of horizontal and vertical writing frames, the character lines within each frame can be extracted with high precision, and therefore the i8g! Accuracy can be increased.
第1図は本発明の一実施例を示すブロック図、第2図は
処理のフローチャート、第3図は横書きと縦書きが混在
した表の例を示す図である。
101・・・2値画像入力部、
102・・・2値イメージメモリ、
103・・・表領域認識部。
104・・・表領域イメージメモリ、
105・・・主走査方向線分抽出部、
106・・・主走査方向線分座標メモリ、107・・・
副走査方向線分抽出部、
108・・・副走査方向線分座標メモリ、109・・・
枠認識部、 110・・・枠座標メモリ、1・・・枠領
域抽出部、
2・・・枠領域画像メモリ、
3・・・行方向判定部、
4・・・外接矩形抽出部、
5・・・外接矩形座標メモリ、
6・・・行画像抽出部、
7・・・行画像メモリ、 118・・・認識部。
第1図FIG. 1 is a block diagram showing an embodiment of the present invention, FIG. 2 is a flowchart of processing, and FIG. 3 is a diagram showing an example of a table in which horizontal writing and vertical writing are mixed. 101... Binary image input unit, 102... Binary image memory, 103... Table area recognition unit. 104...Table area image memory, 105...Main scanning direction line segment extraction unit, 106...Main scanning direction line segment coordinate memory, 107...
Sub-scanning direction line segment extraction unit, 108... Sub-scanning direction line segment coordinate memory, 109...
Frame recognition unit, 110...Frame coordinate memory, 1...Frame area extraction unit, 2...Frame area image memory, 3...Line direction determination unit, 4...Circumscribing rectangle extraction unit, 5. . . . Circumscribed rectangle coordinate memory, 6 . . . Line image extraction section, 7 . . . Line image memory, 118 . . . Recognition section. Figure 1
Claims (1)
向の線分で囲まれた枠を抽出し、各枠内の文字行を抽出
して文字認識する表内文字認識方法において、各枠の主
走査方向の長さ及び副走査方向の長さによって各枠内の
文字行が横書きであるか縦書きであるかを判別し、この
判別の結果に応じて各枠内の文字行の抽出方法を切り替
えることを特徴とする表内文字認識方法。(1) In the table character recognition method, a frame surrounded by line segments in the main scanning direction and sub-scanning direction is extracted from a table area of a document image, and character lines within each frame are extracted and characters are recognized. It is determined whether the character line in each frame is written horizontally or vertically depending on the length of the frame in the main scanning direction and the length in the sub-scanning direction, and the character line in each frame is determined according to the result of this determination. A method for recognizing characters in tables characterized by switching extraction methods.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2134876A JP2931041B2 (en) | 1990-05-24 | 1990-05-24 | Character recognition method in table |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2134876A JP2931041B2 (en) | 1990-05-24 | 1990-05-24 | Character recognition method in table |
Publications (2)
Publication Number | Publication Date |
---|---|
JPH0433080A true JPH0433080A (en) | 1992-02-04 |
JP2931041B2 JP2931041B2 (en) | 1999-08-09 |
Family
ID=15138558
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
JP2134876A Expired - Lifetime JP2931041B2 (en) | 1990-05-24 | 1990-05-24 | Character recognition method in table |
Country Status (1)
Country | Link |
---|---|
JP (1) | JP2931041B2 (en) |
-
1990
- 1990-05-24 JP JP2134876A patent/JP2931041B2/en not_active Expired - Lifetime
Also Published As
Publication number | Publication date |
---|---|
JP2931041B2 (en) | 1999-08-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP2940936B2 (en) | Tablespace identification method | |
JPH03122773A (en) | Image forming device | |
JP2812982B2 (en) | Table recognition method | |
JP3058791B2 (en) | Method of extracting figure of image recognition device | |
JPH0433080A (en) | Method for recognizing character in table | |
JP3095470B2 (en) | Character recognition device | |
JPH03160582A (en) | Method for separating ruled line and character in document picture data | |
JPH05128305A (en) | Area dividing method | |
JPH0433079A (en) | Table processing system | |
JP3140079B2 (en) | Ruled line recognition method and table processing method | |
JPH05274472A (en) | Image recognizing device | |
JP2001236464A (en) | Method and device for character extraction and storage medium | |
JP3162414B2 (en) | Ruled line recognition method and table processing method | |
JP2918363B2 (en) | Character classification method and character recognition device | |
JPH11306282A (en) | Method and device for extracting character string area | |
JPH0266681A (en) | Drawing processor | |
JPH04303277A (en) | Method for recognizing table | |
JP2004240500A (en) | Device and program for processing image, and storing medium | |
JPH1049602A (en) | Method for recognizing document | |
JPH05135202A (en) | Document picture reader | |
JPH04156694A (en) | Character recognition system | |
JPH0728933A (en) | Character recognition device | |
JPH103517A (en) | Device for detecting tilt angle of document picture | |
JPH02166583A (en) | Character recognizing device | |
JPH04346189A (en) | Character string type identification device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20080521 Year of fee payment: 9 |
|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20090521 Year of fee payment: 10 |
|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20100521 Year of fee payment: 11 |
|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20110521 Year of fee payment: 12 |
|
EXPY | Cancellation because of completion of term | ||
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20110521 Year of fee payment: 12 |