JPH0433080A - Method for recognizing character in table - Google Patents

Method for recognizing character in table

Info

Publication number
JPH0433080A
JPH0433080A JP2134876A JP13487690A JPH0433080A JP H0433080 A JPH0433080 A JP H0433080A JP 2134876 A JP2134876 A JP 2134876A JP 13487690 A JP13487690 A JP 13487690A JP H0433080 A JPH0433080 A JP H0433080A
Authority
JP
Japan
Prior art keywords
frame
character
line
scanning direction
main scanning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP2134876A
Other languages
Japanese (ja)
Other versions
JP2931041B2 (en
Inventor
Goro Bessho
吾朗 別所
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ricoh Co Ltd
Original Assignee
Ricoh Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ricoh Co Ltd filed Critical Ricoh Co Ltd
Priority to JP2134876A priority Critical patent/JP2931041B2/en
Publication of JPH0433080A publication Critical patent/JPH0433080A/en
Application granted granted Critical
Publication of JP2931041B2 publication Critical patent/JP2931041B2/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Landscapes

  • Character Input (AREA)

Abstract

PURPOSE:To enable character recognition by exactly extracting not only the line of laterally describing characters but also the line of longitudinally describing characters in a table by discriminating the lengths of each frame in main scanning and sub scanning directions and switching the method of extracting the character line in each frame corresponding to the result of the discrimination. CONSTITUTION:The frame surrounded by line segments in the main scanning and sub scanning directions is extracted from the table area of a document picture. Then, the lengths of each frame in the main scanning and sub scanning directions are discriminated and corresponding to the discriminated result, the method for extracting the character line in each frame is switched. Namely, it is automatically discriminated whether the character printed in each frame of the table is described laterally or longitudinally, and the character line extraction method suitable for the discriminated direction is applied. Thus, even in the table mixing the laterally and longitudinally describing frames, the character line in each frame is exactly extracted, and the accuracy for the character recognition is improved.

Description

【発明の詳細な説明】 〔産業上の利用分野〕 本発明は、文字認識装置における文書の表内文字認識方
法に関する。
DETAILED DESCRIPTION OF THE INVENTION [Field of Industrial Application] The present invention relates to a method for recognizing characters within a document in a character recognition device.

〔従来の技術〕[Conventional technology]

文字認識装置においては5文書画像を文字領域、写真や
図形などのイメージ領域、表領域などに分割し、それぞ
れの領域に別の処理を行うことが多い。
In character recognition devices, five document images are often divided into a character area, an image area such as a photograph or figure, a table area, etc., and different processing is performed on each area.

表領域に関しては、表を構成する罫線の位置を認識し、
罫線で囲まれた枠内の画像に対して連結した黒画素の追
跡を行い、黒画素連結の外接矩形を求め、それを統合し
て文字行を抽出し、文字認識している。
Regarding the table area, it recognizes the position of the borders that make up the table,
Connected black pixels are traced in an image within a frame surrounded by ruled lines, a circumscribed rectangle of connected black pixels is found, and the circumscribed rectangles are integrated to extract character lines and character recognition is performed.

(発明が解決しようとする課題〕 しかし従来は1表の各枠内の文字が横書きまたは縦書き
のいずれか一方で印字されていることを前提に文字行抽
出処理をしているため、例えば第3図に示すような横書
きの文字行と縦書きの文字行が混在した表の場合、横書
き(縦書き)を前提としているときには縦書き文字行(
横書き文字行)の抽出が正確に行われず、結果として文
字認識が正確に行われないことがある。
(Problem to be solved by the invention) However, conventionally, character line extraction processing is performed on the assumption that the characters in each frame of a table are printed either horizontally or vertically. In the case of a table with a mixture of horizontal and vertical text lines as shown in Figure 3, if horizontal text (vertical writing) is assumed, the vertical text lines (
(Horizontal character lines) may not be extracted accurately, and as a result, character recognition may not be performed accurately.

本発明の目的は、表内の横書き文字行も縦書き文字行も
正確に抽出して文字認識することができる表内文字認識
方法を提供することにある。
An object of the present invention is to provide a table character recognition method that can accurately extract and recognize both horizontally written character lines and vertically written character lines in a table.

〔課題を解決するための手段〕[Means to solve the problem]

本発明は、文書画像の表領域より、主走査方向及び副走
査方向の線分で囲まれた枠を抽出し、各枠内の文字行を
抽出して文字認識する表内文字認識方法において、各粋
の主走査方向の長さ及び副走査方向の長さによって各枠
内の文字行が横書きであるか縦書きであるかを判別し、
この判別の結果に応じて各枠内の文字行の抽出方法を切
り替えることを特徴とする。
The present invention provides a table character recognition method that extracts frames surrounded by line segments in the main scanning direction and sub-scanning direction from a table area of a document image, extracts character lines within each frame, and recognizes the characters. Determine whether the character line in each frame is written horizontally or vertically based on the length in the main scanning direction and the length in the sub-scanning direction of each frame,
The method is characterized in that the method for extracting character lines within each frame is switched depending on the result of this determination.

〔作 用〕[For production]

本発明によれば、表中の各枠内に印字された文字が横書
きであるか縦書きであるかが自動的に判別され5判別さ
れた方向に適した文字行抽出方法が適用されることによ
り、横書きの枠と縦書きの枠が混在した表においても、
各枠内の文字行が正確に抽出され、したがって各枠内の
文字の切り出し及び文字認識の精度が上がる。
According to the present invention, it is automatically determined whether the characters printed in each frame in the table are written horizontally or vertically, and a character line extraction method suitable for the determined direction is applied. Therefore, even in tables with a mixture of horizontal and vertical frames,
The character lines within each frame are accurately extracted, thus increasing the accuracy of character extraction and character recognition within each frame.

〔実施例〕〔Example〕

第1図は本発明の一実施例を示すブロック図。 FIG. 1 is a block diagram showing one embodiment of the present invention.

第2図は処理のフローチャートである。FIG. 2 is a flowchart of the process.

スキャナーなどの2値画像入力部101によって文書を
読取り、その2値画像を2値イメージメモリ102に格
納する(処理ステップ201)。
A document is read by a binary image input unit 101 such as a scanner, and the binary image is stored in the binary image memory 102 (processing step 201).

この文書画像に対して1表領域認識部103はランレン
グス分布などを利用して表領域を自動的に認識するか、
あるいはマウスなどを用いて操作者から指定された領域
を表領域として認識し、表領域のイメージを表領域イメ
ージメモリ104に格納する(処理ステップ202)。
For this document image, the table area recognition unit 103 automatically recognizes the table area using run length distribution or the like.
Alternatively, an area specified by the operator using a mouse or the like is recognized as a table area, and an image of the table area is stored in the table area image memory 104 (processing step 202).

この表領域のイメージに対し、主走査方向線分抽出部1
05において、主走査方向に連結した黒画素を追跡して
主走査方向の線分を抽出し、その始点及び終点の座標を
主走査方向線分座標メモリ106に格納する(処理ステ
ップ203)。同様に副走査方向線分抽出部107にお
いて、表領域イメージ内の副走査方向に連結した黒画素
を追跡して副走査方向の線分を抽出し、その始点及び終
点の座標を副走査方向線分座標メモリ108に格納する
(処理ステップ204)。
For this image of the table area, the line segment extraction unit 1 in the main scanning direction
05, black pixels connected in the main scanning direction are traced to extract a line segment in the main scanning direction, and the coordinates of the starting point and end point are stored in the main scanning direction line segment coordinate memory 106 (processing step 203). Similarly, the sub-scanning direction line segment extraction unit 107 traces the black pixels connected in the sub-scanning direction in the table area image, extracts a line segment in the sub-scanning direction, and calculates the coordinates of the starting point and end point along the sub-scanning direction line. The coordinates are stored in the minute coordinate memory 108 (processing step 204).

次に枠認識部109において、各メモリ106゜108
に格納された線分座標を参照し、主走査方向線分と副走
査方向線分の組合せにより表の枠を認識し、枠の座標例
えば対角頂点の座標を枠座標メモリ110に格納する(
処理ステップ205)。
Next, in the frame recognition unit 109, each memory 106° 108
The frame of the table is recognized by the combination of the line segment in the main scanning direction and the line segment in the sub-scanning direction, and the coordinates of the frame, for example, the coordinates of the diagonal vertices, are stored in the frame coordinate memory 110 (
Process step 205).

また枠領域抽出部111において、枠座標メモリ115
内の枠座標を参照することにより、表領域イメージメモ
リ104より枠の領域の画像を抽出して枠領域画像メモ
リ112に格納する(処理ステップ206)。
In addition, in the frame area extraction unit 111, the frame coordinate memory 115
By referring to the frame coordinates in , the image of the frame area is extracted from the table area image memory 104 and stored in the frame area image memory 112 (processing step 206).

次に行方向判定部113において、枠座標メモリ110
を参照して全ての枠に対して主走査方向及び副走査方向
の長さのヒストグラムを作成する(処理ステップ207
,208)、そして、&大度数の副走査方向の長さを持
つ枠は全て行方向が横書きの枠であると判別しく処理ス
テップ209゜210)、その長さと同じ主走査方向の
長さを持つ枠は行方向が縦書きの枠であると判別しく処
理ステップ211,212)、残った枠はそれまでに判
別された枠数が多いほうの行方向の枠であると判別する
(処理ステップ213)、なお、処理ステップ207,
208でヒストグラムを求める際には各走査方向の長さ
にある程度の幅を持たせ。
Next, in the row direction determination unit 113, the frame coordinate memory 110
A histogram of lengths in the main scanning direction and the sub-scanning direction is created for all frames with reference to (processing step 207
, 208), and all frames with a length in the sub-scanning direction of & large degrees are determined to be frames with horizontal writing in the line direction. The remaining frames are determined to be frames with vertical writing in the line direction (processing steps 211 and 212), and the remaining frames are determined to be vertical frames with the largest number of frames determined up to that point (processing steps 211 and 212). 213), and processing steps 207,
When calculating the histogram in step 208, a certain amount of width is given to the length in each scanning direction.

同様に処理ステップ211で長さを判別する際にも、比
較する長さの差がある幅の範囲内のときは一致すると判
定する。求められた行方向の情報は外接矩形抽出部11
4を経由して行画像抽出部116へ伝えられる。
Similarly, when determining the length in processing step 211, if the difference in length to be compared is within a certain width range, it is determined that they match. The obtained row direction information is sent to the circumscribed rectangle extraction unit 11
4 to the row image extraction unit 116.

例えば第3図に示した表の場合、横書きの枠の副走査方
向の長さは全て同一(あるいは、はぼ同一)であるので
、その頻度は最大である。したがって、この表の横書き
の枠はすべて処理ステップ210で横書きと判別される
。また、この表の縦書き文字列″データ″が印刷された
枠の主走査方向の長さは、最大頻度の副走査方向の長さ
とほぼ同一である(差が一定の幅の範囲である)ので。
For example, in the case of the table shown in FIG. 3, the lengths of the horizontally written frames in the sub-scanning direction are all the same (or almost the same), so the frequency is maximum. Therefore, all horizontal writing frames in this table are determined to be horizontal writing in processing step 210. Also, the length of the frame in which the vertically written character string "data" in this table is printed in the main scanning direction is almost the same as the length of the maximum frequency in the sub-scanning direction (the difference is within a certain width range) So.

処理ステップ212で縦書きの枠と判別される。In processing step 212, it is determined that the frame is for vertical writing.

次に外接矩形抽出部114において、枠領域画像メモリ
112を参照し、各枠内の画像に対して連結した黒画素
を追跡し、黒画素連結の外接矩形を抽出して、その対角
頂点の座標を外接矩形座標メモリ115に格納する(処
理ステップ214)。
Next, the circumscribed rectangle extraction unit 114 refers to the frame area image memory 112, traces the connected black pixels for the image within each frame, extracts the circumscribed rectangle of connected black pixels, and extracts the circumscribed rectangle of the connected black pixels, and The coordinates are stored in the circumscribed rectangle coordinate memory 115 (processing step 214).

次に行画像抽出部116において、外接矩形座標メモリ
115を参照し、枠領域画像メモリ112内の各枠領域
画像に対して黒画素連結の外接矩形を1行方向判別部1
13により判別された行方向へ統合することにより、枠
内の文字行画像(文字列画像)を抽出し行画像メモリ1
17に格納する(処理ステップ215,216)− このように各枠毎に行方向すなわち横書き・縦書きのい
ずれであるかの判別を行い、判別した行方向に適した方
法により文字行画像抽出を行うため、横書きの枠と縦書
きの枠が混在した表領域において、いずれの行方向の文
字行画像も正確に抽出することが可能となる。
Next, the row image extraction unit 116 refers to the circumscribed rectangle coordinate memory 115 and extracts a circumscribed rectangle connected to black pixels for each frame area image in the frame area image memory 112 by the line direction determination unit 11.
By integrating in the row direction determined by 13, character row images (character string images) within the frame are extracted and stored in row image memory 1.
17 (processing steps 215, 216) - In this way, the line direction for each frame is determined, that is, whether it is horizontal writing or vertical writing, and character line image extraction is performed using a method suitable for the determined line direction. Therefore, in a table area where horizontal writing frames and vertical writing frames coexist, it becomes possible to accurately extract character line images in any row direction.

次に文字認識部118において、行画像メモリ117内
の各粋の文字行画像より文字画像を切り出すが、前段の
文字行画像抽出が正確であるため。
Next, in the character recognition unit 118, a character image is cut out from each type of character line image in the line image memory 117, since the character line image extraction in the previous stage is accurate.

この文字画像切出しも正確に行うことができる。This character image cutting out can also be performed accurately.

そして、切り出した文字画像の特徴を抽出し、認識辞書
とのマツチングを行って認識し、認識結果を外部へ出力
する(処理ステップ217)。
Then, the features of the cut out character image are extracted, recognized by matching with the recognition dictionary, and the recognition result is output to the outside (processing step 217).

〔発明の効果〕〔Effect of the invention〕

以上説明した如く、本発明によれば、表中の各枠内に印
字された文字行が横書きであるか縦書きであるかを自動
的に判別し、判別した方向に応じた方法により文字行抽
出を行うので、横書き枠と縦書き枠が混在した表におい
ても、各枠内の文字行の切出しを精度良く行うことがで
き、したがって枠内文字のi8g!精度と上げることが
できる。
As explained above, according to the present invention, it is automatically determined whether a character line printed in each frame in a table is written horizontally or vertically, and the character line is Since extraction is performed, even in a table with a mixture of horizontal and vertical writing frames, the character lines within each frame can be extracted with high precision, and therefore the i8g! Accuracy can be increased.

【図面の簡単な説明】[Brief explanation of drawings]

第1図は本発明の一実施例を示すブロック図、第2図は
処理のフローチャート、第3図は横書きと縦書きが混在
した表の例を示す図である。 101・・・2値画像入力部、 102・・・2値イメージメモリ、 103・・・表領域認識部。 104・・・表領域イメージメモリ、 105・・・主走査方向線分抽出部、 106・・・主走査方向線分座標メモリ、107・・・
副走査方向線分抽出部、 108・・・副走査方向線分座標メモリ、109・・・
枠認識部、 110・・・枠座標メモリ、1・・・枠領
域抽出部、 2・・・枠領域画像メモリ、 3・・・行方向判定部、 4・・・外接矩形抽出部、 5・・・外接矩形座標メモリ、 6・・・行画像抽出部、 7・・・行画像メモリ、 118・・・認識部。 第1図
FIG. 1 is a block diagram showing an embodiment of the present invention, FIG. 2 is a flowchart of processing, and FIG. 3 is a diagram showing an example of a table in which horizontal writing and vertical writing are mixed. 101... Binary image input unit, 102... Binary image memory, 103... Table area recognition unit. 104...Table area image memory, 105...Main scanning direction line segment extraction unit, 106...Main scanning direction line segment coordinate memory, 107...
Sub-scanning direction line segment extraction unit, 108... Sub-scanning direction line segment coordinate memory, 109...
Frame recognition unit, 110...Frame coordinate memory, 1...Frame area extraction unit, 2...Frame area image memory, 3...Line direction determination unit, 4...Circumscribing rectangle extraction unit, 5. . . . Circumscribed rectangle coordinate memory, 6 . . . Line image extraction section, 7 . . . Line image memory, 118 . . . Recognition section. Figure 1

Claims (1)

【特許請求の範囲】[Claims] (1)文書画像の表領域より、主走査方向及び副走査方
向の線分で囲まれた枠を抽出し、各枠内の文字行を抽出
して文字認識する表内文字認識方法において、各枠の主
走査方向の長さ及び副走査方向の長さによって各枠内の
文字行が横書きであるか縦書きであるかを判別し、この
判別の結果に応じて各枠内の文字行の抽出方法を切り替
えることを特徴とする表内文字認識方法。
(1) In the table character recognition method, a frame surrounded by line segments in the main scanning direction and sub-scanning direction is extracted from a table area of a document image, and character lines within each frame are extracted and characters are recognized. It is determined whether the character line in each frame is written horizontally or vertically depending on the length of the frame in the main scanning direction and the length in the sub-scanning direction, and the character line in each frame is determined according to the result of this determination. A method for recognizing characters in tables characterized by switching extraction methods.
JP2134876A 1990-05-24 1990-05-24 Character recognition method in table Expired - Lifetime JP2931041B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2134876A JP2931041B2 (en) 1990-05-24 1990-05-24 Character recognition method in table

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2134876A JP2931041B2 (en) 1990-05-24 1990-05-24 Character recognition method in table

Publications (2)

Publication Number Publication Date
JPH0433080A true JPH0433080A (en) 1992-02-04
JP2931041B2 JP2931041B2 (en) 1999-08-09

Family

ID=15138558

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2134876A Expired - Lifetime JP2931041B2 (en) 1990-05-24 1990-05-24 Character recognition method in table

Country Status (1)

Country Link
JP (1) JP2931041B2 (en)

Also Published As

Publication number Publication date
JP2931041B2 (en) 1999-08-09

Similar Documents

Publication Publication Date Title
JP2940936B2 (en) Tablespace identification method
JPH03122773A (en) Image forming device
JP2812982B2 (en) Table recognition method
JP3058791B2 (en) Method of extracting figure of image recognition device
JPH0433080A (en) Method for recognizing character in table
JP3095470B2 (en) Character recognition device
JPH03160582A (en) Method for separating ruled line and character in document picture data
JPH05128305A (en) Area dividing method
JPH0433079A (en) Table processing system
JP3140079B2 (en) Ruled line recognition method and table processing method
JPH05274472A (en) Image recognizing device
JP2001236464A (en) Method and device for character extraction and storage medium
JP3162414B2 (en) Ruled line recognition method and table processing method
JP2918363B2 (en) Character classification method and character recognition device
JPH11306282A (en) Method and device for extracting character string area
JPH0266681A (en) Drawing processor
JPH04303277A (en) Method for recognizing table
JP2004240500A (en) Device and program for processing image, and storing medium
JPH1049602A (en) Method for recognizing document
JPH05135202A (en) Document picture reader
JPH04156694A (en) Character recognition system
JPH0728933A (en) Character recognition device
JPH103517A (en) Device for detecting tilt angle of document picture
JPH02166583A (en) Character recognizing device
JPH04346189A (en) Character string type identification device

Legal Events

Date Code Title Description
FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20080521

Year of fee payment: 9

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20090521

Year of fee payment: 10

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20100521

Year of fee payment: 11

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20110521

Year of fee payment: 12

EXPY Cancellation because of completion of term
FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20110521

Year of fee payment: 12