JPH03142691A - Table format document recognizing system - Google Patents

Table format document recognizing system

Info

Publication number
JPH03142691A
JPH03142691A JP1282736A JP28273689A JPH03142691A JP H03142691 A JPH03142691 A JP H03142691A JP 1282736 A JP1282736 A JP 1282736A JP 28273689 A JP28273689 A JP 28273689A JP H03142691 A JPH03142691 A JP H03142691A
Authority
JP
Japan
Prior art keywords
ruled line
projection data
ruled
vertical
candidates
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP1282736A
Other languages
Japanese (ja)
Other versions
JP2796561B2 (en
Inventor
Koji Katano
片野 浩司
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fuji Electric Co Ltd
Fuji Facom Corp
Original Assignee
Fuji Electric Co Ltd
Fuji Facom Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fuji Electric Co Ltd, Fuji Facom Corp filed Critical Fuji Electric Co Ltd
Priority to JP1282736A priority Critical patent/JP2796561B2/en
Publication of JPH03142691A publication Critical patent/JPH03142691A/en
Application granted granted Critical
Publication of JP2796561B2 publication Critical patent/JP2796561B2/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Landscapes

  • Character Input (AREA)
  • Character Discrimination (AREA)
  • Image Analysis (AREA)

Abstract

PURPOSE:To perform table recognition in a short time and accurately by dividing image information into belt areas with prescribed width, and extracting a rule mark candidate from projection data in a direction intersecting orthogonally to a longitudinal direction. CONSTITUTION:A vertical rule mark candidate can be accurately extracted by providing an image scanner 1 and a graphic separation part 2, dividing read binary image information into, for example, the plural belt areas in a horizontal direction, extracting the projection data for each belt area in the direction intersecting orthogonally to the longitudinal direction, and extracting, for example, the rule mark candidate with narrow width in a vertical direction from the projection data, and a vertical rule mark can be recognized by connecting extracted rule mark candidates in neighboring belt areas. A horizontal rule mark can be extracted similarly. Also, when a table is the one in which a character string is partitioned with a null area, rule mark recognition can be easily performed by extracting a temporary rule mark candidate by retrieving the null area even when it is the table not partitioned with a line. In such a way, the character recognition rate and recognition speed of a character recognition device can be improved.

Description

【発明の詳細な説明】 〔産業上の利用分野〕 この発明は、文字、表、図等が混在する文書を読込んで
文書画像情報とし、この文書画像情報から個々の情報を
分離して認識する表形式文書認識方式に関する。
[Detailed Description of the Invention] [Industrial Application Field] This invention reads a document containing a mixture of characters, tables, figures, etc. as document image information, and separates and recognizes individual information from this document image information. Concerning tabular document recognition methods.

〔従来の技術〕[Conventional technology]

従来、光学文字認識装置(OCR)等の文字認識装置で
は、文書を画像情報として読取り、この画像情報を行或
いは列と文字群とに分離し、さらに文字と思われる所で
細分化して文字認識を行うようにしているが、表中の文
字を認識する場合には、罫線が含まれていることから枠
内の文字については行間や文字間がことなるため、誤認
識も多くなり、しかも表の構造を認識しているわけでは
ないので、その表自身が持つ意味を失わせてしまうとい
う課題があった。
Conventionally, character recognition devices such as optical character recognition devices (OCR) read a document as image information, separate this image information into rows or columns and character groups, and further subdivide the parts that are considered to be characters for character recognition. However, when recognizing characters in a table, since the characters in the frame include ruled lines, the line spacing and character spacing are different, resulting in many misrecognitions. Since the structure of the table is not recognized, the problem is that the table itself loses its meaning.

この課題を解決するために、従来、昭和63年電子情報
通信学会春季全国大会予稿集第2−224頁に記載され
ているように、文書画像を水平(又は垂直)方向に走査
し、黒画素数のヒストグラムを形成し、次いでヒストグ
ラムの山の部分について、高さが闇値以上である区間を
山の幅として、その幅が小さいものを抽出し、それに対
応する文書画像の領域を線分の領域とし、さらに線分領
域を線分の方向に走査し、白画素が闇値以上連続する部
分を終点として線分を抽出し、その後抽出した線分から
表の外接矩形を求めて欄情報を生威し、裏構造を認識す
ることが提案されている。
In order to solve this problem, conventionally, as described in the Proceedings of the 1986 Institute of Electronics, Information and Communication Engineers Spring National Conference, pages 2-224, document images are scanned in the horizontal (or vertical) direction, and black pixels are Form a histogram of the numbers, and then, for the mountain part of the histogram, the section whose height is greater than or equal to the darkness value is set as the mountain width, and the one with the smaller width is extracted, and the corresponding area of the document image is divided into line segments. The line segment area is then scanned in the direction of the line segment, and the line segment is extracted with the end point being the part where white pixels are continuous over the darkness value.Then, the circumscribed rectangle of the table is calculated from the extracted line segment and column information is generated. It has been proposed to recognize the underlying structure.

〔発明が解決しようとする課題〕[Problem to be solved by the invention]

しかしながら、上記従来の表形式文書認識方式にあって
は、文書画像を水平又は垂直方向に走査する関係で、水
平又は垂直方向に連続する罫線の場合には、罫線を切り
出すことができるが、例えば水平方向の2本の罫線間を
結ぶ短い垂直方向の罫線がある場合にはこれを検出する
ことができないので、細かく区分けされた表を正確に認
識するには、文書画像を短い罫線を認識可能な大きさの
領域に分割し、各領域毎に水平及び垂直方向の走査を行
ってヒストグラムを作成する必要があり、画像認識を行
う場合の処理時間が長くなると共に、分割領域の周縁部
に罫線があるときには判別が困難であり、また各分割領
域で抽出した罫線候補を連結する場合の処理時間も長く
なり、さらには文字群が実線、点線等の線によって区分
けされず、空白領域による仮想罫線によって区分けされ
た表については表認識を行うことができないという未解
決の課題があった。
However, in the conventional tabular document recognition method described above, when a document image is scanned horizontally or vertically, if the ruled lines are continuous in the horizontal or vertical direction, the ruled lines can be cut out. If there is a short vertical ruled line connecting two horizontal ruled lines, it cannot be detected, so in order to accurately recognize a finely divided table, it is necessary to recognize short ruled lines in the document image. It is necessary to divide each area into areas of a certain size and create a histogram by scanning each area in the horizontal and vertical directions, which increases the processing time for image recognition and creates ruled lines around the edges of the divided areas. In addition, it takes a long time to process when connecting the ruled line candidates extracted from each divided area, and furthermore, character groups are not separated by solid lines, dotted lines, etc., and virtual ruled lines are created by blank areas. There was an unresolved problem that table recognition could not be performed for tables divided by .

そこで、この発明は、上記従来例の未解決の課題に着目
してなされたものであり、表認識を短時間で正確に行う
ことができ、また空白領域による仮想罫線であっても容
易に表認識を行うことができる表形式文書認識方式を提
供することを目的としている。
Therefore, this invention was made by focusing on the unresolved problems of the conventional example described above, and it is possible to perform table recognition accurately in a short time, and also to easily display even virtual ruled lines made of blank areas. The purpose is to provide a tabular document recognition method that can perform recognition.

〔課題を解決するための手段〕[Means to solve the problem]

上記目的を遠戚するために、請求項(1)に係る表形式
文書認識方式は、文字と罫線とが混在する表形式文書を
画像読取装置で読取り、該画像読取装置からの画像情報
中の罫線を切り出し、該罫線によって両底された領域の
文字情報を文字認識装置に送出するようにした表形式文
書認識方式であって、前記画像情報を所定幅の帯状領域
に分割し、各帯状領域をその長手方向と直交する方向に
投影した投影データを形成する投影データ形成手段と、
該投影データ形成手段で形成した投影データから罫線候
補を抽出する罫線候補抽出手段と、該罫線候補抽出手段
で抽出した罫線候補を隣接する帯状領域間で連結して罫
線を認識する罫線認識手段とを備えたことを特徴として
いる。
In order to achieve the above object, the tabular document recognition method according to claim (1) reads a tabular document containing a mixture of characters and ruled lines with an image reading device, and uses the image information in the image reading device to read the tabular document in which characters and ruled lines are mixed. This is a tabular document recognition method in which ruled lines are cut out and character information in areas that are bottomed by the ruled lines is sent to a character recognition device. projection data forming means for forming projection data in which is projected in a direction perpendicular to the longitudinal direction;
ruled line candidate extraction means for extracting ruled line candidates from the projection data formed by the projection data forming means; ruled line recognition means for recognizing ruled lines by connecting the ruled line candidates extracted by the ruled line candidate extraction means between adjacent strip areas; It is characterized by having the following.

また、請求項(2)に係る表形式文書認識方式は、前記
請求項(1)において、罫線候補抽出手段は、文字列間
に空白領域があるときに、当該空白領域から仮想罫線候
補を抽出するように構成されていることを特徴としてい
る。
Further, in the tabular document recognition method according to claim (2), in the claim (1), when there is a blank area between character strings, the ruled line candidate extracting means extracts a virtual ruled line candidate from the blank area. It is characterized by being configured to do so.

〔作用〕[Effect]

請求項(1)に係る表形式文書認識方式においては、画
像読取装置で読取った2値化画像情報を例えば水平方向
の複数の帯状領域に分割し、これら帯状領域の夫々につ
いてその長手方向と直交する方向に投影データをとり、
この投影データから例えば幅狭の垂直方向の罫線候補を
抽出することにより、垂直罫線候補を正確に抽出するこ
とができ、この抽出した罫線候補を隣接する帯状領域間
で連結することにより、垂直罫線を認識することができ
る。
In the tabular document recognition method according to claim (1), the binarized image information read by the image reading device is divided into, for example, a plurality of strip-shaped regions in the horizontal direction, and each of these strip-shaped regions is divided orthogonally to its longitudinal direction. Take the projection data in the direction of
For example, by extracting narrow vertical ruled line candidates from this projection data, vertical ruled line candidates can be accurately extracted, and by connecting these extracted ruled line candidates between adjacent strip areas, vertical ruled line candidates can be can be recognized.

また、水平罫線を抽出するには、2値化画像情報を垂直
方向の複数の帯状領域に分割し、その長手方向と直交す
る投影データから水平方向罫線候補を抽出し、各帯状領
域の罫線候補を連結することにより、水平罫線を抽出す
ることができる。この水平罫線を抽出する際に、予め前
記処理によって抽出した垂直罫線を前記2値化画像情報
から消去しておくことにより、罫線認識の時間を短縮す
ることができる。
In addition, in order to extract horizontal ruled lines, the binarized image information is divided into multiple strip-shaped regions in the vertical direction, horizontal ruled line candidates are extracted from projection data perpendicular to the longitudinal direction, and ruled line candidates for each strip-shaped region are extracted. By concatenating , horizontal ruled lines can be extracted. When extracting this horizontal ruled line, by deleting the vertical ruled line extracted by the process from the binarized image information in advance, the time for recognizing the ruled line can be shortened.

また、請求項(2)に係る表形式文書認識方式において
は、文字列が空白領域で区分されている表である場合に
、その空白領域を探索して仮想罫線候補を抽出すること
により、線で区分けされていない表であっても、罫線認
識を容易に行うことかできる。
In addition, in the tabular document recognition method according to claim (2), when the character string is a table divided by blank areas, by searching the blank areas and extracting virtual ruled line candidates, line Even if the table is not divided into sections, it is possible to easily recognize ruled lines.

〔実施例〕〔Example〕

以下、この発明の実施例を図面に基づいて説明する。 Embodiments of the present invention will be described below based on the drawings.

第1図はこの発明の概略構成を示すブロック図である。FIG. 1 is a block diagram showing a schematic configuration of the present invention.

図中、lはイメージスキャナーであって、パーソナルコ
ンピュータ等で構成される処理装置4によって制御され
、第2図に示すように、垂直罫線Rv及び水平罫線RH
によって形成された表T内に任意の文字が描かれた表形
式文書りを読取り、これを所定レベルで黒白画素に2値
化した2値化画像情報を出力する。
In the figure, l is an image scanner, which is controlled by a processing device 4 composed of a personal computer or the like, and as shown in FIG. 2, vertical ruled lines Rv and horizontal ruled lines RH
A tabular document in which arbitrary characters are drawn in a table T formed by is read, and this is binarized into black and white pixels at a predetermined level to output binarized image information.

このイメージスキャナー1の2値化画像情報は、図形分
離部2に送出され、この図形分離部2で表情報とこれに
よって区分けされた文字情報とに分割され、文字情報が
光学文字認識装置(OCR)3に入力され、この光学文
字認識装置3で文字認識が行われて認識された文字デー
タが処理装置4に入力される。
The binarized image information from the image scanner 1 is sent to the figure separator 2, where it is divided into table information and character information separated by the table information. ) 3, the optical character recognition device 3 performs character recognition, and the recognized character data is input to the processing device 4.

図形分離部2は、例えばインタフェース回路2a、演算
処理装置2b及びドツト対応の画像メモリを含む記憶装
置2Cを少なくとも有し、イメージスキャナーlからの
2値化画像情報が入力されると、この2値化画像情報を
記憶装置2Cの画像メモリに格納し、次いで演算処理装
置2bで第3図の処理を実行する。すなわち、先ずステ
ップので垂直罫線Rvを抽出する垂直罫線抽出処理を実
行し、次いでステップ■に移行して水平罫線Rgを抽出
する水平罫線抽出処理を実行し、次いでステップ■に移
行して罫線によらず罫線に対応する空白領域による仮想
罫線R,を抽出する仮想罫線抽出処理を実行し、次いで
ステップ■に移行して各処理で抽出した垂直罫線RV、
水平罫線R工及び仮想罫線R7で囲まれる欄内の画像情
報を光学文字読取装置3に送出する。
The figure separator 2 has at least an interface circuit 2a, an arithmetic processing unit 2b, and a storage device 2C including a dot-compatible image memory, and when the binary image information from the image scanner 1 is input, the figure separator 2 converts the binary image information into The converted image information is stored in the image memory of the storage device 2C, and then the processing shown in FIG. 3 is executed by the arithmetic processing device 2b. That is, first, step 2 executes vertical ruled line extraction processing to extract vertical ruled lines Rv, then moves to step 2, executes horizontal ruled line extraction processing to extract horizontal ruled lines Rg, and then moves to step 2, which extracts horizontal ruled lines Rv. A virtual ruled line extraction process is executed to extract a virtual ruled line R, which is a blank area corresponding to the ruled line, and then the process proceeds to step 2, where vertical ruled lines RV, extracted in each process are extracted.
The image information in the field surrounded by the horizontal ruled line R and the virtual ruled line R7 is sent to the optical character reading device 3.

ここで、垂直罫線抽出処理は、第4図に示すようにステ
ップ■で画像メモリに格納されている2値化画像情報を
、第7図(a)に示すように、所定幅の水平方向(X方
向)に延長する複数N個の帯状領域B X I−B X
 Nに分割するように水平方向に順次垂直方向の所定幅
W分づつ読出し、夫々の帯状領域BX+”’BXNにつ
いてその長手方向(X方向)と直交する垂直方向(Y方
向)に投影して第7図(b)に示すように、黒画素とな
る論理値“l”の垂直投影データを形成し、次いでステ
ップ@に移行して、この投影データから垂直罫線候補を
抽出する。この垂直罫線候補の抽出は、投影データを予
め設定した閾値で再度2値化し、この2値化データの“
l”が連続する部分を探索し、罫線は通常幅狭であるこ
とから′°1”の連続幅が予め設定した幅以下であると
きに、これを垂直罫線候補として抽出し、その他を文字
候補として抽出する。
Here, in the vertical ruled line extraction process, as shown in FIG. 4, the binarized image information stored in the image memory in step A plurality of N band-shaped regions B X I-B X extending in the X direction)
A predetermined width W in the vertical direction is read out sequentially in the horizontal direction so as to divide it into As shown in FIG. 7(b), vertical projection data of the logical value "l" which is a black pixel is formed, and then the process proceeds to step @, where a vertical ruled line candidate is extracted from this projection data.This vertical ruled line candidate To extract the data, the projection data is binarized again using a preset threshold, and this binarized data is
Search for consecutive parts of ``l'', and since ruled lines are usually narrow, if the continuous width of ``°1'' is less than a preset width, this is extracted as a vertical ruled line candidate, and the others are used as character candidates. Extract as.

次いで、ステップ■に移行して、文字候補中に含まれる
垂直罫線候補を抽出する罫線候補再抽出処理を実行する
。すなわち、第7図(a)に示すように、ある帯状領域
BX、内に水平罫線R□を含む場合には、その投影デー
タは、第7図(b)に示すように、水平罫線RHを表す
論理値“1”分だけ嵩上げされるので、第7図(ロ)で
点線図示の所定の閾値T、で再2値化を行っても、各垂
直罫線R8間で論理値“l”が連続することになり、全
体が文字候補として認識されてしまう、このため、罫線
候補再抽出処理では、論理値“1”が連続する幅区間内
の最大値(罫線を含んでいるときには、その罫線部分が
通常は高いピークとなる)と、その幅区間内の平均値と
で再々2値化を行うことにより、第7図(C)に示すよ
うに、文字候補中に含まれる垂直罫線候補を抽出するこ
とができる。
Next, the process moves to step (2), and a ruled line candidate re-extraction process is executed to extract vertical ruled line candidates included in the character candidates. That is, as shown in FIG. 7(a), when a certain strip area BX includes a horizontal ruled line R□, its projection data includes a horizontal ruled line RH as shown in FIG. 7(b). Since the logical value "1" is increased by the amount represented by the logical value "1", even if binarization is performed again at the predetermined threshold value T shown by the dotted line in FIG. Therefore, in the ruled line candidate re-extraction process, the maximum value in the width interval where the logical value "1" is continuous (if it includes a ruled line, the ruled line By performing binarization again using the average value within the width section (which usually has a high peak) and the average value within the width section, vertical ruled line candidates included in the character candidates can be determined, as shown in Figure 7 (C). can be extracted.

次いで、ステップ[相]に移行して、各帯状領域BX1
〜BX、について、罫線候補の抽出が完了したか否かを
判定し、罫線候補抽出が完了していない帯状領域がある
ときには、ステップ[相]に移行して次の画像メモリの
続出アドレスを次の帯状領域BX11となるように設定
したから前記ステップ■に戻り、全ての帯状領域B X
 t〜BX、について罫線候補の抽出が完了したときに
は、ステップ[相]に移行して各帯状領域BX、〜BX
、で抽出した垂直罫線候補を垂直方向に連結して垂直罫
線候補列を形成する。この垂直罫線候補列を形成するに
は、隣接する帯状領域間で各罫線候補群から垂直方向(
Y方向)に重なりを持つ垂直罫線候補を抽出して両者を
連結することにより行う。
Next, proceeding to step [phase], each band-shaped region BX1
~BX, it is determined whether or not extraction of ruled line candidates has been completed, and if there is a band-shaped area for which ruled line candidate extraction has not been completed, the process moves to step [phase] and the successive addresses of the next image memory are Since the setting is made so that the strip area BX11 is set, return to the step
When the extraction of ruled line candidates for t~BX is completed, the process moves to step [phase] and the extraction of ruled line candidates for each strip area BX, ~BX is completed.
The vertical ruled line candidates extracted in , are vertically connected to form a vertical ruled line candidate sequence. To form this vertical ruled line candidate row, from each ruled line candidate group in the vertical direction (
This is done by extracting vertical ruled line candidates that overlap in the Y direction) and connecting them.

次いで、ステップ@に移行して、垂直罫線候補列をその
延長方向と直交する方向即ちX方向に投影して投影デー
タをとり、この投影データがあるレベル以上連続してい
るときに罫線として認識し、その投影データから始端及
び終端座標を求めることで垂直罫線を決定してから処理
を終了して第3図の処理に復帰する。
Next, in step @, the vertical ruled line candidate row is projected in a direction perpendicular to its extension direction, that is, in the X direction to obtain projection data, and when this projection data is continuous at a certain level or more, it is recognized as a ruled line. , the vertical ruled lines are determined by determining the start and end coordinates from the projection data, and then the process is terminated and the process returns to the process of FIG. 3.

また、水平罫線抽出処理は、第5図に示すように先ずス
テップ■で、画像メモリに格納されている2値化画像情
報から前記垂直罫線候補抽出処理で決定された垂直罫線
Rvを消去した消去2値化画像情報を作成し、これを画
像メモリに格納する。
In the horizontal ruled line extraction process, as shown in FIG. Binarized image information is created and stored in the image memory.

次いで、ステップ@に移行して、消去2値化画像情報に
ついて、前記垂直罫線候補抽出処理における帯状領域B
X、〜B X sに代えて所定幅Wで垂直方向に分割し
て垂直帯状領域BY、〜BY。
Next, the process moves to step @, and for the erased binarized image information, the strip area B in the vertical ruled line candidate extraction process is
Instead of X, ~B

を形成するように、消去2値化画像情報を垂直方向に所
定幅分づつ読出し、その長手方向(Y方向)とは直交す
るX方向に投影して投影データを作威する。
The erased binarized image information is read out vertically by a predetermined width and projected in the X direction perpendicular to the longitudinal direction (Y direction) to create projection data.

次いで、ステップ0に移行して、投影データに基づいて
前記垂直罫線候補抽出処理と同様の処理を行うことによ
り水平罫線候補を抽出する。
Next, in step 0, horizontal ruled line candidates are extracted by performing the same process as the vertical ruled line candidate extraction process based on the projection data.

次いで、ステップ[相]に移行して全ての垂直帯状領域
BY、−BY、について水平罫線候補の抽出を完了した
か否かを判定し、抽出が未完了であるときには、ステッ
プ@に移行して次の垂直帯状領域B Y j+ rに対
応する続出アドレスを設定してから前記ステップ@に戻
り、全ての垂直帯状領域BY1〜BY、について水平罫
線候補の抽出が完了したときには、ステップ[相]に移
行する。
Next, the process moves to step [phase] to determine whether extraction of horizontal ruled line candidates has been completed for all the vertical strip areas BY, -BY, and if the extraction has not been completed, the process moves to step @. After setting the successive address corresponding to the next vertical strip area B Y j + r, return to step @, and when the extraction of horizontal ruled line candidates for all vertical strip areas BY1 to BY is completed, proceed to step [phase]. Transition.

このステップ[相]では、隣接する垂直帯状領域の水平
罫線候補を抽出して連結することにより、水平罫線候補
列を作威し、次いでステップOに移行して水平罫線候補
列をX方向に投影することにより、水平罫線を決定する
In this step [phase], a horizontal ruled line candidate row is created by extracting and concatenating horizontal ruled line candidates in adjacent vertical strip areas, and then the process moves to step O to project the horizontal ruled line candidate row in the X direction. By doing this, the horizontal ruled lines are determined.

さらに、仮想罫線抽出処理は、第6図に示すように、先
ずステップ[相]で、第8図に示すように、上記垂直罫
線候補抽出処理及び水平罫線候補抽出処理によって求め
られた垂直罫線Rv及び水平罫線RMの最外罫線に外接
する矩形の最外罫線枠Fの内側を仮想罫線処理領域に設
定し、次いでステップ@に移行して、画像メモリに格納
されている垂直罫線消去2値化画像情報から水平罫L9
. R,を消去して第9図に示すような2値化画像情報
を形成し、次いでステップ0に移行して、2値化画像情
報を前述した水平罫線候補抽出処理と同様に、複数M個
の帯状領域BY、−BY、に分割し、これら帯状領域B
Y、〜BYイの夫々についてX方向の投影データを形成
し、次いでステップ[相]に移行して、投影データから
論理値゛0”を表す領域を仮想罫線候補として抽出する
。この結果、例えば各帯状領域BYi−,,BYえ及び
B Y t −tで第1O図(a)に示すように、空白
部の仮想罫線候補を抽出することができる。
Furthermore, as shown in FIG. 6, in the virtual ruled line extraction process, first, in step [phase], as shown in FIG. Then, the inside of the rectangular outermost ruled line frame F circumscribing the outermost ruled line of the horizontal ruled line RM is set as a virtual ruled line processing area, and then the process moves to step @, and the vertical ruled line stored in the image memory is erased and binarized. Horizontal rule L9 from image information
.. R, is deleted to form binarized image information as shown in FIG. into strip regions BY, -BY, and these strip regions B
Projection data in the X direction is formed for each of Y, ~BYA, and then the process moves to step [phase], where an area representing a logical value "0" is extracted as a virtual ruled line candidate from the projection data.As a result, for example, As shown in FIG. 1O(a), virtual ruled line candidates for blank areas can be extracted from each band-shaped area BYi-, , BYe, and BYt-t.

次いで、ステップ[相]に移行して全ての垂直帯状領域
BY、−BY、について仮想罫線候補の抽出が完了した
か否かを判定し、仮想罫線候補の抽出が未完了の垂直帯
状領域があるときには、ステップ[相]に移行して画像
メモリに対する続出アドレスを次の帯状領域BY、、、
に対応するアドレスに更新してからステップ0に戻り、
全ての垂直帯状領域B Y +〜BYwについて仮想罫
線候補の抽出が完了したときには、ステップ■に移行す
る。
Next, the process moves to step [phase] to determine whether extraction of virtual ruled line candidates has been completed for all vertical strip areas BY, -BY, and there is a vertical strip area for which extraction of virtual ruled line candidates has not been completed. Sometimes, the process moves to step [phase] and successive addresses for the image memory are transferred to the next strip area BY,...
Update to the address corresponding to , then return to step 0,
When the extraction of virtual ruled line candidates is completed for all vertical strip areas B Y + to BYw, the process moves to step (2).

このステップ■では、前記ステップ[相]で抽出した仮
想罫線候補は、空白部の全てが候補となるため、候補の
大きさの違いが著しくなり、例えば第10図(a)にお
ける帯状領域BY□では、全てが仮想罫線候補となって
しまうので、この帯状領域BYtの仮想罫線候補と、こ
れに隣接する帯状領域BYえ−、及びBY、。1の仮想
罫線候補CA r及びCAZの数を算出し、両者の仮想
罫線候補数を比較して何れか多い方の仮想罫線候補に合
わせて仮想罫線候補を分割して仮想罫線分割候補CA。
In this step (2), the virtual ruled line candidates extracted in the step [phase] include all the blank areas, so the difference in the size of the candidates becomes significant. For example, in the strip area BY□ in FIG. 10(a) Then, since all of them become virtual ruled line candidates, the virtual ruled line candidates of this strip area BYt and the adjacent strip areas BYe and BY. The numbers of virtual ruled line candidates CA r and CAZ are calculated, the numbers of both virtual ruled line candidates are compared, and the virtual ruled line candidates are divided according to the larger virtual ruled line candidate to create virtual ruled line division candidates CA.

及びCAz’を選定する。and select CAz'.

次いで、ステップ[相]に移行して、抽出した仮想罫線
分割候補を含む全ての仮想罫線候補のうち、隣接する帯
状wI域の互いに水平方向連接する仮想罫線候補を探索
して連結し、仮想罫線候補列を形成する。
Next, the process moves to step [phase], and among all virtual ruled line candidates including the extracted virtual ruled line division candidates, virtual ruled line candidates that are horizontally connected to each other in adjacent strip wI areas are searched and connected, and the virtual ruled line candidates are searched and connected. Form candidate columns.

次いで、ステップ0に移行して、仮想罫線候補列のうち
前述した罫線候補抽出処理で決定した垂直罫線Rv及び
水平罫線R9と重なる仮想罫線候補列を消去し、残りの
仮想罫線候補列について仮想罫線を決定してから第4図
の処理に戻る。この仮想罫線の決定は、仮想罫線候補列
が罫線に比較して幅広であるので、仮想罫線候補列の始
端及び終端における仮想罫線候補の中点を代表点とし、
この代表点の座標を求めることにより、両代表点を結ぶ
線分を仮想罫線として決定する。
Next, the process moves to step 0, in which virtual ruled line candidate columns that overlap with the vertical ruled line Rv and horizontal ruled line R9 determined in the above-described ruled line candidate extraction process are deleted from the virtual ruled line candidate columns, and virtual ruled line candidate columns are deleted for the remaining virtual ruled line candidate columns. After determining this, the process returns to the process shown in FIG. In determining this virtual ruled line, since the virtual ruled line candidate row is wider than the ruled line, the midpoint of the virtual ruled line candidate at the start and end of the virtual ruled line candidate row is used as a representative point,
By determining the coordinates of this representative point, a line segment connecting both representative points is determined as a virtual ruled line.

なお、仮想罫線を決定するときに、福に第11図(a)
に示すように幅広の仮想罫線候補CA、に対して幅狭の
複数の仮想罫線候補CA + z及びCA13が連接す
る場合があり、この場合には幅広の仮想罫線候補CA 
+ +の中点と幅狭の仮想罫線候補CA2.の中点とを
代表点に選定すると、仮想罫線を誤抽出することになる
ので、幅広の仮想罫線候補と幅狭の仮想罫線候補とが連
結されて仮想罫線候補列が形成される場合には、第11
図(ロ)に示すように、幅狭の仮想罫線候補の何れか一
方の中点を唯一の代表点として選択し、仮想罫線を決定
すると共に、残りの幅狭の仮想罫線候補に対しては幅広
罫線候補に接する境界位置を仮想罫線の終点又は始点座
標として決定する。
In addition, when determining the virtual ruled lines, please refer to Fig. 11(a).
As shown in the figure, there are cases where multiple narrow virtual ruled line candidates CA + z and CA13 are connected to the wide virtual ruled line candidate CA, and in this case, the wide virtual ruled line candidate CA
+ Midpoint of + and narrow virtual ruled line candidate CA2. If the midpoint of is selected as the representative point, virtual ruled lines will be extracted incorrectly, so when a virtual ruled line candidate string is formed by connecting wide virtual ruled line candidates and narrow virtual ruled line candidates, , 11th
As shown in Figure (B), the midpoint of one of the narrow virtual ruled line candidates is selected as the only representative point to determine the virtual ruled line, and the remaining narrow virtual ruled line candidates are The boundary position touching the wide ruled line candidate is determined as the end point or starting point coordinates of the virtual ruled line.

このようにして、全ての罫線及び仮想罫線を決定するこ
とにより表形式の認識が終了したら、前記第4図のステ
ップ■に移行して、各罫線で囲まれる欄の文字情報を一
括して光学文字認識装置3に出力する。
In this way, when the recognition of the tabular format is completed by determining all the ruled lines and virtual ruled lines, the process moves to step (2) in FIG. Output to character recognition device 3.

したがって、光学文字認識装置3では、表形式文書りの
各欄毎に文字1mを行うことになるので、文字認識率を
向上させることができると共に、裏構造を認識すること
ができるので、表形式文書の行又は列に対しての項目毎
に文字のみ或いは数字のみなどの指定を行うことが可能
となり、文字の選択数を限定することができ、認識率と
認識速度の向上を図ることができる。さらに、光学文字
認識装置3で認識した文字を制御コードに変換してパー
ソナルコンピュータ等で構成される演算処理装置4に入
力することにより、住所録や電話帳などの氏名、住所、
電話番号等のデータベースを作成する場合に、これらの
文書をイメージスキャナーで読込むだけでよく、データ
ベースの作成を容易に行うことができる。
Therefore, the optical character recognition device 3 processes 1 m of characters for each column of the tabular document, which improves the character recognition rate and recognizes the back structure. It is now possible to specify only letters or only numbers for each item in a row or column of a document, and the number of selected characters can be limited, improving recognition rate and speed. . Furthermore, by converting the characters recognized by the optical character recognition device 3 into control codes and inputting them into the arithmetic processing device 4 composed of a personal computer or the like, names and addresses in address books, telephone directories, etc.
When creating a database of telephone numbers, etc., it is only necessary to read these documents with an image scanner, and the database can be created easily.

なお、上記実施例においては、画像読取装置としてイメ
ージスキャナー1を適用した場合について説明したが、
これに限定されるものではなく、テレビカメラ、固体撮
像装置等の任意の画像読取装置を適用することができる
Note that in the above embodiment, the case where the image scanner 1 is applied as the image reading device has been described.
The present invention is not limited to this, and any image reading device such as a television camera, solid-state imaging device, etc. can be applied.

また、上記実施例においては、表形式文書の罫線が実線
である場合について説明したが、これに限定されるもの
ではなく、破線、鎖線等の罫線である場合にも投影デー
タのピークが実線の場合に比較して小さくなるだけであ
るので、容易に罫線を認識することができる。
Furthermore, in the above embodiment, the case where the ruled line of the tabular document is a solid line has been explained, but the invention is not limited to this, and even when the ruled line is a broken line, a chain line, etc., the peak of the projection data is a solid line. The ruled lines can be easily recognized because they are only smaller than in the case where the ruled lines are displayed.

さらに、上記実施例においては、イメージスキャナー1
の読取方向と表形式文書の罫線方向とが一致している場
合について説明したが、これに限らず表形式文書の2値
化画像情報がイメージスキャナー1の読取方向とずれて
いる場合には、垂直罫線Rv、水平罫線RH及び仮想罫
IIRIの始終端座標を決定する場合に、罫線候補列の
罫線候補を抽出する場合の投影方向とは直交する方向の
投影データをとり、これによって始端及び終端座標を求
めると共に、罫線候補列の方程式を求めることにより、
始点及び終点座標を決定することができる。但し、仮想
罫線候補については、前述したように実在する罫線がな
いことにより、代表点となる仮想罫線候補列の始端及び
終端の中点を結ぶ直線の方程式を利用する。
Furthermore, in the above embodiment, the image scanner 1
The case where the reading direction of the image scanner 1 matches the ruled line direction of the tabular document has been described; When determining the start and end coordinates of vertical ruled lines Rv, horizontal ruled lines RH, and virtual ruled lines IIRI, projection data in a direction perpendicular to the projection direction used when extracting ruled line candidates from the ruled line candidate array is taken, and the starting and ending points are determined using projection data. By finding the coordinates and the equation of the ruled line candidate row,
Starting point and ending point coordinates can be determined. However, as for the virtual ruled line candidates, since there are no real ruled lines as described above, an equation of a straight line connecting the midpoints of the start and end points of the virtual ruled line candidate row, which are representative points, is used.

〔発明の効果〕〔Effect of the invention〕

以上説明したように、請求項(1)に係る表形式文書認
識方式にあっては、表形式文書の表を形成する垂直罫線
及び水平罫線をそれぞれ両者に直交する帯状領域に分割
した2値化画像情報をその長手方向と直交する方向に投
影した投影データに基づいて罫線候補を抽出するように
したので、2値化画像情報の分割数を少なくすることが
でき、表形式文書の認識速度を向上させることができ、
しかも裏構造を認識してこの裏構造の各欄毎に文字認識
を行うので、行又は列に対しての項目毎に文字のみ或い
は数字のみ等の指定を行うことができ、文字認識装置で
の文字の選択数を限定できることで、文学誌m装置の文
字認識率及び認識速度を向上させることができ、さらに
は表形式を認識することができるので、表形式文書から
所定項目毎のデータベースを作成する場合に、表形式文
書を文字読取装置で読取るだけで容易に作成することが
できる効果が得られる。
As explained above, in the tabular document recognition method according to claim (1), the vertical ruled lines and the horizontal ruled lines forming the table of the tabular document are binarized by dividing them into strip-shaped areas perpendicular to both. Since ruled line candidates are extracted based on projection data obtained by projecting image information in a direction perpendicular to its longitudinal direction, the number of divisions of binarized image information can be reduced, and the recognition speed of tabular documents can be increased. can be improved,
Moreover, since the back structure is recognized and character recognition is performed for each column of this back structure, it is possible to specify only letters or only numbers for each item in a row or column, and the character recognition device By being able to limit the number of selected characters, it is possible to improve the character recognition rate and recognition speed of the literary journal m device, and it is also possible to recognize tabular formats, so it is possible to create a database for each predetermined item from tabular documents. In this case, an advantage can be obtained that the tabular document can be easily created simply by reading it with a character reading device.

また、請求項(2)に係る表形式文書認識方式にあって
は、前記請求項(1)の効果に加えて、罫線によって区
分けされておらず、空白部分による仮想罫線によって区
分けされている表形式であっても、罫線があるものとし
て認識することができるので、文字認識装置での文字認
識率及び認識速度をより向上させることができる効果が
得られる。
In addition to the effect of claim (1), the tabular document recognition method according to claim (2) also provides a table format that is not divided by ruled lines but is divided by virtual ruled lines formed by blank areas. Even if the character is in a different format, it can be recognized as having ruled lines, so that the character recognition rate and recognition speed of the character recognition device can be further improved.

【図面の簡単な説明】[Brief explanation of the drawing]

第1図はこの発明の概略構成を示すブロック図、第2図
は表形式文書を示す説明図、第3図〜第6図は図形分離
部における処理手順の一例を示すフローチャート、第7
図(a)、 (b)及び(C)は夫々この発明の詳細な
説明に供する水平帯状領域、その垂直投影データ及び再
々2値化データを示す説明図、第8図は仮想罫線処理領
域を示す説明図、第9図は消去2値化情報を示す説明図
、第10図(a)及び働)並びに第11図(a)及び(
b)は夫々仮想罫線抽出処理の説明に供する説明図であ
る。 図中、1はイメージスキャナー(文字読取装置)2は図
形分離部、3は光学文字認識装置、4は演算処理装置、
Dは表形式文書、Rvは垂直罫線、RNは水平罫線であ
る。
FIG. 1 is a block diagram showing a schematic configuration of the present invention, FIG. 2 is an explanatory diagram showing a tabular document, FIGS.
Figures (a), (b), and (C) are explanatory diagrams showing a horizontal band-shaped area, its vertical projection data, and re-binarized data, respectively, to provide a detailed explanation of the present invention, and Fig. 8 shows a virtual ruled line processing area. FIG. 9 is an explanatory diagram showing erased binarized information, FIG. 10 (a) and operation) and FIG. 11 (a) and (
b) is an explanatory diagram for explaining the virtual ruled line extraction process. In the figure, 1 is an image scanner (character reading device), 2 is a graphic separation unit, 3 is an optical character recognition device, 4 is an arithmetic processing unit,
D is a tabular document, Rv is a vertical ruled line, and RN is a horizontal ruled line.

Claims (2)

【特許請求の範囲】[Claims] (1)文字と罫線とが混在する表形式文書を画像読取装
置で読取り、該画像読取装置からの画像情報中の罫線を
切り出し、該罫線によって画成された領域の文字情報を
文字認識装置に送出するようにした表形式文書認識方式
であって、前記画像情報を所定幅の帯状領域に分割し、
各帯状領域をその長手方向と直交する方向に投影した投
影データを形成する投影データ形成手段と、該投影デー
タ形成手段で形成した投影データから罫線候補を抽出す
る罫線候補抽出手段と、該罫線候補抽出手段で抽出した
罫線候補領域を隣接する帯状領域間で連結して罫線を決
定する罫線決定手段とを備えたことを特徴とする表形式
文書認識方式。
(1) Read a tabular document containing a mixture of characters and ruled lines with an image reading device, cut out the ruled lines in the image information from the image reading device, and send the character information in the area defined by the ruled lines to a character recognition device. A tabular document recognition method configured to send the image information into strip-shaped areas of a predetermined width;
projection data forming means for forming projection data obtained by projecting each strip area in a direction orthogonal to the longitudinal direction thereof; a ruled line candidate extracting means for extracting ruled line candidates from the projection data formed by the projection data forming means; and the ruled line candidates. 1. A tabular document recognition method, comprising a ruled line determining means for determining a ruled line by connecting ruled line candidate areas extracted by the extracting means between adjacent strip areas.
(2)前記罫線候補抽出手段は、文字列間に空白領域が
あるときに、当該空白領域から仮想罫線候補を抽出する
ように構成されている請求項(1)記載の表形式文書認
識方式。
(2) The tabular document recognition method according to claim 1, wherein the ruled line candidate extracting means is configured to extract virtual ruled line candidates from the blank area when there is a blank area between character strings.
JP1282736A 1989-10-30 1989-10-30 Tabular document recognition method Expired - Lifetime JP2796561B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP1282736A JP2796561B2 (en) 1989-10-30 1989-10-30 Tabular document recognition method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP1282736A JP2796561B2 (en) 1989-10-30 1989-10-30 Tabular document recognition method

Publications (2)

Publication Number Publication Date
JPH03142691A true JPH03142691A (en) 1991-06-18
JP2796561B2 JP2796561B2 (en) 1998-09-10

Family

ID=17656380

Family Applications (1)

Application Number Title Priority Date Filing Date
JP1282736A Expired - Lifetime JP2796561B2 (en) 1989-10-30 1989-10-30 Tabular document recognition method

Country Status (1)

Country Link
JP (1) JP2796561B2 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5668892A (en) * 1992-10-27 1997-09-16 Fuji Xerox Co., Ltd. Table recognition apparatus
JP2003030585A (en) * 2001-07-12 2003-01-31 Canon Inc Image processor and image processing method, and program and storage medium therefor
CN105426856A (en) * 2015-11-25 2016-03-23 成都数联铭品科技有限公司 Image table character identification method
CN107622233A (en) * 2017-09-11 2018-01-23 畅捷通信息技术股份有限公司 A kind of Table recognition method, identifying system and computer installation

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS61269779A (en) * 1985-05-24 1986-11-29 Agency Of Ind Science & Technol Document picture analyzing system
JPS63101983A (en) * 1986-10-17 1988-05-06 Fujitsu Ltd Character string extracting system
JPH01169686A (en) * 1987-12-25 1989-07-04 Fujitsu Ltd Character line detecting system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS61269779A (en) * 1985-05-24 1986-11-29 Agency Of Ind Science & Technol Document picture analyzing system
JPS63101983A (en) * 1986-10-17 1988-05-06 Fujitsu Ltd Character string extracting system
JPH01169686A (en) * 1987-12-25 1989-07-04 Fujitsu Ltd Character line detecting system

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5668892A (en) * 1992-10-27 1997-09-16 Fuji Xerox Co., Ltd. Table recognition apparatus
US5708730A (en) * 1992-10-27 1998-01-13 Fuji Xerox Co., Ltd. Table recognition apparatus
JP2003030585A (en) * 2001-07-12 2003-01-31 Canon Inc Image processor and image processing method, and program and storage medium therefor
CN105426856A (en) * 2015-11-25 2016-03-23 成都数联铭品科技有限公司 Image table character identification method
CN107622233A (en) * 2017-09-11 2018-01-23 畅捷通信息技术股份有限公司 A kind of Table recognition method, identifying system and computer installation

Also Published As

Publication number Publication date
JP2796561B2 (en) 1998-09-10

Similar Documents

Publication Publication Date Title
JP4271878B2 (en) Character search method and apparatus in video, and character search processing program
EP0843275A2 (en) Pattern extraction apparatus and method for extracting patterns
JP2926066B2 (en) Table recognition device
JPH03142691A (en) Table format document recognizing system
JPH08320914A (en) Table recognition method and device
Nguyen et al. Enhanced character segmentation for format-free Japanese text recognition
JP3904397B2 (en) Table recognition method
JP2890306B2 (en) Table space separation apparatus and table space separation method
CN109409370B (en) Remote desktop character recognition method and device
JP2786361B2 (en) Document image area extraction method
JPS62121589A (en) Character segmenting system
JP2006106931A (en) Character string retrieval device and method and program for the method
JP3083609B2 (en) Information processing apparatus and character recognition apparatus using the same
JP2789622B2 (en) Character / graphic area determination device
JPH0697470B2 (en) Character string extractor
JPH03268181A (en) Document reader
JPH11242716A (en) Image processing method and storage medium
JP4763113B2 (en) High speed labeling method
JP2022051198A (en) Ocr processor, ocr processing method, and program
JP2993533B2 (en) Information processing device and character recognition device
JPS6343788B2 (en)
JP3976804B2 (en) Image processing method
JPH01295383A (en) Area extracting method
JPH03210688A (en) Line detecting device
JPH0394393A (en) Character recognizing device